diff --git a/TensorFlow/contrib/cv/InsightFace_TF/LICENSE b/TensorFlow/contrib/cv/InsightFace_TF/LICENSE
new file mode 100644
index 0000000000000000000000000000000000000000..1562c52dff26c5d846489922bd15eeb8c16f1a80
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/LICENSE
@@ -0,0 +1,21 @@
+MIT License
+
+Copyright (c) 2018 Jiankang Deng and Jia Guo
+
+Permission is hereby granted, free of charge, to any person obtaining a copy
+of this software and associated documentation files (the "Software"), to deal
+in the Software without restriction, including without limitation the rights
+to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
+copies of the Software, and to permit persons to whom the Software is
+furnished to do so, subject to the following conditions:
+
+The above copyright notice and this permission notice shall be included in all
+copies or substantial portions of the Software.
+
+THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
+AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
+OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+SOFTWARE.
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/README.md b/TensorFlow/contrib/cv/InsightFace_TF/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..d1a40c91b1a9ff2dd5c6da9e290ffb237d830f4a
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/README.md
@@ -0,0 +1,197 @@
+- [基本信息](#基本信息.md)
+- [概述](#概述.md)
+- [训练环境准备](#训练环境准备.md)
+- [快速上手](#快速上手.md)
+- [高级参考](#高级参考.md)
+
基本信息
+
+**发布者(Publisher):Huawei**
+
+**应用领域(Application Domain):Object Detection**
+
+**版本(Version):1.1**
+
+**修改时间(Modified) :2022.11.29**
+
+**大小(Size):648KB**
+
+**框架(Framework):TensorFlow-gpu_1.14.0**
+
+**模型格式(Model Format):ckpt**
+
+**精度(Precision):Mixed**
+
+**处理器(Processor):昇腾910**
+
+**应用级别(Categories):Official**
+
+**描述(Description):人脸识别算法**
+
+概述
+
+## 简述
+
+InsightFace 是基于 MXNet 框架实现的业界主流人脸识别解决方案。
+
+- 参考论文:
+
+ [https://arxiv.org/abs/1801.07698](ArcFace: Additive Angular Margin Loss for Deep Face Recognition)
+
+- 参考实现:
+
+ https://github.com/auroua/InsightFace_TF
+
+- 适配昇腾 AI 处理器的实现:
+
+ https://gitee.com/ascend/ModelZoo-TensorFlow/tree/master/TensorFlow/contrib/cv/InsightFace_TF
+
+- 通过Git获取对应commit\_id的代码方法如下:
+
+ git clone {repository_url} # 克隆仓库的代码
+ cd {repository_name} # 切换到模型的代码仓目录
+ git checkout {branch} # 切换到对应分支
+ git reset --hard {commit_id} # 代码设置到对应的commit_id
+ cd {code_path} # 切换到模型代码所在路径,若仓库下只有该模型,则无需切换
+
+
+## 默认配置
+
+- 训练超参(单卡):
+ - Batch size: 32
+ - net_depth:100
+ - epochs:200
+ - momentum:0.9
+ - weight_deacy:5e-4
+ - eval_datasets:lfw
+ - eval_db_path:/datasets/faces_ms1m_112x112
+ - image_size:112,112
+ - tfrecords_file_path: ./datasets/tfrecords_webface
+ - summary_path:./output/summary
+ - ckpt_path:./output/ckpt
+ - log_file_path:./output/logs
+ - saver_maxkeep: 100
+ - buffer_size: 10000
+ - log_device_mapping:False
+ - summary_interval:300
+ - ckpt_interval:10000
+ - validate_interval:2000
+ - show_info_interval:20
+
+## 精度
+
+- GPU上运行精度:0.9880000000000001
+- NPU上运行精度:0.9876666666666665
+## 性能
+
+- GPU:80.461 samples/sec
+- NPU:205.789 samples/sec
+
+
+训练环境准备
+
+- 硬件环境和运行环境准备请参见《[CANN软件安装指南](https://support.huawei.com/enterprise/zh/ascend-computing/cann-pid-251168373?category=installation-update)》
+- 运行以下命令安装依赖。
+```
+pip3 install requirements.txt
+```
+说明:依赖配置文件requirements.txt文件位于模型的根目录
+
+快速上手
+
+## 数据集准备
+
+训练数据集路径
+ datasets/tfrecords_webface
+测试数据集路径
+ datasets/faces_ms1m_112x112
+
+
+## 模型训练
+
+- 运行train_nets.py文件
+- 开始训练。
+
+ - 启动训练之前,首先要配置程序运行相关环境变量。
+
+ 环境变量配置信息参见:
+
+ [Ascend 910训练平台环境变量设置](https://gitee.com/ascend/ModelZoo-TensorFlow/wikis/01.%E8%AE%AD%E7%BB%83%E8%84%9A%E6%9C%AC%E8%BF%81%E7%A7%BB%E6%A1%88%E4%BE%8B/Ascend%20910%E8%AE%AD%E7%BB%83%E5%B9%B3%E5%8F%B0%E7%8E%AF%E5%A2%83%E5%8F%98%E9%87%8F%E8%AE%BE%E7%BD%AE)
+
+ - 单卡训练
+
+ 1.首先在脚本train_nets.py中,配置训练数据集和测试数据集参数如下所示:
+
+ ```
+
+ --tfrecords_file_path=./datasets/tfrecords_webface --eval_db_path=./datasets/faces_ms1m_112x112
+
+ ```
+
+ 2.启动训练
+
+ 启动单卡训练 (脚本为train_nets.py)
+
+ ```
+ python3 train_nets.py
+
+ ```
+
+
+高级参考
+
+## 脚本和示例代码
+
+```
+|--README.md #说明文档
+|--train_nets.py #训练代码
+|--train_nets_mgpu.py
+|--train_nets_mgpu_new.py
+|--train_nets_mgpu_new.py
+|--verification.py
+|--eval_ckpt_file.py
+|--requirements.txt #所需依赖
+|--datasets #训练需要的数据集
+|--data #训练需要的数据集
+| |--eval_data_reader.py
+| |--mx2tfrecords.py #数据集转tfrecords格式脚本
+|--figures
+|--losses
+| |--face_losses.py #损失函数
+|--nets
+| |--imagenet_classes.py
+| |--L_Resnet_E_IR.py
+| |--L_Resnet_E_IR_fix_issue9.py
+| |--L_Resnet_E_IR_GBN.py
+| |--L_Resnet_E_IR_MGPU.py
+| |--L_Resnet_E_IR_RBN.py
+| |--nets_utils.py
+| |--networks.py
+| |--resnet.py
+| |--tl_layers_modify.py
+| |--vgg16.py
+| |--vgg19.py
+|--output
+|--test
+| |--memory_usage_test.py
+| |--resnet_test_static.py
+| |--test_losses.py
+| |--multiple_gpu_test
+| | |--test_mgpu_mnist.py
+| | |--test_tensorlayer.py
+| |--benchmark
+| | |--gluon_batchsize_test.py
+| | |--mxnet_batchsize_test.py
+| | |--resnet_slim_benchmark.py
+| | |--resnet_tl_benchmark.py
+| | |--tensorlayer_batchsize_test.py
+| | |--utils_final.py
+| | |--vgg19_slim_benchmark.py
+| | |--vgg19_tl_benchmark.py
+```
+
+
+
+## 训练过程
+
+通过“模型训练”中的训练指令启动单卡或者多卡训练。
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/data/__init__.py b/TensorFlow/contrib/cv/InsightFace_TF/data/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/data/eval_data_reader.py b/TensorFlow/contrib/cv/InsightFace_TF/data/eval_data_reader.py
new file mode 100644
index 0000000000000000000000000000000000000000..d2ef5b4df481b719400c986928b4d8cd5e97190f
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/data/eval_data_reader.py
@@ -0,0 +1,145 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import numpy as np
+import pickle
+import argparse
+import os
+import mxnet as mx
+import cv2
+import io
+import PIL.Image
+import mxnet.ndarray as nd
+
+
+def get_parser():
+ parser = argparse.ArgumentParser(description='evluation data parser')
+ parser.add_argument('--eval_datasets', default=['lfw', 'cfp_ff', 'cfp_fp', 'agedb_30'], help='evluation datasets')
+ # parser.add_argument('--eval_datasets', default=['cfp_fp'], help='evluation datasets')
+ parser.add_argument('--eval_db_path', default='../datasets/faces_ms1m_112x112', help='evluate datasets base path')
+ parser.add_argument('--image_size', default=[112, 112], help='the image size')
+ parser.add_argument('--tfrecords_file_path', default='../datasets/tfrecords/eval', help='the image size')
+ parser.add_argument('--db_base_path', default='../datasets/faces_ms1m_112x112', help='the image size')
+ args = parser.parse_args()
+ return args
+
+
+def load_bin(path, image_size):
+ '''
+ :param path: the input file path
+ :param image_size: the input image size
+ :return: the returned datasets is opencv format BGR [112, 112, 3]
+ '''
+ bins, issame_list = pickle.load(open(path, 'rb'), encoding='bytes')
+ issame_list_int = list(map(int, issame_list))
+ data_list = []
+ for _ in [0, 1]:
+ data = np.zeros(shape=[len(issame_list)*2, *image_size, 3])
+ data_list.append(data)
+ for i in range(len(issame_list)*2):
+ _bin = bins[i]
+ tf_images = tf.image.decode_jpeg(_bin)
+ tf_images = tf.reshape(tf_images, shape=(112, 112, 3))
+ sess = tf.Session(config=npu_config_proto())
+ images = sess.run(tf_images)
+ img_cv = cv2.cvtColor(images, cv2.COLOR_RGB2BGR)
+ print(np.min(img_cv), np.max(img_cv), img_cv.dtype)
+ cv2.imshow('test', img_cv)
+ cv2.waitKey(0)
+ for flip in [0,1]:
+ if flip == 1:
+ # print(i, flip)
+ img_cv = np.fliplr(img_cv)
+ # cv2.imshow('test', img_cv)
+ # cv2.waitKey(0)
+ data_list[flip][i][:] = img_cv
+ i += 1
+ if i % 1000 == 0:
+ print('loading bin', i)
+ print(data_list[0].shape)
+ return data_list, issame_list
+
+
+def mx2tfrecords(imgidx, imgrec, args):
+ output_path = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ writer = tf.python_io.TFRecordWriter(output_path)
+ for i in imgidx:
+ img_info = imgrec.read_idx(i)
+ header, img = mx.recordio.unpack(img_info)
+ encoded_jpg_io = io.BytesIO(img)
+ image = PIL.Image.open(encoded_jpg_io)
+ np_img = np.array(image)
+ img = cv2.cvtColor(np_img, cv2.COLOR_RGB2BGR)
+ img_raw = img.tobytes()
+ label = int(header.label)
+ example = tf.train.Example(features=tf.train.Features(feature={
+ 'image_raw': tf.train.Feature(bytes_list=tf.train.BytesList(value=[img_raw])),
+ "label": tf.train.Feature(int64_list=tf.train.Int64List(value=[label]))
+ }))
+ writer.write(example.SerializeToString()) # Serialize To String
+ if i % 10000 == 0:
+ print('%d num image processed' % i)
+ writer.close()
+
+
+def mx2tfrecords_eval_data(args, db_name):
+ '''
+ Change evaluation data to tfrecords
+ :param args:
+ :param type: lfw, ......
+ :return:
+ '''
+ bins, issame_list = pickle.load(open(os.path.join(args.db_base_path, db_name+'.bin'), 'rb'), encoding='bytes')
+ output_image_path = os.path.join(args.tfrecords_file_path, db_name+'_eval_data.tfrecords')
+ writer_img = tf.python_io.TFRecordWriter(output_image_path)
+ for i in range(len(bins)):
+ img_info = bins[i]
+ img = mx.image.imdecode(img_info).asnumpy()
+ img = cv2.cvtColor(img, cv2.COLOR_RGB2BGR)
+ img_b = img.tobytes()
+ # # decode test
+ # sess = tf.Session()
+ # img_2 = tf.decode_raw(img_b, out_type=tf.uint8)
+ # img_2 = tf.reshape(img_2, shape=(112, 112, 3))
+ # img_2 = tf.image.flip_left_right(img_2)
+ # img_2_np = sess.run(img_2)
+ # print(img_2_np.shape)
+ # cv2.imshow('test', img_2_np)
+ # cv2.waitKey(0)
+ example = tf.train.Example(features=tf.train.Features(feature={
+ 'image_raw': tf.train.Feature(bytes_list=tf.train.BytesList(value=[img_b]))
+ }))
+ writer_img.write(example.SerializeToString()) # Serialize To String
+ if i % 1000 == 0:
+ print('%d num image processed' % i)
+ writer_img.close()
+
+
+def load_bin(db_name, image_size, args):
+ bins, issame_list = pickle.load(open(os.path.join(args.eval_db_path, db_name+'.bin'), 'rb'), encoding='bytes')
+ data_list = []
+ for _ in [0,1]:
+ data = np.empty((len(issame_list)*2, image_size[0], image_size[1], 3))
+ data_list.append(data)
+ for i in range(len(issame_list)*2):
+ _bin = bins[i]
+ img = mx.image.imdecode(_bin).asnumpy()
+ img = cv2.cvtColor(img, cv2.COLOR_RGB2BGR)
+ for flip in [0,1]:
+ if flip == 1:
+ img = np.fliplr(img)
+ data_list[flip][i, ...] = img
+ i += 1
+ if i % 1000 == 0:
+ print('loading bin', i)
+ print(data_list[0].shape)
+ return data_list, issame_list
+
+
+if __name__ == '__main__':
+ args = get_parser()
+ ver_list = []
+ ver_name_list = []
+ for db in args.eval_datasets:
+ print('begin db %s convert.' % db)
+ # mx2tfrecords_eval_data(args, db)
+ data_set = load_bin(db, args.image_size)
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/data/mx2tfrecords.py b/TensorFlow/contrib/cv/InsightFace_TF/data/mx2tfrecords.py
new file mode 100644
index 0000000000000000000000000000000000000000..3050cc6fa92de5d486cbb8eded4454f9b1c30245
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/data/mx2tfrecords.py
@@ -0,0 +1,128 @@
+from npu_bridge.npu_init import *
+import mxnet as mx
+import argparse
+import PIL.Image
+import io
+import numpy as np
+import cv2
+import tensorflow as tf
+import os
+
+
+def parse_args():
+ parser = argparse.ArgumentParser(
+ formatter_class=argparse.ArgumentDefaultsHelpFormatter,
+ description='data path information'
+ )
+ parser.add_argument('--bin_path', default='../datasets/faces_ms1m_112x112/train.rec', type=str,
+ help='path to the binary image file')
+ parser.add_argument('--idx_path', default='../datasets/faces_ms1m_112x112/train.idx', type=str,
+ help='path to the image index path')
+ parser.add_argument('--tfrecords_file_path', default='../datasets/tfrecords', type=str,
+ help='path to the output of tfrecords file path')
+ args = parser.parse_args()
+ return args
+
+
+def mx2tfrecords_old(imgidx, imgrec, args):
+ output_path = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ writer = tf.python_io.TFRecordWriter(output_path)
+ for i in imgidx:
+ img_info = imgrec.read_idx(i)
+ header, img = mx.recordio.unpack(img_info)
+ encoded_jpg_io = io.BytesIO(img)
+ image = PIL.Image.open(encoded_jpg_io)
+ np_img = np.array(image)
+ img = cv2.cvtColor(np_img, cv2.COLOR_RGB2BGR)
+ img_raw = img.tobytes()
+ label = int(header.label)
+ example = tf.train.Example(features=tf.train.Features(feature={
+ 'image_raw': tf.train.Feature(bytes_list=tf.train.BytesList(value=[img_raw])),
+ "label": tf.train.Feature(int64_list=tf.train.Int64List(value=[label]))
+ }))
+ writer.write(example.SerializeToString()) # Serialize To String
+ if i % 10000 == 0:
+ print('%d num image processed' % i)
+ writer.close()
+
+
+def mx2tfrecords(imgidx, imgrec, args):
+ output_path = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ writer = tf.python_io.TFRecordWriter(output_path)
+ for i in imgidx:
+ img_info = imgrec.read_idx(i)
+ header, img = mx.recordio.unpack(img_info)
+ label = int(header.label)
+ example = tf.train.Example(features=tf.train.Features(feature={
+ 'image_raw': tf.train.Feature(bytes_list=tf.train.BytesList(value=[img])),
+ "label": tf.train.Feature(int64_list=tf.train.Int64List(value=[label]))
+ }))
+ writer.write(example.SerializeToString()) # Serialize To String
+ if i % 10000 == 0:
+ print('%d num image processed' % i)
+ writer.close()
+
+
+def parse_function(example_proto):
+ features = {'image_raw': tf.FixedLenFeature([], tf.string),
+ 'label': tf.FixedLenFeature([], tf.int64)}
+ features = tf.parse_single_example(example_proto, features)
+ # You can do more image distortion here for training data
+ img = tf.image.decode_jpeg(features['image_raw'])
+ img = tf.reshape(img, shape=(112, 112, 3))
+ r, g, b = tf.split(img, num_or_size_splits=3, axis=-1)
+ img = tf.concat([b, g, r], axis=-1)
+ img = tf.cast(img, dtype=tf.float32)
+ img = tf.subtract(img, 127.5)
+ img = tf.multiply(img, 0.0078125)
+ img = tf.image.random_flip_left_right(img)
+ label = tf.cast(features['label'], tf.int64)
+ return img, label
+
+
+if __name__ == '__main__':
+ # define parameters
+ id2range = {}
+ data_shape = (3, 112, 112)
+ args = parse_args()
+ imgrec = mx.recordio.MXIndexedRecordIO(args.idx_path, args.bin_path, 'r')
+ s = imgrec.read_idx(0)
+ header, _ = mx.recordio.unpack(s)
+ print(header.label)
+ imgidx = list(range(1, int(header.label[0])))
+ seq_identity = range(int(header.label[0]), int(header.label[1]))
+ for identity in seq_identity:
+ s = imgrec.read_idx(identity)
+ header, _ = mx.recordio.unpack(s)
+ a, b = int(header.label[0]), int(header.label[1])
+ id2range[identity] = (a, b)
+ print('id2range', len(id2range))
+
+ # generate tfrecords
+ mx2tfrecords(imgidx, imgrec, args)
+
+ config = tf.ConfigProto(allow_soft_placement=True)
+ sess = tf.Session(config=npu_config_proto(config_proto=config))
+ # training datasets api config
+ tfrecords_f = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ dataset = tf.data.TFRecordDataset(tfrecords_f)
+ dataset = dataset.map(parse_function)
+ dataset = dataset.shuffle(buffer_size=30000)
+ dataset = dataset.batch(32, drop_remainder=True)
+ iterator = dataset.make_initializable_iterator()
+ next_element = iterator.get_next()
+ # begin iteration
+ for i in range(1000):
+ sess.run(iterator.initializer)
+ while True:
+ try:
+ images, labels = sess.run(next_element)
+ cv2.imshow('test', images[1, ...])
+ cv2.waitKey(0)
+ except tf.errors.OutOfRangeError:
+ print("End of dataset")
+
+
+
+
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/eval_ckpt_file.py b/TensorFlow/contrib/cv/InsightFace_TF/eval_ckpt_file.py
new file mode 100644
index 0000000000000000000000000000000000000000..88a09c75ff144f3f3889fd612408095574059e97
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/eval_ckpt_file.py
@@ -0,0 +1,66 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import argparse
+from data.eval_data_reader import load_bin
+from losses.face_losses import arcface_loss
+from nets.L_Resnet_E_IR import get_resnet
+import tensorlayer as tl
+from verification import ver_test
+
+
+def get_args():
+ parser = argparse.ArgumentParser(description='input information')
+ parser.add_argument('--ckpt_file', default='/home/aurora/workspaces2018/InsightFace_TF/output/ckpt_model_c/InsightFace_iter_best_',
+ type=str, help='the ckpt file path')
+ # parser.add_argument('--eval_datasets', default=['lfw', 'cfp_ff', 'cfp_fp', 'agedb_30'], help='evluation datasets')
+ parser.add_argument('--eval_datasets', default=['agedb_30'], help='evluation datasets')
+ parser.add_argument('--eval_db_path', default='./datasets/faces_ms1m_112x112', help='evluate datasets base path')
+ parser.add_argument('--image_size', default=[112, 112], help='the image size')
+ parser.add_argument('--net_depth', default=50, help='resnet depth, default is 50')
+ parser.add_argument('--num_output', default=85164, help='the image size')
+ parser.add_argument('--batch_size', default=32, help='batch size to train network')
+ parser.add_argument('--ckpt_index_list',
+ default=['1950000.ckpt'], help='ckpt file indexes')
+ args = parser.parse_args()
+ return args
+
+
+if __name__ == '__main__':
+ args = get_args()
+ ver_list = []
+ ver_name_list = []
+ for db in args.eval_datasets:
+ print('begin db %s convert.' % db)
+ data_set = load_bin(db, args.image_size, args)
+ ver_list.append(data_set)
+ ver_name_list.append(db)
+
+ images = tf.placeholder(name='img_inputs', shape=[None, *args.image_size, 3], dtype=tf.float32)
+ labels = tf.placeholder(name='img_labels', shape=[None, ], dtype=tf.int64)
+ dropout_rate = tf.placeholder(name='dropout_rate', dtype=tf.float32)
+
+ w_init_method = tf.contrib.layers.xavier_initializer(uniform=False)
+ net = get_resnet(images, args.net_depth, type='ir', w_init=w_init_method, trainable=False, keep_rate=dropout_rate)
+ embedding_tensor = net.outputs
+ # mv_mean = tl.layers.get_variables_with_name('resnet_v1_50/bn0/moving_mean', False, True)[0]
+ # 3.2 get arcface loss
+ logit = arcface_loss(embedding=net.outputs, labels=labels, w_init=w_init_method, out_num=args.num_output)
+
+ sess = tf.Session(config=npu_config_proto())
+ saver = tf.train.Saver()
+
+ result_index = []
+ for file_index in args.ckpt_index_list:
+ feed_dict_test = {}
+ path = args.ckpt_file + file_index
+ saver.restore(sess, path)
+ print('ckpt file %s restored!' % file_index)
+ feed_dict_test.update(tl.utils.dict_to_one(net.all_drop))
+ feed_dict_test[dropout_rate] = 1.0
+ results = ver_test(ver_list=ver_list, ver_name_list=ver_name_list, nbatch=0, sess=sess,
+ embedding_tensor=embedding_tensor, batch_size=args.batch_size, feed_dict=feed_dict_test,
+ input_placeholder=images)
+ result_index.append(results)
+ print(result_index)
+
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/figures/if_gpu.log b/TensorFlow/contrib/cv/InsightFace_TF/figures/if_gpu.log
new file mode 100644
index 0000000000000000000000000000000000000000..9ee5e1253da28e03360c39ee8e3523a8bff52e08
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/figures/if_gpu.log
@@ -0,0 +1,17005 @@
+nohup: ignoring input
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:516: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint8 = np.dtype([("qint8", np.int8, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:517: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_quint8 = np.dtype([("quint8", np.uint8, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:518: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint16 = np.dtype([("qint16", np.int16, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:519: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_quint16 = np.dtype([("quint16", np.uint16, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:520: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint32 = np.dtype([("qint32", np.int32, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:525: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ np_resource = np.dtype([("resource", np.ubyte, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:541: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint8 = np.dtype([("qint8", np.int8, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:542: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_quint8 = np.dtype([("quint8", np.uint8, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:543: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint16 = np.dtype([("qint16", np.int16, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:544: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_quint16 = np.dtype([("quint16", np.uint16, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:545: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint32 = np.dtype([("qint32", np.int32, 1)])
+/home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:550: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ np_resource = np.dtype([("resource", np.ubyte, 1)])
+WARNING:tensorflow:From /home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorlayer/layers.py:28: The name tf.GraphKeys is deprecated. Please use tf.compat.v1.GraphKeys instead.
+
+WARNING:tensorflow:From /home/wangdawei/.conda/envs/insightface/lib/python3.7/site-packages/tensorlayer/layers.py:4030: The name tf.nn.max_pool is deprecated. Please use tf.nn.max_pool2d instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:50: The name tf.assign_add is deprecated. Please use tf.compat.v1.assign_add instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:51: The name tf.placeholder is deprecated. Please use tf.compat.v1.placeholder instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/data/mx2tfrecords.py:63: The name tf.FixedLenFeature is deprecated. Please use tf.io.FixedLenFeature instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/data/mx2tfrecords.py:65: The name tf.parse_single_example is deprecated. Please use tf.io.parse_single_example instead.
+
+2022-11-30 21:16:48.665426: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcuda.so.1
+2022-11-30 21:16:48.807033: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1640] Found device 0 with properties:
+name: NVIDIA TITAN Xp major: 6 minor: 1 memoryClockRate(GHz): 1.582
+pciBusID: 0000:04:00.0
+2022-11-30 21:16:48.807238: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudart.so.10.1
+2022-11-30 21:16:48.808938: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcublas.so.10
+2022-11-30 21:16:48.810485: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcufft.so.10
+2022-11-30 21:16:48.810740: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcurand.so.10
+2022-11-30 21:16:48.812521: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusolver.so.10
+2022-11-30 21:16:48.813586: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusparse.so.10
+2022-11-30 21:16:48.817749: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
+2022-11-30 21:16:48.824898: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1763] Adding visible gpu devices: 0
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:64: DatasetV1.make_initializable_iterator (from tensorflow.python.data.ops.dataset_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Use `for ... in dataset:` to iterate over a dataset. If using `tf.estimator`, return the `Dataset` object directly from your input function. As a last resort, you can use `tf.compat.v1.data.make_initializable_iterator(dataset)`.
+begin db lfw convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+(12000, 112, 112, 3)
+begin db cfp_ff convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+loading bin 13000
+loading bin 14000
+(14000, 112, 112, 3)
+begin db cfp_fp convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+loading bin 13000
+loading bin 14000
+(14000, 112, 112, 3)
+begin db agedb_30 convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+(12000, 112, 112, 3)
+ [TL] InputLayer resnet_v1_100/input_layer: (?, 112, 112, 3)
+ [TL] Conv2dLayer resnet_v1_100/conv1: shape:[3, 3, 3, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/bn0: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/bn0/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/bn0/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/prelu0: channel_shared:False
+ [TL] PoolLayer resnet_v1_100/block1/unit_1/bottleneck_v1/shortcut: ksize:[1, 1, 1, 1] strides:[1, 2, 2, 1] padding:SAME pool:max_pool
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block1/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_1/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_2/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_3/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 64, 128] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block2/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_1/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_2/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_3/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_4/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_5/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_6/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_7/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_8/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_9/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_10/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_11/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_12/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_13/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 128, 256] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1: shape:[3, 3, 128, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block3/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_1/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_2/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_3/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_4/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_5/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_6/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_7/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_8/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_9/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_10/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_11/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_12/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_13/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_14/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_15/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_16/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_17/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_18/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_19/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_20/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_21/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_22/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_23/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_24/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_25/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_26/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_27/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_28/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_29/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_30/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 256, 512] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1: shape:[3, 3, 256, 512] strides:[1, 1, 1, 1] pad:SAME act:identityWARNING:tensorflow:From /home/wangdawei/InsightFace_TF/nets/L_Resnet_E_IR_fix_issue9.py:314: calling dropout (from tensorflow.python.ops.nn_ops) with keep_prob is deprecated and will be removed in a future version.
+Instructions for updating:
+Please use `rate` instead of `keep_prob`. Rate should be set to `rate = 1 - keep_prob`.
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/losses/face_losses.py:20: calling norm (from tensorflow.python.ops.linalg_ops) with keep_dims is deprecated and will be removed in a future version.
+Instructions for updating:
+keep_dims is deprecated, use keepdims instead
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/losses/face_losses.py:21: div (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Deprecated in favor of operator or tf.math.divide.
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/losses/face_losses.py:40: add_dispatch_support..wrapper (from tensorflow.python.ops.array_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Use tf.where in 2.0, which has the same broadcast rule as np.where
+
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block4/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_1/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_2/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_3/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/E_BN1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/E_BN1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/E_BN1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ReshapeLayer resnet_v1_100/E_Reshapelayer: (?, 25088)
+ [TL] DenseLayer resnet_v1_100/E_DenseLayer: 512 identity
+ [TL] BatchNormLayer resnet_v1_100/E_BN2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/E_BN2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/E_BN2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] InputLayer resnet_v1_100/input_layer: (?, 112, 112, 3)
+ [TL] Conv2dLayer resnet_v1_100/conv1: shape:[3, 3, 3, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/bn0: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/prelu0: channel_shared:False
+ [TL] PoolLayer resnet_v1_100/block1/unit_1/bottleneck_v1/shortcut: ksize:[1, 1, 1, 1] strides:[1, 2, 2, 1] padding:SAME pool:max_pool
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block1/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_1/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_2/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_3/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 64, 128] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block2/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_1/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_2/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_3/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_4/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_5/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_6/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_7/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_8/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_9/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_10/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_11/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_12/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_13/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 128, 256] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1: shape:[3, 3, 128, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block3/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_1/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_2/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_3/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_4/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_5/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_6/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_7/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_8/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_9/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_10/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_11/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_12/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_13/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_14/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_15/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_16/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_17/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_18/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_19/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_20/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_21/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_22/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_23/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_24/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_25/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_26/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_27/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_28/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_29/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_30/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 256, 512] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1: shape:[3, 3, 256, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block4/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_1/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_2/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_3/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/E_BN1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ReshapeLayer resnet_v1_100/E_Reshapelayer: (?, 25088)
+ [TL] DenseLayer resnet_v1_100/E_DenseLayer: 512 identity
+ [TL] BatchNormLayer resnet_v1_100/E_BN2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [*] geting variables with W_conv2d
+ got 0: resnet_v1_100/conv1/W_conv2d:0 (3, 3, 3, 64)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 2: resnet_v1_100/block1/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 3: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 4: resnet_v1_100/block1/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 5: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 6: resnet_v1_100/block1/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 7: resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 64, 128)
+ got 8: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 128)
+ got 9: resnet_v1_100/block2/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 10: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 11: resnet_v1_100/block2/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 12: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 13: resnet_v1_100/block2/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 14: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 15: resnet_v1_100/block2/unit_4/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 16: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 17: resnet_v1_100/block2/unit_5/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 18: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 19: resnet_v1_100/block2/unit_6/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 20: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 21: resnet_v1_100/block2/unit_7/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 22: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 23: resnet_v1_100/block2/unit_8/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 24: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 25: resnet_v1_100/block2/unit_9/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 26: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 27: resnet_v1_100/block2/unit_10/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 28: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 29: resnet_v1_100/block2/unit_11/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 30: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 31: resnet_v1_100/block2/unit_12/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 32: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 33: resnet_v1_100/block2/unit_13/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 34: resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 128, 256)
+ got 35: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 256)
+ got 36: resnet_v1_100/block3/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 37: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 38: resnet_v1_100/block3/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 39: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 40: resnet_v1_100/block3/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 41: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 42: resnet_v1_100/block3/unit_4/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 43: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 44: resnet_v1_100/block3/unit_5/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 45: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 46: resnet_v1_100/block3/unit_6/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 47: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 48: resnet_v1_100/block3/unit_7/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 49: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 50: resnet_v1_100/block3/unit_8/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 51: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 52: resnet_v1_100/block3/unit_9/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 53: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 54: resnet_v1_100/block3/unit_10/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 55: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 56: resnet_v1_100/block3/unit_11/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 57: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 58: resnet_v1_100/block3/unit_12/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 59: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 60: resnet_v1_100/block3/unit_13/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 61: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 62: resnet_v1_100/block3/unit_14/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 63: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 64: resnet_v1_100/block3/unit_15/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 65: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 66: resnet_v1_100/block3/unit_16/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 67: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 68: resnet_v1_100/block3/unit_17/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 69: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 70: resnet_v1_100/block3/unit_18/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 71: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 72: resnet_v1_100/block3/unit_19/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 73: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 74: resnet_v1_100/block3/unit_20/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 75: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 76: resnet_v1_100/block3/unit_21/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 77: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 78: resnet_v1_100/block3/unit_22/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 79: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 80: resnet_v1_100/block3/unit_23/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 81: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 82: resnet_v1_100/block3/unit_24/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 83: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 84: resnet_v1_100/block3/unit_25/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 85: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 86: resnet_v1_100/block3/unit_26/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 87: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 88: resnet_v1_100/block3/unit_27/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 89: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 90: resnet_v1_100/block3/unit_28/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 91: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 92: resnet_v1_100/block3/unit_29/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 93: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 94: resnet_v1_100/block3/unit_30/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 95: resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 256, 512)
+ got 96: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 512)
+ got 97: resnet_v1_100/block4/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ got 98: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 512, 512)
+ got 99: resnet_v1_100/block4/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ got 100: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 512, 512)
+ got 101: resnet_v1_100/block4/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ [*] geting variables with resnet_v1_50/E_DenseLayer/W
+ [*] geting variables with embedding_weights
+ got 0: arcface_loss/embedding_weights:0 (512, 85164)
+ [*] geting variables with gamma
+ got 0: resnet_v1_100/bn0/gamma:0 (64,)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 2: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 3: resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 4: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 5: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 6: resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 7: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 8: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 9: resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 10: resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (128,)
+ got 11: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 12: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 13: resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 14: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 15: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 16: resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 17: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 18: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 19: resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 20: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 21: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 22: resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 23: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 24: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 25: resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 26: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 27: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 28: resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 29: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 30: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 31: resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 32: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 33: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 34: resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 35: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 36: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 37: resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 38: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 39: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 40: resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 41: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 42: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 43: resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 44: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 45: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 46: resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 47: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 48: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 49: resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 50: resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (256,)
+ got 51: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 52: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 53: resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 54: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 55: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 56: resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 57: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 58: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 59: resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 60: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 61: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 62: resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 63: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 64: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 65: resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 66: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 67: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 68: resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 69: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 70: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 71: resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 72: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 73: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 74: resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 75: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 76: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 77: resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 78: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 79: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 80: resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 81: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 82: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 83: resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 84: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 85: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 86: resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 87: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 88: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 89: resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 90: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 91: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 92: resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 93: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 94: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 95: resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 96: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 97: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 98: resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 99: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 100: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 101: resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 102: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 103: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 104: resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 105: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 106: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 107: resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 108: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 109: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 110: resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 111: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 112: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 113: resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 114: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 115: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 116: resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 117: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 118: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 119: resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 120: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 121: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 122: resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 123: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 124: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 125: resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 126: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 127: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 128: resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 129: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 130: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 131: resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 132: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 133: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 134: resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 135: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 136: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 137: resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 138: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 139: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 140: resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 141: resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (512,)
+ got 142: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 143: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 144: resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 145: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (512,)
+ got 146: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 147: resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 148: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (512,)
+ got 149: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 150: resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 151: resnet_v1_100/E_BN1/gamma:0 (512,)
+ [*] geting variables with alphas
+ got 0: resnet_v1_100/prelu0/alphas:0 (64,)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 2: resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 3: resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 4: resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 5: resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 6: resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 7: resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 8: resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 9: resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 10: resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 11: resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 12: resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 13: resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 14: resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 15: resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 16: resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 17: resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 18: resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 19: resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 20: resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 21: resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 22: resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer/alphas:0 (256,)WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:113: The name tf.train.piecewise_constant is deprecated. Please use tf.compat.v1.train.piecewise_constant instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:115: The name tf.train.MomentumOptimizer is deprecated. Please use tf.compat.v1.train.MomentumOptimizer instead.
+
+2022-11-30 21:18:28.811504: I tensorflow/core/platform/cpu_feature_guard.cc:142] Your CPU supports instructions that this TensorFlow binary was not compiled to use: SSE4.1 SSE4.2 AVX AVX2 FMA
+2022-11-30 21:18:28.845049: I tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency: 2999970000 Hz
+2022-11-30 21:18:28.848269: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x9c70f70 executing computations on platform Host. Devices:
+2022-11-30 21:18:28.848328: I tensorflow/compiler/xla/service/service.cc:175] StreamExecutor device (0): ,
+2022-11-30 21:18:29.175044: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x2829c90 executing computations on platform CUDA. Devices:
+2022-11-30 21:18:29.175112: I tensorflow/compiler/xla/service/service.cc:175] StreamExecutor device (0): NVIDIA TITAN Xp, Compute Capability 6.1
+2022-11-30 21:18:29.179947: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1640] Found device 0 with properties:
+name: NVIDIA TITAN Xp major: 6 minor: 1 memoryClockRate(GHz): 1.582
+pciBusID: 0000:04:00.0
+2022-11-30 21:18:29.180246: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudart.so.10.1
+2022-11-30 21:18:29.180312: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcublas.so.10
+2022-11-30 21:18:29.180397: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcufft.so.10
+2022-11-30 21:18:29.180456: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcurand.so.10
+2022-11-30 21:18:29.180510: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusolver.so.10
+2022-11-30 21:18:29.180576: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusparse.so.10
+2022-11-30 21:18:29.180633: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
+2022-11-30 21:18:29.191273: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1763] Adding visible gpu devices: 0
+2022-11-30 21:18:29.191323: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudart.so.10.1
+2022-11-30 21:18:29.194778: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1181] Device interconnect StreamExecutor with strength 1 edge matrix:
+2022-11-30 21:18:29.194796: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1187] 0
+2022-11-30 21:18:29.194803: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1200] 0: N
+2022-11-30 21:18:29.201064: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1326] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 10829 MB memory) -> physical GPU (device: 0, name: NVIDIA TITAN Xp, pci bus id: 0000:04:00.0, compute capability: 6.1)
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:132: The name tf.summary.FileWriter is deprecated. Please use tf.compat.v1.summary.FileWriter instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:137: The name tf.summary.histogram is deprecated. Please use tf.compat.v1.summary.histogram instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:142: The name tf.summary.scalar is deprecated. Please use tf.compat.v1.summary.scalar instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:147: The name tf.summary.merge is deprecated. Please use tf.compat.v1.summary.merge instead.
+
+WARNING:tensorflow:From /home/wangdawei/InsightFace_TF/train_nets.py:149: The name tf.train.Saver is deprecated. Please use tf.compat.v1.train.Saver instead.
+
+2022-11-30 21:18:38.110777: W tensorflow/compiler/jit/mark_for_compilation_pass.cc:1412] (One-time warning): Not using XLA:CPU for cluster because envvar TF_XLA_FLAGS=--tf_xla_cpu_global_jit was not set. If you want XLA:CPU, either set that envvar, or use experimental_jit_scope to enable XLA:CPU. To confirm that XLA is active, pass --vmodule=xla_compilation_cache=1 (as a proper command-line flag, not via TF_XLA_FLAGS) or set the envvar XLA_FLAGS=--xla_hlo_profile.
+2022-11-30 21:19:14.281975: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcublas.so.10
+2022-11-30 21:19:14.561981: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
+
+ got 23: resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 24: resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 25: resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 26: resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 27: resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 28: resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 29: resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 30: resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 31: resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 32: resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 33: resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 34: resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 35: resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 36: resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 37: resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 38: resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 39: resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 40: resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 41: resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 42: resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 43: resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 44: resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 45: resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 46: resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 47: resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer/alphas:0 (512,)
+ got 48: resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer/alphas:0 (512,)
+ got 49: resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer/alphas:0 (512,)
+[640000, 960000, 1280000]
+epoch 0, total_step 20, total loss is 58.21 , inference loss is 43.74, weight deacy loss is 14.47, training accuracy is 0.000000, time 56.410 samples/sec
+epoch 0, total_step 40, total loss is 57.38 , inference loss is 42.91, weight deacy loss is 14.47, training accuracy is 0.000000, time 57.553 samples/sec
+epoch 0, total_step 60, total loss is 55.19 , inference loss is 40.72, weight deacy loss is 14.47, training accuracy is 0.000000, time 55.945 samples/sec
+epoch 0, total_step 80, total loss is 54.33 , inference loss is 39.86, weight deacy loss is 14.47, training accuracy is 0.000000, time 56.522 samples/sec
+epoch 0, total_step 100, total loss is 51.04 , inference loss is 36.57, weight deacy loss is 14.47, training accuracy is 0.000000, time 56.807 samples/sec
+epoch 0, total_step 120, total loss is 53.72 , inference loss is 39.25, weight deacy loss is 14.47, training accuracy is 0.000000, time 55.201 samples/sec
+epoch 0, total_step 140, total loss is 52.19 , inference loss is 37.72, weight deacy loss is 14.47, training accuracy is 0.000000, time 56.912 samples/sec
+epoch 0, total_step 160, total loss is 48.87 , inference loss is 34.40, weight deacy loss is 14.47, training accuracy is 0.000000, time 57.123 samples/sec
+epoch 0, total_step 180, total loss is 49.36 , inference loss is 34.89, weight deacy loss is 14.47, training accuracy is 0.000000, time 56.634 samples/sec
+epoch 0, total_step 200, total loss is 50.67 , inference loss is 36.21, weight deacy loss is 14.47, training accuracy is 0.000000, time 55.820 samples/sec
+epoch 0, total_step 220, total loss is 49.16 , inference loss is 34.69, weight deacy loss is 14.46, training accuracy is 0.000000, time 56.435 samples/sec
+epoch 0, total_step 240, total loss is 50.29 , inference loss is 35.83, weight deacy loss is 14.46, training accuracy is 0.000000, time 55.984 samples/sec
+epoch 0, total_step 260, total loss is 50.15 , inference loss is 35.69, weight deacy loss is 14.46, training accuracy is 0.000000, time 56.288 samples/sec
+epoch 0, total_step 280, total loss is 49.38 , inference loss is 34.92, weight deacy loss is 14.46, training accuracy is 0.000000, time 57.451 samples/sec
+epoch 0, total_step 300, total loss is 46.74 , inference loss is 32.28, weight deacy loss is 14.46, training accuracy is 0.000000, time 55.443 samples/sec
+epoch 0, total_step 320, total loss is 49.28 , inference loss is 34.83, weight deacy loss is 14.46, training accuracy is 0.000000, time 56.080 samples/sec
+epoch 0, total_step 340, total loss is 51.66 , inference loss is 37.21, weight deacy loss is 14.46, training accuracy is 0.031250, time 56.472 samples/sec
+epoch 0, total_step 360, total loss is 49.01 , inference loss is 34.56, weight deacy loss is 14.45, training accuracy is 0.000000, time 55.941 samples/sec
+epoch 0, total_step 380, total loss is 47.85 , inference loss is 33.40, weight deacy loss is 14.45, training accuracy is 0.000000, time 56.119 samples/sec
+epoch 0, total_step 400, total loss is 45.20 , inference loss is 30.74, weight deacy loss is 14.45, training accuracy is 0.031250, time 56.649 samples/sec
+epoch 0, total_step 420, total loss is 51.02 , inference loss is 36.57, weight deacy loss is 14.45, training accuracy is 0.000000, time 56.551 samples/sec
+epoch 0, total_step 440, total loss is 47.74 , inference loss is 33.30, weight deacy loss is 14.45, training accuracy is 0.000000, time 56.638 samples/sec
+epoch 0, total_step 460, total loss is 48.47 , inference loss is 34.02, weight deacy loss is 14.45, training accuracy is 0.031250, time 55.755 samples/sec
+epoch 0, total_step 480, total loss is 46.88 , inference loss is 32.44, weight deacy loss is 14.44, training accuracy is 0.062500, time 56.403 samples/sec
+epoch 0, total_step 500, total loss is 46.05 , inference loss is 31.61, weight deacy loss is 14.44, training accuracy is 0.031250, time 56.126 samples/sec
+epoch 0, total_step 520, total loss is 50.54 , inference loss is 36.10, weight deacy loss is 14.44, training accuracy is 0.000000, time 56.829 samples/sec
+epoch 0, total_step 540, total loss is 48.20 , inference loss is 33.76, weight deacy loss is 14.44, training accuracy is 0.000000, time 53.533 samples/sec
+epoch 0, total_step 560, total loss is 49.94 , inference loss is 35.50, weight deacy loss is 14.44, training accuracy is 0.000000, time 56.925 samples/sec
+epoch 0, total_step 580, total loss is 51.03 , inference loss is 36.59, weight deacy loss is 14.44, training accuracy is 0.000000, time 55.968 samples/sec
+epoch 0, total_step 600, total loss is 47.38 , inference loss is 32.94, weight deacy loss is 14.43, training accuracy is 0.062500, time 53.567 samples/sec
+epoch 0, total_step 620, total loss is 46.22 , inference loss is 31.78, weight deacy loss is 14.43, training accuracy is 0.000000, time 54.024 samples/sec
+epoch 0, total_step 640, total loss is 43.97 , inference loss is 29.54, weight deacy loss is 14.43, training accuracy is 0.000000, time 56.529 samples/sec
+epoch 0, total_step 660, total loss is 47.43 , inference loss is 33.00, weight deacy loss is 14.43, training accuracy is 0.000000, time 56.209 samples/sec
+epoch 0, total_step 680, total loss is 52.52 , inference loss is 38.09, weight deacy loss is 14.43, training accuracy is 0.000000, time 56.710 samples/sec
+epoch 0, total_step 700, total loss is 50.66 , inference loss is 36.23, weight deacy loss is 14.43, training accuracy is 0.000000, time 56.823 samples/sec
+epoch 0, total_step 720, total loss is 49.54 , inference loss is 35.12, weight deacy loss is 14.42, training accuracy is 0.000000, time 57.090 samples/sec
+epoch 0, total_step 740, total loss is 47.01 , inference loss is 32.59, weight deacy loss is 14.42, training accuracy is 0.000000, time 55.788 samples/sec
+epoch 0, total_step 760, total loss is 48.82 , inference loss is 34.39, weight deacy loss is 14.42, training accuracy is 0.000000, time 56.488 samples/sec
+epoch 0, total_step 780, total loss is 48.11 , inference loss is 33.69, weight deacy loss is 14.42, training accuracy is 0.000000, time 57.274 samples/sec
+epoch 0, total_step 800, total loss is 47.04 , inference loss is 32.63, weight deacy loss is 14.42, training accuracy is 0.000000, time 55.574 samples/sec
+epoch 0, total_step 820, total loss is 46.96 , inference loss is 32.54, weight deacy loss is 14.42, training accuracy is 0.000000, time 56.380 samples/sec
+epoch 0, total_step 840, total loss is 46.60 , inference loss is 32.18, weight deacy loss is 14.42, training accuracy is 0.000000, time 55.713 samples/sec
+epoch 0, total_step 860, total loss is 48.05 , inference loss is 33.63, weight deacy loss is 14.41, training accuracy is 0.031250, time 56.100 samples/sec
+epoch 0, total_step 880, total loss is 47.77 , inference loss is 33.36, weight deacy loss is 14.41, training accuracy is 0.000000, time 56.159 samples/sec
+epoch 0, total_step 900, total loss is 45.44 , inference loss is 31.03, weight deacy loss is 14.41, training accuracy is 0.000000, time 56.687 samples/sec
+epoch 0, total_step 920, total loss is 50.99 , inference loss is 36.58, weight deacy loss is 14.41, training accuracy is 0.000000, time 54.686 samples/sec
+epoch 0, total_step 940, total loss is 45.35 , inference loss is 30.94, weight deacy loss is 14.41, training accuracy is 0.000000, time 56.457 samples/sec
+epoch 0, total_step 960, total loss is 44.71 , inference loss is 30.30, weight deacy loss is 14.41, training accuracy is 0.000000, time 55.984 samples/sec
+epoch 0, total_step 980, total loss is 45.67 , inference loss is 31.27, weight deacy loss is 14.40, training accuracy is 0.031250, time 56.513 samples/sec
+epoch 0, total_step 1000, total loss is 46.44 , inference loss is 32.04, weight deacy loss is 14.40, training accuracy is 0.000000, time 56.570 samples/sec
+epoch 0, total_step 1020, total loss is 46.85 , inference loss is 32.45, weight deacy loss is 14.40, training accuracy is 0.031250, time 55.947 samples/sec
+epoch 0, total_step 1040, total loss is 46.83 , inference loss is 32.43, weight deacy loss is 14.40, training accuracy is 0.000000, time 57.063 samples/sec
+epoch 0, total_step 1060, total loss is 49.62 , inference loss is 35.23, weight deacy loss is 14.40, training accuracy is 0.031250, time 56.729 samples/sec
+epoch 0, total_step 1080, total loss is 44.60 , inference loss is 30.20, weight deacy loss is 14.40, training accuracy is 0.031250, time 56.032 samples/sec
+epoch 0, total_step 1100, total loss is 46.83 , inference loss is 32.43, weight deacy loss is 14.39, training accuracy is 0.000000, time 55.926 samples/sec
+epoch 0, total_step 1120, total loss is 46.29 , inference loss is 31.90, weight deacy loss is 14.39, training accuracy is 0.000000, time 55.506 samples/sec
+epoch 0, total_step 1140, total loss is 46.27 , inference loss is 31.88, weight deacy loss is 14.39, training accuracy is 0.000000, time 56.381 samples/sec
+epoch 0, total_step 1160, total loss is 48.53 , inference loss is 34.14, weight deacy loss is 14.39, training accuracy is 0.000000, time 56.856 samples/sec
+epoch 0, total_step 1180, total loss is 46.41 , inference loss is 32.02, weight deacy loss is 14.39, training accuracy is 0.000000, time 57.186 samples/sec
+epoch 0, total_step 1200, total loss is 43.89 , inference loss is 29.50, weight deacy loss is 14.39, training accuracy is 0.000000, time 56.548 samples/sec
+epoch 0, total_step 1220, total loss is 51.58 , inference loss is 37.20, weight deacy loss is 14.38, training accuracy is 0.000000, time 56.559 samples/sec
+epoch 0, total_step 1240, total loss is 46.62 , inference loss is 32.24, weight deacy loss is 14.38, training accuracy is 0.000000, time 56.403 samples/sec
+epoch 0, total_step 1260, total loss is 49.93 , inference loss is 35.55, weight deacy loss is 14.38, training accuracy is 0.000000, time 56.704 samples/sec
+epoch 0, total_step 1280, total loss is 43.50 , inference loss is 29.12, weight deacy loss is 14.38, training accuracy is 0.093750, time 55.614 samples/sec
+epoch 0, total_step 1300, total loss is 42.16 , inference loss is 27.78, weight deacy loss is 14.38, training accuracy is 0.000000, time 56.365 samples/sec
+epoch 0, total_step 1320, total loss is 45.78 , inference loss is 31.40, weight deacy loss is 14.38, training accuracy is 0.000000, time 56.670 samples/sec
+epoch 0, total_step 1340, total loss is 44.76 , inference loss is 30.38, weight deacy loss is 14.37, training accuracy is 0.000000, time 56.124 samples/sec
+epoch 0, total_step 1360, total loss is 44.90 , inference loss is 30.53, weight deacy loss is 14.37, training accuracy is 0.031250, time 56.140 samples/sec
+epoch 0, total_step 1380, total loss is 45.29 , inference loss is 30.92, weight deacy loss is 14.37, training accuracy is 0.031250, time 57.506 samples/sec
+epoch 0, total_step 1400, total loss is 43.49 , inference loss is 29.12, weight deacy loss is 14.37, training accuracy is 0.125000, time 56.094 samples/sec
+epoch 0, total_step 1420, total loss is 47.50 , inference loss is 33.13, weight deacy loss is 14.37, training accuracy is 0.000000, time 56.965 samples/sec
+epoch 0, total_step 1440, total loss is 48.81 , inference loss is 34.45, weight deacy loss is 14.37, training accuracy is 0.000000, time 56.227 samples/sec
+epoch 0, total_step 1460, total loss is 43.86 , inference loss is 29.49, weight deacy loss is 14.36, training accuracy is 0.062500, time 54.591 samples/sec
+epoch 0, total_step 1480, total loss is 43.90 , inference loss is 29.54, weight deacy loss is 14.36, training accuracy is 0.031250, time 56.506 samples/sec
+epoch 0, total_step 1500, total loss is 48.40 , inference loss is 34.04, weight deacy loss is 14.36, training accuracy is 0.000000, time 56.474 samples/sec
+epoch 0, total_step 1520, total loss is 51.01 , inference loss is 36.65, weight deacy loss is 14.36, training accuracy is 0.000000, time 56.398 samples/sec
+epoch 0, total_step 1540, total loss is 46.51 , inference loss is 32.15, weight deacy loss is 14.36, training accuracy is 0.000000, time 55.367 samples/sec
+epoch 0, total_step 1560, total loss is 46.69 , inference loss is 32.34, weight deacy loss is 14.36, training accuracy is 0.000000, time 57.218 samples/sec
+epoch 0, total_step 1580, total loss is 46.58 , inference loss is 32.23, weight deacy loss is 14.35, training accuracy is 0.000000, time 55.824 samples/sec
+epoch 0, total_step 1600, total loss is 47.70 , inference loss is 33.35, weight deacy loss is 14.35, training accuracy is 0.000000, time 56.118 samples/sec
+epoch 0, total_step 1620, total loss is 45.72 , inference loss is 31.37, weight deacy loss is 14.35, training accuracy is 0.000000, time 56.702 samples/sec
+epoch 0, total_step 1640, total loss is 42.48 , inference loss is 28.13, weight deacy loss is 14.35, training accuracy is 0.093750, time 57.345 samples/sec
+epoch 0, total_step 1660, total loss is 41.84 , inference loss is 27.49, weight deacy loss is 14.35, training accuracy is 0.000000, time 57.415 samples/sec
+epoch 0, total_step 1680, total loss is 45.08 , inference loss is 30.73, weight deacy loss is 14.35, training accuracy is 0.031250, time 54.068 samples/sec
+epoch 0, total_step 1700, total loss is 47.54 , inference loss is 33.20, weight deacy loss is 14.34, training accuracy is 0.000000, time 55.844 samples/sec
+epoch 0, total_step 1720, total loss is 45.82 , inference loss is 31.47, weight deacy loss is 14.34, training accuracy is 0.000000, time 55.514 samples/sec
+epoch 0, total_step 1740, total loss is 39.64 , inference loss is 25.30, weight deacy loss is 14.34, training accuracy is 0.093750, time 56.209 samples/sec
+epoch 0, total_step 1760, total loss is 48.26 , inference loss is 33.92, weight deacy loss is 14.34, training accuracy is 0.031250, time 55.824 samples/sec
+epoch 0, total_step 1780, total loss is 46.75 , inference loss is 32.42, weight deacy loss is 14.34, training accuracy is 0.031250, time 57.303 samples/sec
+epoch 0, total_step 1800, total loss is 40.62 , inference loss is 26.28, weight deacy loss is 14.34, training accuracy is 0.062500, time 55.803 samples/sec
+epoch 0, total_step 1820, total loss is 36.88 , inference loss is 22.55, weight deacy loss is 14.33, training accuracy is 0.000000, time 57.519 samples/sec
+epoch 0, total_step 1840, total loss is 47.11 , inference loss is 32.78, weight deacy loss is 14.33, training accuracy is 0.031250, time 55.633 samples/sec
+epoch 0, total_step 1860, total loss is 40.11 , inference loss is 25.78, weight deacy loss is 14.33, training accuracy is 0.062500, time 56.696 samples/sec
+epoch 0, total_step 1880, total loss is 47.33 , inference loss is 33.00, weight deacy loss is 14.33, training accuracy is 0.000000, time 56.337 samples/sec
+epoch 0, total_step 1900, total loss is 50.77 , inference loss is 36.44, weight deacy loss is 14.33, training accuracy is 0.000000, time 56.012 samples/sec
+epoch 0, total_step 1920, total loss is 46.90 , inference loss is 32.57, weight deacy loss is 14.33, training accuracy is 0.000000, time 56.489 samples/sec
+epoch 0, total_step 1940, total loss is 43.12 , inference loss is 28.79, weight deacy loss is 14.33, training accuracy is 0.031250, time 56.697 samples/sec
+epoch 0, total_step 1960, total loss is 45.30 , inference loss is 30.97, weight deacy loss is 14.32, training accuracy is 0.093750, time 56.410 samples/sec
+epoch 0, total_step 1980, total loss is 47.55 , inference loss is 33.23, weight deacy loss is 14.32, training accuracy is 0.000000, time 56.225 samples/sec
+testing verification..
+(12000, 512)
+infer time 76.44929499999998
+best_threshold_index 137 0.8868518518518519
+best_threshold_index 137 0.8887037037037037
+best_threshold_index 137 0.8870370370370371
+best_threshold_index 137 0.8861111111111111
+best_threshold_index 137 0.8907407407407407
+best_threshold_index 137 0.8857407407407407
+best_threshold_index 137 0.8864814814814815
+best_threshold_index 137 0.8870370370370371
+best_threshold_index 137 0.8864814814814815
+best_threshold_index 137 0.8864814814814815
+[lfw][2000]XNorm: 23.024599
+[lfw][2000]Accuracy-Flip: 0.88717+-0.01267
+testing verification..
+(14000, 512)
+infer time 86.38507599999997
+best_threshold_index 117 0.8625396825396825
+best_threshold_index 117 0.8563492063492063
+best_threshold_index 123 0.8593650793650793
+best_threshold_index 117 0.8588888888888889
+best_threshold_index 117 0.86
+best_threshold_index 117 0.8568253968253968
+best_threshold_index 117 0.8592063492063492
+best_threshold_index 117 0.8587301587301587
+best_threshold_index 117 0.8607936507936508
+best_threshold_index 117 0.8617460317460317
+[cfp_ff][2000]XNorm: 24.346275
+[cfp_ff][2000]Accuracy-Flip: 0.85857+-0.01678
+testing verification..
+(14000, 512)
+infer time 86.35054500000004
+best_threshold_index 161 0.6736507936507936
+best_threshold_index 161 0.6711111111111111
+best_threshold_index 161 0.6677777777777778
+best_threshold_index 161 0.6706349206349206
+best_threshold_index 161 0.6722222222222223
+best_threshold_index 161 0.6660317460317461
+best_threshold_index 161 0.6719047619047619
+best_threshold_index 160 0.6658730158730158
+best_threshold_index 161 0.6733333333333333
+best_threshold_index 161 0.6676190476190477
+[cfp_fp][2000]XNorm: 22.376803
+[cfp_fp][2000]Accuracy-Flip: 0.66929+-0.02434
+testing verification..
+(12000, 512)
+infer time 73.88304800000002
+best_threshold_index 161 0.6687037037037037
+best_threshold_index 161 0.6731481481481482
+best_threshold_index 161 0.6746296296296296
+best_threshold_index 161 0.6744444444444444
+best_threshold_index 161 0.6716666666666666
+best_threshold_index 161 0.6687037037037037
+best_threshold_index 161 0.6705555555555556
+best_threshold_index 161 0.6668518518518518
+best_threshold_index 161 0.6701851851851852
+best_threshold_index 161 0.6711111111111111
+[agedb_30][2000]XNorm: 19.643857
+[agedb_30][2000]Accuracy-Flip: 0.67100+-0.02181
+test accuracy is: 0.8871666666666667
+epoch 0, total_step 2000, total loss is 44.28 , inference loss is 29.96, weight deacy loss is 14.32, training accuracy is 0.093750, time 54.216 samples/sec
+epoch 0, total_step 2020, total loss is 44.37 , inference loss is 30.05, weight deacy loss is 14.32, training accuracy is 0.062500, time 56.289 samples/sec
+epoch 0, total_step 2040, total loss is 42.24 , inference loss is 27.92, weight deacy loss is 14.32, training accuracy is 0.093750, time 56.334 samples/sec
+epoch 0, total_step 2060, total loss is 46.65 , inference loss is 32.33, weight deacy loss is 14.32, training accuracy is 0.062500, time 56.525 samples/sec
+epoch 0, total_step 2080, total loss is 41.29 , inference loss is 26.97, weight deacy loss is 14.31, training accuracy is 0.000000, time 57.446 samples/sec
+epoch 0, total_step 2100, total loss is 40.91 , inference loss is 26.60, weight deacy loss is 14.31, training accuracy is 0.125000, time 55.881 samples/sec
+epoch 0, total_step 2120, total loss is 37.68 , inference loss is 23.37, weight deacy loss is 14.31, training accuracy is 0.156250, time 55.727 samples/sec
+epoch 0, total_step 2140, total loss is 43.45 , inference loss is 29.15, weight deacy loss is 14.31, training accuracy is 0.062500, time 57.020 samples/sec
+epoch 0, total_step 2160, total loss is 44.68 , inference loss is 30.37, weight deacy loss is 14.31, training accuracy is 0.062500, time 56.527 samples/sec
+epoch 0, total_step 2180, total loss is 40.57 , inference loss is 26.26, weight deacy loss is 14.31, training accuracy is 0.125000, time 53.559 samples/sec
+epoch 0, total_step 2200, total loss is 41.40 , inference loss is 27.10, weight deacy loss is 14.30, training accuracy is 0.031250, time 56.348 samples/sec
+epoch 0, total_step 2220, total loss is 42.85 , inference loss is 28.54, weight deacy loss is 14.30, training accuracy is 0.000000, time 56.409 samples/sec
+epoch 0, total_step 2240, total loss is 42.88 , inference loss is 28.58, weight deacy loss is 14.30, training accuracy is 0.000000, time 54.427 samples/sec
+epoch 0, total_step 2260, total loss is 41.45 , inference loss is 27.15, weight deacy loss is 14.30, training accuracy is 0.062500, time 55.611 samples/sec
+epoch 0, total_step 2280, total loss is 42.62 , inference loss is 28.32, weight deacy loss is 14.30, training accuracy is 0.156250, time 55.782 samples/sec
+epoch 0, total_step 2300, total loss is 39.56 , inference loss is 25.26, weight deacy loss is 14.30, training accuracy is 0.031250, time 56.137 samples/sec
+epoch 0, total_step 2320, total loss is 36.61 , inference loss is 22.31, weight deacy loss is 14.29, training accuracy is 0.093750, time 56.791 samples/sec
+epoch 0, total_step 2340, total loss is 41.08 , inference loss is 26.79, weight deacy loss is 14.29, training accuracy is 0.000000, time 53.716 samples/sec
+epoch 0, total_step 2360, total loss is 42.29 , inference loss is 28.00, weight deacy loss is 14.29, training accuracy is 0.000000, time 55.312 samples/sec
+epoch 0, total_step 2380, total loss is 38.22 , inference loss is 23.94, weight deacy loss is 14.29, training accuracy is 0.031250, time 57.556 samples/sec
+epoch 0, total_step 2400, total loss is 39.31 , inference loss is 25.03, weight deacy loss is 14.29, training accuracy is 0.000000, time 56.745 samples/sec
+epoch 0, total_step 2420, total loss is 42.58 , inference loss is 28.29, weight deacy loss is 14.29, training accuracy is 0.062500, time 56.503 samples/sec
+epoch 0, total_step 2440, total loss is 46.93 , inference loss is 32.65, weight deacy loss is 14.28, training accuracy is 0.031250, time 57.007 samples/sec
+epoch 0, total_step 2460, total loss is 39.48 , inference loss is 25.19, weight deacy loss is 14.28, training accuracy is 0.093750, time 56.456 samples/sec
+epoch 0, total_step 2480, total loss is 40.80 , inference loss is 26.52, weight deacy loss is 14.28, training accuracy is 0.031250, time 56.024 samples/sec
+epoch 0, total_step 2500, total loss is 44.02 , inference loss is 29.74, weight deacy loss is 14.28, training accuracy is 0.031250, time 56.348 samples/sec
+epoch 0, total_step 2520, total loss is 41.86 , inference loss is 27.59, weight deacy loss is 14.28, training accuracy is 0.031250, time 56.129 samples/sec
+epoch 0, total_step 2540, total loss is 41.15 , inference loss is 26.87, weight deacy loss is 14.28, training accuracy is 0.031250, time 55.851 samples/sec
+epoch 0, total_step 2560, total loss is 39.83 , inference loss is 25.56, weight deacy loss is 14.27, training accuracy is 0.093750, time 55.929 samples/sec
+epoch 0, total_step 2580, total loss is 47.11 , inference loss is 32.84, weight deacy loss is 14.27, training accuracy is 0.031250, time 56.128 samples/sec
+epoch 0, total_step 2600, total loss is 37.92 , inference loss is 23.64, weight deacy loss is 14.27, training accuracy is 0.062500, time 57.060 samples/sec
+epoch 0, total_step 2620, total loss is 45.19 , inference loss is 30.92, weight deacy loss is 14.27, training accuracy is 0.000000, time 55.656 samples/sec
+epoch 0, total_step 2640, total loss is 39.56 , inference loss is 25.29, weight deacy loss is 14.27, training accuracy is 0.000000, time 56.625 samples/sec
+epoch 0, total_step 2660, total loss is 40.61 , inference loss is 26.35, weight deacy loss is 14.27, training accuracy is 0.031250, time 54.757 samples/sec
+epoch 0, total_step 2680, total loss is 40.39 , inference loss is 26.13, weight deacy loss is 14.26, training accuracy is 0.031250, time 56.885 samples/sec
+epoch 0, total_step 2700, total loss is 40.74 , inference loss is 26.47, weight deacy loss is 14.26, training accuracy is 0.000000, time 56.391 samples/sec
+epoch 0, total_step 2720, total loss is 39.33 , inference loss is 25.07, weight deacy loss is 14.26, training accuracy is 0.062500, time 56.627 samples/sec
+epoch 0, total_step 2740, total loss is 40.59 , inference loss is 26.34, weight deacy loss is 14.26, training accuracy is 0.031250, time 56.455 samples/sec
+epoch 0, total_step 2760, total loss is 43.82 , inference loss is 29.56, weight deacy loss is 14.26, training accuracy is 0.000000, time 56.264 samples/sec
+epoch 0, total_step 2780, total loss is 39.45 , inference loss is 25.19, weight deacy loss is 14.26, training accuracy is 0.000000, time 56.396 samples/sec
+epoch 0, total_step 2800, total loss is 36.45 , inference loss is 22.20, weight deacy loss is 14.25, training accuracy is 0.062500, time 56.767 samples/sec
+epoch 0, total_step 2820, total loss is 40.73 , inference loss is 26.48, weight deacy loss is 14.25, training accuracy is 0.000000, time 56.145 samples/sec
+epoch 0, total_step 2840, total loss is 44.04 , inference loss is 29.79, weight deacy loss is 14.25, training accuracy is 0.000000, time 55.447 samples/sec
+epoch 0, total_step 2860, total loss is 37.26 , inference loss is 23.01, weight deacy loss is 14.25, training accuracy is 0.031250, time 56.933 samples/sec
+epoch 0, total_step 2880, total loss is 40.01 , inference loss is 25.76, weight deacy loss is 14.25, training accuracy is 0.031250, time 56.063 samples/sec
+epoch 0, total_step 2900, total loss is 36.48 , inference loss is 22.24, weight deacy loss is 14.25, training accuracy is 0.062500, time 56.414 samples/sec
+epoch 0, total_step 2920, total loss is 38.51 , inference loss is 24.26, weight deacy loss is 14.25, training accuracy is 0.000000, time 55.361 samples/sec
+epoch 0, total_step 2940, total loss is 34.62 , inference loss is 20.38, weight deacy loss is 14.24, training accuracy is 0.093750, time 53.055 samples/sec
+epoch 0, total_step 2960, total loss is 38.94 , inference loss is 24.70, weight deacy loss is 14.24, training accuracy is 0.062500, time 56.672 samples/sec
+epoch 0, total_step 2980, total loss is 40.97 , inference loss is 26.73, weight deacy loss is 14.24, training accuracy is 0.062500, time 56.945 samples/sec
+epoch 0, total_step 3000, total loss is 36.97 , inference loss is 22.73, weight deacy loss is 14.24, training accuracy is 0.062500, time 56.639 samples/sec
+epoch 0, total_step 3020, total loss is 35.32 , inference loss is 21.09, weight deacy loss is 14.24, training accuracy is 0.062500, time 56.741 samples/sec
+epoch 0, total_step 3040, total loss is 43.38 , inference loss is 29.14, weight deacy loss is 14.24, training accuracy is 0.031250, time 53.729 samples/sec
+epoch 0, total_step 3060, total loss is 41.18 , inference loss is 26.94, weight deacy loss is 14.23, training accuracy is 0.031250, time 56.215 samples/sec
+epoch 0, total_step 3080, total loss is 40.66 , inference loss is 26.42, weight deacy loss is 14.23, training accuracy is 0.031250, time 56.525 samples/sec
+epoch 0, total_step 3100, total loss is 41.20 , inference loss is 26.97, weight deacy loss is 14.23, training accuracy is 0.031250, time 56.467 samples/sec
+epoch 0, total_step 3120, total loss is 36.07 , inference loss is 21.84, weight deacy loss is 14.23, training accuracy is 0.093750, time 55.714 samples/sec
+epoch 0, total_step 3140, total loss is 39.05 , inference loss is 24.83, weight deacy loss is 14.23, training accuracy is 0.156250, time 56.567 samples/sec
+epoch 0, total_step 3160, total loss is 44.96 , inference loss is 30.73, weight deacy loss is 14.23, training accuracy is 0.093750, time 52.928 samples/sec
+epoch 0, total_step 3180, total loss is 41.84 , inference loss is 27.61, weight deacy loss is 14.22, training accuracy is 0.062500, time 56.424 samples/sec
+epoch 0, total_step 3200, total loss is 40.57 , inference loss is 26.35, weight deacy loss is 14.22, training accuracy is 0.093750, time 56.506 samples/sec
+epoch 0, total_step 3220, total loss is 41.62 , inference loss is 27.40, weight deacy loss is 14.22, training accuracy is 0.093750, time 56.658 samples/sec
+epoch 0, total_step 3240, total loss is 38.93 , inference loss is 24.71, weight deacy loss is 14.22, training accuracy is 0.062500, time 56.241 samples/sec
+epoch 0, total_step 3260, total loss is 41.51 , inference loss is 27.29, weight deacy loss is 14.22, training accuracy is 0.062500, time 56.180 samples/sec
+epoch 0, total_step 3280, total loss is 43.38 , inference loss is 29.17, weight deacy loss is 14.22, training accuracy is 0.031250, time 55.865 samples/sec
+epoch 0, total_step 3300, total loss is 43.21 , inference loss is 29.00, weight deacy loss is 14.22, training accuracy is 0.031250, time 54.844 samples/sec
+epoch 0, total_step 3320, total loss is 45.54 , inference loss is 31.33, weight deacy loss is 14.21, training accuracy is 0.062500, time 56.816 samples/sec
+epoch 0, total_step 3340, total loss is 40.29 , inference loss is 26.08, weight deacy loss is 14.21, training accuracy is 0.031250, time 56.963 samples/sec
+epoch 0, total_step 3360, total loss is 41.68 , inference loss is 27.47, weight deacy loss is 14.21, training accuracy is 0.062500, time 55.107 samples/sec
+epoch 0, total_step 3380, total loss is 46.98 , inference loss is 32.77, weight deacy loss is 14.21, training accuracy is 0.031250, time 56.300 samples/sec
+epoch 0, total_step 3400, total loss is 40.87 , inference loss is 26.67, weight deacy loss is 14.21, training accuracy is 0.031250, time 55.255 samples/sec
+epoch 0, total_step 3420, total loss is 43.03 , inference loss is 28.82, weight deacy loss is 14.21, training accuracy is 0.000000, time 57.182 samples/sec
+epoch 0, total_step 3440, total loss is 38.38 , inference loss is 24.18, weight deacy loss is 14.21, training accuracy is 0.062500, time 56.645 samples/sec
+epoch 0, total_step 3460, total loss is 40.72 , inference loss is 26.52, weight deacy loss is 14.20, training accuracy is 0.125000, time 56.800 samples/sec
+epoch 0, total_step 3480, total loss is 42.91 , inference loss is 28.71, weight deacy loss is 14.20, training accuracy is 0.000000, time 55.800 samples/sec
+epoch 0, total_step 3500, total loss is 40.19 , inference loss is 25.99, weight deacy loss is 14.20, training accuracy is 0.031250, time 55.524 samples/sec
+epoch 0, total_step 3520, total loss is 40.93 , inference loss is 26.73, weight deacy loss is 14.20, training accuracy is 0.031250, time 57.469 samples/sec
+epoch 0, total_step 3540, total loss is 39.32 , inference loss is 25.12, weight deacy loss is 14.20, training accuracy is 0.062500, time 56.062 samples/sec
+epoch 0, total_step 3560, total loss is 45.26 , inference loss is 31.06, weight deacy loss is 14.20, training accuracy is 0.031250, time 56.894 samples/sec
+epoch 0, total_step 3580, total loss is 42.14 , inference loss is 27.94, weight deacy loss is 14.19, training accuracy is 0.000000, time 56.281 samples/sec
+epoch 0, total_step 3600, total loss is 40.34 , inference loss is 26.15, weight deacy loss is 14.19, training accuracy is 0.000000, time 55.544 samples/sec
+epoch 0, total_step 3620, total loss is 40.32 , inference loss is 26.12, weight deacy loss is 14.19, training accuracy is 0.000000, time 50.609 samples/sec
+epoch 0, total_step 3640, total loss is 43.10 , inference loss is 28.91, weight deacy loss is 14.19, training accuracy is 0.031250, time 54.969 samples/sec
+epoch 0, total_step 3660, total loss is 44.04 , inference loss is 29.85, weight deacy loss is 14.19, training accuracy is 0.062500, time 56.995 samples/sec
+epoch 0, total_step 3680, total loss is 44.87 , inference loss is 30.68, weight deacy loss is 14.19, training accuracy is 0.062500, time 54.744 samples/sec
+epoch 0, total_step 3700, total loss is 36.78 , inference loss is 22.59, weight deacy loss is 14.19, training accuracy is 0.031250, time 52.689 samples/sec
+epoch 0, total_step 3720, total loss is 42.02 , inference loss is 27.83, weight deacy loss is 14.18, training accuracy is 0.000000, time 56.860 samples/sec
+epoch 0, total_step 3740, total loss is 39.70 , inference loss is 25.52, weight deacy loss is 14.18, training accuracy is 0.062500, time 56.594 samples/sec
+epoch 0, total_step 3760, total loss is 38.54 , inference loss is 24.35, weight deacy loss is 14.18, training accuracy is 0.031250, time 56.025 samples/sec
+epoch 0, total_step 3780, total loss is 41.42 , inference loss is 27.24, weight deacy loss is 14.18, training accuracy is 0.000000, time 56.267 samples/sec
+epoch 0, total_step 3800, total loss is 47.06 , inference loss is 32.88, weight deacy loss is 14.18, training accuracy is 0.000000, time 56.171 samples/sec
+epoch 0, total_step 3820, total loss is 40.81 , inference loss is 26.63, weight deacy loss is 14.18, training accuracy is 0.000000, time 56.642 samples/sec
+epoch 0, total_step 3840, total loss is 42.65 , inference loss is 28.47, weight deacy loss is 14.18, training accuracy is 0.062500, time 56.917 samples/sec
+epoch 0, total_step 3860, total loss is 46.09 , inference loss is 31.91, weight deacy loss is 14.17, training accuracy is 0.000000, time 57.180 samples/sec
+epoch 0, total_step 3880, total loss is 43.08 , inference loss is 28.90, weight deacy loss is 14.17, training accuracy is 0.000000, time 55.112 samples/sec
+epoch 0, total_step 3900, total loss is 43.47 , inference loss is 29.30, weight deacy loss is 14.17, training accuracy is 0.031250, time 55.659 samples/sec
+epoch 0, total_step 3920, total loss is 45.22 , inference loss is 31.06, weight deacy loss is 14.17, training accuracy is 0.000000, time 54.661 samples/sec
+epoch 0, total_step 3940, total loss is 46.13 , inference loss is 31.97, weight deacy loss is 14.17, training accuracy is 0.000000, time 57.101 samples/sec
+epoch 0, total_step 3960, total loss is 43.52 , inference loss is 29.35, weight deacy loss is 14.17, training accuracy is 0.000000, time 56.946 samples/sec
+epoch 0, total_step 3980, total loss is 42.86 , inference loss is 28.70, weight deacy loss is 14.16, training accuracy is 0.062500, time 56.212 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.90206299999991
+best_threshold_index 132 0.9083333333333333
+best_threshold_index 132 0.909074074074074
+best_threshold_index 132 0.9092592592592592
+best_threshold_index 132 0.9077777777777778
+best_threshold_index 132 0.9125925925925926
+best_threshold_index 132 0.907962962962963
+best_threshold_index 132 0.9057407407407407
+best_threshold_index 132 0.9092592592592592
+best_threshold_index 132 0.9068518518518518
+best_threshold_index 132 0.9081481481481481
+[lfw][4000]XNorm: 23.709450
+[lfw][4000]Accuracy-Flip: 0.90850+-0.01546
+testing verification..
+(14000, 512)
+infer time 86.30891600000012
+best_threshold_index 126 0.9047619047619048
+best_threshold_index 125 0.8998412698412699
+best_threshold_index 125 0.9025396825396825
+best_threshold_index 125 0.902063492063492
+best_threshold_index 128 0.9014285714285715
+best_threshold_index 128 0.900952380952381
+best_threshold_index 128 0.903968253968254
+best_threshold_index 125 0.9011111111111111
+best_threshold_index 128 0.9017460317460317
+best_threshold_index 125 0.9034920634920635
+[cfp_ff][4000]XNorm: 23.354800
+[cfp_ff][4000]Accuracy-Flip: 0.90014+-0.01318
+testing verification..
+(14000, 512)
+infer time 86.99717900000006
+best_threshold_index 160 0.7287301587301588
+best_threshold_index 160 0.7282539682539683
+best_threshold_index 158 0.7234920634920635
+best_threshold_index 160 0.7236507936507937
+best_threshold_index 158 0.7271428571428571
+best_threshold_index 158 0.7236507936507937
+best_threshold_index 160 0.7247619047619047
+best_threshold_index 158 0.7241269841269842
+best_threshold_index 160 0.7247619047619047
+best_threshold_index 160 0.7231746031746031
+[cfp_fp][4000]XNorm: 21.126693
+[cfp_fp][4000]Accuracy-Flip: 0.72286+-0.01760
+testing verification..
+(12000, 512)
+infer time 74.83907600000006
+best_threshold_index 155 0.7337037037037037
+best_threshold_index 158 0.7344444444444445
+best_threshold_index 158 0.732037037037037
+best_threshold_index 157 0.7357407407407407
+best_threshold_index 157 0.74
+best_threshold_index 157 0.7359259259259259
+best_threshold_index 155 0.7368518518518519
+best_threshold_index 158 0.7379629629629629
+best_threshold_index 159 0.7368518518518519
+best_threshold_index 158 0.7346296296296296
+[agedb_30][4000]XNorm: 20.727606
+[agedb_30][4000]Accuracy-Flip: 0.73167+-0.01942
+test accuracy is: 0.9085000000000001
+epoch 0, total_step 4000, total loss is 43.73 , inference loss is 29.57, weight deacy loss is 14.16, training accuracy is 0.000000, time 53.918 samples/sec
+epoch 0, total_step 4020, total loss is 45.65 , inference loss is 31.49, weight deacy loss is 14.16, training accuracy is 0.062500, time 56.283 samples/sec
+epoch 0, total_step 4040, total loss is 40.65 , inference loss is 26.49, weight deacy loss is 14.16, training accuracy is 0.000000, time 56.465 samples/sec
+epoch 0, total_step 4060, total loss is 47.50 , inference loss is 33.34, weight deacy loss is 14.16, training accuracy is 0.000000, time 56.203 samples/sec
+epoch 0, total_step 4080, total loss is 42.40 , inference loss is 28.24, weight deacy loss is 14.16, training accuracy is 0.062500, time 57.506 samples/sec
+epoch 0, total_step 4100, total loss is 36.87 , inference loss is 22.71, weight deacy loss is 14.16, training accuracy is 0.031250, time 55.585 samples/sec
+epoch 0, total_step 4120, total loss is 41.52 , inference loss is 27.37, weight deacy loss is 14.15, training accuracy is 0.062500, time 55.442 samples/sec
+epoch 0, total_step 4140, total loss is 39.20 , inference loss is 25.05, weight deacy loss is 14.15, training accuracy is 0.031250, time 56.308 samples/sec
+epoch 0, total_step 4160, total loss is 42.90 , inference loss is 28.75, weight deacy loss is 14.15, training accuracy is 0.000000, time 56.705 samples/sec
+epoch 0, total_step 4180, total loss is 40.47 , inference loss is 26.32, weight deacy loss is 14.15, training accuracy is 0.000000, time 56.058 samples/sec
+epoch 0, total_step 4200, total loss is 35.68 , inference loss is 21.53, weight deacy loss is 14.15, training accuracy is 0.062500, time 55.605 samples/sec
+epoch 0, total_step 4220, total loss is 38.15 , inference loss is 24.01, weight deacy loss is 14.15, training accuracy is 0.000000, time 54.038 samples/sec
+epoch 0, total_step 4240, total loss is 39.15 , inference loss is 25.01, weight deacy loss is 14.14, training accuracy is 0.000000, time 55.444 samples/sec
+epoch 0, total_step 4260, total loss is 47.28 , inference loss is 33.14, weight deacy loss is 14.14, training accuracy is 0.062500, time 56.562 samples/sec
+epoch 0, total_step 4280, total loss is 42.33 , inference loss is 28.19, weight deacy loss is 14.14, training accuracy is 0.000000, time 57.072 samples/sec
+epoch 0, total_step 4300, total loss is 45.80 , inference loss is 31.66, weight deacy loss is 14.14, training accuracy is 0.031250, time 56.638 samples/sec
+epoch 0, total_step 4320, total loss is 41.09 , inference loss is 26.96, weight deacy loss is 14.14, training accuracy is 0.031250, time 55.597 samples/sec
+epoch 0, total_step 4340, total loss is 41.25 , inference loss is 27.11, weight deacy loss is 14.14, training accuracy is 0.000000, time 56.103 samples/sec
+epoch 0, total_step 4360, total loss is 44.03 , inference loss is 29.90, weight deacy loss is 14.14, training accuracy is 0.031250, time 55.798 samples/sec
+epoch 0, total_step 4380, total loss is 40.54 , inference loss is 26.41, weight deacy loss is 14.13, training accuracy is 0.031250, time 55.767 samples/sec
+epoch 0, total_step 4400, total loss is 48.24 , inference loss is 34.10, weight deacy loss is 14.13, training accuracy is 0.062500, time 56.812 samples/sec
+epoch 0, total_step 4420, total loss is 44.24 , inference loss is 30.11, weight deacy loss is 14.13, training accuracy is 0.031250, time 56.761 samples/sec
+epoch 0, total_step 4440, total loss is 38.30 , inference loss is 24.17, weight deacy loss is 14.13, training accuracy is 0.031250, time 56.883 samples/sec
+epoch 0, total_step 4460, total loss is 36.87 , inference loss is 22.74, weight deacy loss is 14.13, training accuracy is 0.093750, time 56.451 samples/sec
+epoch 0, total_step 4480, total loss is 43.56 , inference loss is 29.43, weight deacy loss is 14.13, training accuracy is 0.062500, time 56.128 samples/sec
+epoch 0, total_step 4500, total loss is 41.46 , inference loss is 27.33, weight deacy loss is 14.12, training accuracy is 0.031250, time 56.052 samples/sec
+epoch 0, total_step 4520, total loss is 46.38 , inference loss is 32.26, weight deacy loss is 14.12, training accuracy is 0.000000, time 56.607 samples/sec
+epoch 0, total_step 4540, total loss is 40.44 , inference loss is 26.32, weight deacy loss is 14.12, training accuracy is 0.031250, time 55.440 samples/sec
+epoch 0, total_step 4560, total loss is 43.57 , inference loss is 29.45, weight deacy loss is 14.12, training accuracy is 0.031250, time 56.423 samples/sec
+epoch 0, total_step 4580, total loss is 41.63 , inference loss is 27.51, weight deacy loss is 14.12, training accuracy is 0.031250, time 55.463 samples/sec
+epoch 0, total_step 4600, total loss is 40.61 , inference loss is 26.50, weight deacy loss is 14.12, training accuracy is 0.031250, time 56.412 samples/sec
+epoch 0, total_step 4620, total loss is 45.34 , inference loss is 31.23, weight deacy loss is 14.12, training accuracy is 0.000000, time 56.467 samples/sec
+epoch 0, total_step 4640, total loss is 44.18 , inference loss is 30.06, weight deacy loss is 14.11, training accuracy is 0.031250, time 56.274 samples/sec
+epoch 0, total_step 4660, total loss is 39.67 , inference loss is 25.56, weight deacy loss is 14.11, training accuracy is 0.000000, time 55.187 samples/sec
+epoch 0, total_step 4680, total loss is 43.14 , inference loss is 29.03, weight deacy loss is 14.11, training accuracy is 0.031250, time 56.017 samples/sec
+epoch 0, total_step 4700, total loss is 45.12 , inference loss is 31.01, weight deacy loss is 14.11, training accuracy is 0.031250, time 56.714 samples/sec
+epoch 0, total_step 4720, total loss is 42.68 , inference loss is 28.57, weight deacy loss is 14.11, training accuracy is 0.000000, time 55.461 samples/sec
+epoch 0, total_step 4740, total loss is 45.52 , inference loss is 31.41, weight deacy loss is 14.11, training accuracy is 0.000000, time 55.885 samples/sec
+epoch 0, total_step 4760, total loss is 39.96 , inference loss is 25.86, weight deacy loss is 14.10, training accuracy is 0.031250, time 56.000 samples/sec
+epoch 0, total_step 4780, total loss is 39.65 , inference loss is 25.55, weight deacy loss is 14.10, training accuracy is 0.031250, time 56.588 samples/sec
+epoch 0, total_step 4800, total loss is 42.61 , inference loss is 28.51, weight deacy loss is 14.10, training accuracy is 0.031250, time 55.874 samples/sec
+epoch 0, total_step 4820, total loss is 38.81 , inference loss is 24.71, weight deacy loss is 14.10, training accuracy is 0.031250, time 56.543 samples/sec
+epoch 0, total_step 4840, total loss is 41.73 , inference loss is 27.63, weight deacy loss is 14.10, training accuracy is 0.062500, time 56.521 samples/sec
+epoch 0, total_step 4860, total loss is 37.44 , inference loss is 23.34, weight deacy loss is 14.10, training accuracy is 0.000000, time 54.479 samples/sec
+epoch 0, total_step 4880, total loss is 36.29 , inference loss is 22.19, weight deacy loss is 14.10, training accuracy is 0.000000, time 56.070 samples/sec
+epoch 0, total_step 4900, total loss is 43.58 , inference loss is 29.49, weight deacy loss is 14.09, training accuracy is 0.000000, time 55.674 samples/sec
+epoch 0, total_step 4920, total loss is 47.62 , inference loss is 33.53, weight deacy loss is 14.09, training accuracy is 0.031250, time 56.641 samples/sec
+epoch 0, total_step 4940, total loss is 42.58 , inference loss is 28.49, weight deacy loss is 14.09, training accuracy is 0.000000, time 55.935 samples/sec
+epoch 0, total_step 4960, total loss is 44.85 , inference loss is 30.76, weight deacy loss is 14.09, training accuracy is 0.000000, time 56.162 samples/sec
+epoch 0, total_step 4980, total loss is 46.41 , inference loss is 32.32, weight deacy loss is 14.09, training accuracy is 0.031250, time 56.491 samples/sec
+epoch 0, total_step 5000, total loss is 42.46 , inference loss is 28.38, weight deacy loss is 14.09, training accuracy is 0.031250, time 56.230 samples/sec
+epoch 0, total_step 5020, total loss is 40.18 , inference loss is 26.10, weight deacy loss is 14.09, training accuracy is 0.062500, time 55.927 samples/sec
+epoch 0, total_step 5040, total loss is 40.15 , inference loss is 26.06, weight deacy loss is 14.08, training accuracy is 0.000000, time 56.353 samples/sec
+epoch 0, total_step 5060, total loss is 40.01 , inference loss is 25.92, weight deacy loss is 14.08, training accuracy is 0.000000, time 56.476 samples/sec
+epoch 0, total_step 5080, total loss is 36.13 , inference loss is 22.05, weight deacy loss is 14.08, training accuracy is 0.031250, time 56.511 samples/sec
+epoch 0, total_step 5100, total loss is 39.98 , inference loss is 25.90, weight deacy loss is 14.08, training accuracy is 0.000000, time 55.424 samples/sec
+epoch 0, total_step 5120, total loss is 45.56 , inference loss is 31.49, weight deacy loss is 14.08, training accuracy is 0.031250, time 56.533 samples/sec
+epoch 0, total_step 5140, total loss is 39.51 , inference loss is 25.44, weight deacy loss is 14.08, training accuracy is 0.062500, time 56.276 samples/sec
+epoch 0, total_step 5160, total loss is 37.93 , inference loss is 23.85, weight deacy loss is 14.07, training accuracy is 0.062500, time 53.317 samples/sec
+epoch 0, total_step 5180, total loss is 46.95 , inference loss is 32.88, weight deacy loss is 14.07, training accuracy is 0.031250, time 55.929 samples/sec
+epoch 0, total_step 5200, total loss is 39.33 , inference loss is 25.26, weight deacy loss is 14.07, training accuracy is 0.031250, time 54.129 samples/sec
+epoch 0, total_step 5220, total loss is 44.83 , inference loss is 30.76, weight deacy loss is 14.07, training accuracy is 0.000000, time 55.720 samples/sec
+epoch 0, total_step 5240, total loss is 44.54 , inference loss is 30.48, weight deacy loss is 14.07, training accuracy is 0.031250, time 57.227 samples/sec
+epoch 0, total_step 5260, total loss is 45.41 , inference loss is 31.34, weight deacy loss is 14.07, training accuracy is 0.000000, time 56.505 samples/sec
+epoch 0, total_step 5280, total loss is 43.65 , inference loss is 29.58, weight deacy loss is 14.06, training accuracy is 0.062500, time 55.059 samples/sec
+epoch 0, total_step 5300, total loss is 40.64 , inference loss is 26.58, weight deacy loss is 14.06, training accuracy is 0.000000, time 55.587 samples/sec
+epoch 0, total_step 5320, total loss is 43.91 , inference loss is 29.85, weight deacy loss is 14.06, training accuracy is 0.000000, time 57.596 samples/sec
+epoch 0, total_step 5340, total loss is 39.13 , inference loss is 25.07, weight deacy loss is 14.06, training accuracy is 0.062500, time 56.619 samples/sec
+epoch 0, total_step 5360, total loss is 39.41 , inference loss is 25.35, weight deacy loss is 14.06, training accuracy is 0.031250, time 56.022 samples/sec
+epoch 0, total_step 5380, total loss is 38.17 , inference loss is 24.12, weight deacy loss is 14.06, training accuracy is 0.031250, time 55.684 samples/sec
+epoch 0, total_step 5400, total loss is 39.95 , inference loss is 25.90, weight deacy loss is 14.06, training accuracy is 0.031250, time 56.072 samples/sec
+epoch 0, total_step 5420, total loss is 37.30 , inference loss is 23.25, weight deacy loss is 14.05, training accuracy is 0.093750, time 55.878 samples/sec
+epoch 0, total_step 5440, total loss is 42.35 , inference loss is 28.30, weight deacy loss is 14.05, training accuracy is 0.000000, time 56.131 samples/sec
+epoch 0, total_step 5460, total loss is 44.08 , inference loss is 30.03, weight deacy loss is 14.05, training accuracy is 0.031250, time 56.677 samples/sec
+epoch 0, total_step 5480, total loss is 41.12 , inference loss is 27.07, weight deacy loss is 14.05, training accuracy is 0.000000, time 56.116 samples/sec
+epoch 0, total_step 5500, total loss is 43.50 , inference loss is 29.45, weight deacy loss is 14.05, training accuracy is 0.031250, time 56.642 samples/sec
+epoch 0, total_step 5520, total loss is 40.01 , inference loss is 25.96, weight deacy loss is 14.05, training accuracy is 0.031250, time 56.358 samples/sec
+epoch 0, total_step 5540, total loss is 44.69 , inference loss is 30.64, weight deacy loss is 14.04, training accuracy is 0.062500, time 56.774 samples/sec
+epoch 0, total_step 5560, total loss is 40.23 , inference loss is 26.18, weight deacy loss is 14.04, training accuracy is 0.062500, time 55.602 samples/sec
+epoch 0, total_step 5580, total loss is 39.49 , inference loss is 25.45, weight deacy loss is 14.04, training accuracy is 0.031250, time 56.761 samples/sec
+epoch 0, total_step 5600, total loss is 40.57 , inference loss is 26.53, weight deacy loss is 14.04, training accuracy is 0.031250, time 55.974 samples/sec
+epoch 0, total_step 5620, total loss is 42.16 , inference loss is 28.12, weight deacy loss is 14.04, training accuracy is 0.031250, time 56.626 samples/sec
+epoch 0, total_step 5640, total loss is 37.93 , inference loss is 23.89, weight deacy loss is 14.04, training accuracy is 0.093750, time 56.475 samples/sec
+epoch 0, total_step 5660, total loss is 45.22 , inference loss is 31.18, weight deacy loss is 14.04, training accuracy is 0.031250, time 56.470 samples/sec
+epoch 0, total_step 5680, total loss is 38.07 , inference loss is 24.04, weight deacy loss is 14.03, training accuracy is 0.062500, time 55.118 samples/sec
+epoch 0, total_step 5700, total loss is 38.60 , inference loss is 24.56, weight deacy loss is 14.03, training accuracy is 0.000000, time 55.562 samples/sec
+epoch 0, total_step 5720, total loss is 38.81 , inference loss is 24.78, weight deacy loss is 14.03, training accuracy is 0.062500, time 56.523 samples/sec
+epoch 0, total_step 5740, total loss is 39.41 , inference loss is 25.38, weight deacy loss is 14.03, training accuracy is 0.031250, time 56.759 samples/sec
+epoch 0, total_step 5760, total loss is 37.99 , inference loss is 23.96, weight deacy loss is 14.03, training accuracy is 0.031250, time 56.783 samples/sec
+epoch 0, total_step 5780, total loss is 43.13 , inference loss is 29.11, weight deacy loss is 14.03, training accuracy is 0.031250, time 55.147 samples/sec
+epoch 0, total_step 5800, total loss is 39.14 , inference loss is 25.12, weight deacy loss is 14.03, training accuracy is 0.031250, time 56.802 samples/sec
+epoch 0, total_step 5820, total loss is 39.98 , inference loss is 25.96, weight deacy loss is 14.02, training accuracy is 0.093750, time 56.694 samples/sec
+epoch 0, total_step 5840, total loss is 38.36 , inference loss is 24.34, weight deacy loss is 14.02, training accuracy is 0.031250, time 55.867 samples/sec
+epoch 0, total_step 5860, total loss is 40.52 , inference loss is 26.50, weight deacy loss is 14.02, training accuracy is 0.031250, time 54.522 samples/sec
+epoch 0, total_step 5880, total loss is 39.31 , inference loss is 25.29, weight deacy loss is 14.02, training accuracy is 0.062500, time 53.952 samples/sec
+epoch 0, total_step 5900, total loss is 40.77 , inference loss is 26.76, weight deacy loss is 14.02, training accuracy is 0.093750, time 56.248 samples/sec
+epoch 0, total_step 5920, total loss is 38.70 , inference loss is 24.69, weight deacy loss is 14.02, training accuracy is 0.031250, time 55.113 samples/sec
+epoch 0, total_step 5940, total loss is 40.93 , inference loss is 26.92, weight deacy loss is 14.01, training accuracy is 0.031250, time 55.113 samples/sec
+epoch 0, total_step 5960, total loss is 44.79 , inference loss is 30.78, weight deacy loss is 14.01, training accuracy is 0.000000, time 54.911 samples/sec
+epoch 0, total_step 5980, total loss is 47.41 , inference loss is 33.40, weight deacy loss is 14.01, training accuracy is 0.031250, time 54.481 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.74341299999999
+best_threshold_index 122 0.9248148148148149
+best_threshold_index 122 0.9255555555555556
+best_threshold_index 122 0.9262962962962963
+best_threshold_index 122 0.9266666666666666
+best_threshold_index 122 0.9303703703703704
+best_threshold_index 122 0.9244444444444444
+best_threshold_index 122 0.9246296296296296
+best_threshold_index 122 0.9253703703703704
+best_threshold_index 122 0.9255555555555556
+best_threshold_index 122 0.9246296296296296
+[lfw][6000]XNorm: 23.366604
+[lfw][6000]Accuracy-Flip: 0.92583+-0.01499
+testing verification..
+(14000, 512)
+infer time 87.83107699999998
+best_threshold_index 121 0.9273015873015873
+best_threshold_index 122 0.9241269841269841
+best_threshold_index 125 0.9253968253968254
+best_threshold_index 125 0.923968253968254
+best_threshold_index 122 0.9244444444444444
+best_threshold_index 121 0.9231746031746032
+best_threshold_index 121 0.9253968253968254
+best_threshold_index 122 0.9236507936507936
+best_threshold_index 120 0.9247619047619048
+best_threshold_index 121 0.9261904761904762
+[cfp_ff][6000]XNorm: 22.268322
+[cfp_ff][6000]Accuracy-Flip: 0.92271+-0.01068
+testing verification..
+(14000, 512)
+infer time 87.0244400000001
+best_threshold_index 150 0.7738095238095238
+best_threshold_index 149 0.7736507936507937
+best_threshold_index 150 0.7682539682539683
+best_threshold_index 150 0.7707936507936508
+best_threshold_index 149 0.773968253968254
+best_threshold_index 149 0.7647619047619048
+best_threshold_index 150 0.7674603174603175
+best_threshold_index 149 0.7685714285714286
+best_threshold_index 149 0.7719047619047619
+best_threshold_index 150 0.7704761904761904
+[cfp_fp][6000]XNorm: 20.026400
+[cfp_fp][6000]Accuracy-Flip: 0.76886+-0.02641
+testing verification..
+(12000, 512)
+infer time 74.72280300000011
+best_threshold_index 139 0.7461111111111111
+best_threshold_index 138 0.7479629629629629
+best_threshold_index 139 0.7507407407407407
+best_threshold_index 139 0.7492592592592593
+best_threshold_index 139 0.7509259259259259
+best_threshold_index 139 0.7479629629629629
+best_threshold_index 139 0.7505555555555555
+best_threshold_index 139 0.7487037037037036
+best_threshold_index 139 0.7422222222222222
+best_threshold_index 139 0.7490740740740741
+[agedb_30][6000]XNorm: 20.896915
+[agedb_30][6000]Accuracy-Flip: 0.74750+-0.02231
+test accuracy is: 0.9258333333333333
+epoch 0, total_step 6000, total loss is 40.99 , inference loss is 26.98, weight deacy loss is 14.01, training accuracy is 0.000000, time 54.265 samples/sec
+epoch 0, total_step 6020, total loss is 42.43 , inference loss is 28.42, weight deacy loss is 14.01, training accuracy is 0.031250, time 56.482 samples/sec
+epoch 0, total_step 6040, total loss is 36.52 , inference loss is 22.51, weight deacy loss is 14.01, training accuracy is 0.000000, time 56.205 samples/sec
+epoch 0, total_step 6060, total loss is 39.78 , inference loss is 25.78, weight deacy loss is 14.01, training accuracy is 0.062500, time 55.158 samples/sec
+epoch 0, total_step 6080, total loss is 41.38 , inference loss is 27.37, weight deacy loss is 14.00, training accuracy is 0.000000, time 56.483 samples/sec
+epoch 0, total_step 6100, total loss is 41.64 , inference loss is 27.64, weight deacy loss is 14.00, training accuracy is 0.000000, time 52.215 samples/sec
+epoch 0, total_step 6120, total loss is 36.01 , inference loss is 22.00, weight deacy loss is 14.00, training accuracy is 0.093750, time 56.776 samples/sec
+epoch 0, total_step 6140, total loss is 38.29 , inference loss is 24.29, weight deacy loss is 14.00, training accuracy is 0.062500, time 56.254 samples/sec
+epoch 0, total_step 6160, total loss is 40.76 , inference loss is 26.76, weight deacy loss is 14.00, training accuracy is 0.000000, time 56.381 samples/sec
+epoch 0, total_step 6180, total loss is 43.27 , inference loss is 29.27, weight deacy loss is 14.00, training accuracy is 0.031250, time 55.984 samples/sec
+epoch 0, total_step 6200, total loss is 41.33 , inference loss is 27.33, weight deacy loss is 14.00, training accuracy is 0.031250, time 55.372 samples/sec
+epoch 0, total_step 6220, total loss is 39.87 , inference loss is 25.88, weight deacy loss is 13.99, training accuracy is 0.093750, time 55.954 samples/sec
+epoch 0, total_step 6240, total loss is 40.13 , inference loss is 26.14, weight deacy loss is 13.99, training accuracy is 0.093750, time 55.849 samples/sec
+epoch 0, total_step 6260, total loss is 39.52 , inference loss is 25.53, weight deacy loss is 13.99, training accuracy is 0.062500, time 54.001 samples/sec
+epoch 0, total_step 6280, total loss is 37.74 , inference loss is 23.75, weight deacy loss is 13.99, training accuracy is 0.125000, time 56.938 samples/sec
+epoch 0, total_step 6300, total loss is 37.99 , inference loss is 24.00, weight deacy loss is 13.99, training accuracy is 0.062500, time 56.719 samples/sec
+epoch 0, total_step 6320, total loss is 45.06 , inference loss is 31.07, weight deacy loss is 13.99, training accuracy is 0.000000, time 53.407 samples/sec
+epoch 0, total_step 6340, total loss is 41.53 , inference loss is 27.54, weight deacy loss is 13.98, training accuracy is 0.031250, time 56.089 samples/sec
+epoch 0, total_step 6360, total loss is 43.41 , inference loss is 29.43, weight deacy loss is 13.98, training accuracy is 0.062500, time 56.410 samples/sec
+epoch 0, total_step 6380, total loss is 44.37 , inference loss is 30.39, weight deacy loss is 13.98, training accuracy is 0.031250, time 56.115 samples/sec
+epoch 0, total_step 6400, total loss is 47.05 , inference loss is 33.07, weight deacy loss is 13.98, training accuracy is 0.031250, time 56.276 samples/sec
+epoch 0, total_step 6420, total loss is 40.13 , inference loss is 26.15, weight deacy loss is 13.98, training accuracy is 0.062500, time 56.209 samples/sec
+epoch 0, total_step 6440, total loss is 44.37 , inference loss is 30.39, weight deacy loss is 13.98, training accuracy is 0.031250, time 55.938 samples/sec
+epoch 0, total_step 6460, total loss is 39.83 , inference loss is 25.85, weight deacy loss is 13.98, training accuracy is 0.031250, time 54.150 samples/sec
+epoch 0, total_step 6480, total loss is 38.49 , inference loss is 24.52, weight deacy loss is 13.97, training accuracy is 0.031250, time 56.575 samples/sec
+epoch 0, total_step 6500, total loss is 38.05 , inference loss is 24.08, weight deacy loss is 13.97, training accuracy is 0.031250, time 57.158 samples/sec
+epoch 0, total_step 6520, total loss is 41.31 , inference loss is 27.34, weight deacy loss is 13.97, training accuracy is 0.062500, time 55.975 samples/sec
+epoch 0, total_step 6540, total loss is 38.21 , inference loss is 24.24, weight deacy loss is 13.97, training accuracy is 0.031250, time 56.583 samples/sec
+epoch 0, total_step 6560, total loss is 44.84 , inference loss is 30.87, weight deacy loss is 13.97, training accuracy is 0.000000, time 56.174 samples/sec
+epoch 0, total_step 6580, total loss is 39.23 , inference loss is 25.27, weight deacy loss is 13.97, training accuracy is 0.000000, time 55.008 samples/sec
+epoch 0, total_step 6600, total loss is 43.81 , inference loss is 29.84, weight deacy loss is 13.97, training accuracy is 0.000000, time 56.193 samples/sec
+epoch 0, total_step 6620, total loss is 43.87 , inference loss is 29.90, weight deacy loss is 13.96, training accuracy is 0.062500, time 56.773 samples/sec
+epoch 0, total_step 6640, total loss is 39.34 , inference loss is 25.38, weight deacy loss is 13.96, training accuracy is 0.000000, time 55.907 samples/sec
+epoch 0, total_step 6660, total loss is 39.47 , inference loss is 25.51, weight deacy loss is 13.96, training accuracy is 0.000000, time 56.066 samples/sec
+epoch 0, total_step 6680, total loss is 36.30 , inference loss is 22.34, weight deacy loss is 13.96, training accuracy is 0.062500, time 56.068 samples/sec
+epoch 0, total_step 6700, total loss is 37.41 , inference loss is 23.45, weight deacy loss is 13.96, training accuracy is 0.031250, time 56.555 samples/sec
+epoch 0, total_step 6720, total loss is 36.11 , inference loss is 22.15, weight deacy loss is 13.96, training accuracy is 0.093750, time 56.040 samples/sec
+epoch 0, total_step 6740, total loss is 42.52 , inference loss is 28.56, weight deacy loss is 13.95, training accuracy is 0.031250, time 54.264 samples/sec
+epoch 0, total_step 6760, total loss is 35.95 , inference loss is 21.99, weight deacy loss is 13.95, training accuracy is 0.125000, time 56.209 samples/sec
+epoch 0, total_step 6780, total loss is 40.93 , inference loss is 26.98, weight deacy loss is 13.95, training accuracy is 0.031250, time 56.253 samples/sec
+epoch 0, total_step 6800, total loss is 38.65 , inference loss is 24.70, weight deacy loss is 13.95, training accuracy is 0.187500, time 57.363 samples/sec
+epoch 0, total_step 6820, total loss is 39.68 , inference loss is 25.74, weight deacy loss is 13.95, training accuracy is 0.031250, time 56.653 samples/sec
+epoch 0, total_step 6840, total loss is 42.47 , inference loss is 28.53, weight deacy loss is 13.95, training accuracy is 0.000000, time 55.935 samples/sec
+epoch 0, total_step 6860, total loss is 42.02 , inference loss is 28.07, weight deacy loss is 13.95, training accuracy is 0.000000, time 55.054 samples/sec
+epoch 0, total_step 6880, total loss is 40.07 , inference loss is 26.12, weight deacy loss is 13.94, training accuracy is 0.031250, time 56.280 samples/sec
+epoch 0, total_step 6900, total loss is 48.62 , inference loss is 34.68, weight deacy loss is 13.94, training accuracy is 0.000000, time 55.766 samples/sec
+epoch 0, total_step 6920, total loss is 36.90 , inference loss is 22.96, weight deacy loss is 13.94, training accuracy is 0.031250, time 57.472 samples/sec
+epoch 0, total_step 6940, total loss is 42.13 , inference loss is 28.20, weight deacy loss is 13.94, training accuracy is 0.000000, time 56.302 samples/sec
+epoch 0, total_step 6960, total loss is 47.15 , inference loss is 33.21, weight deacy loss is 13.94, training accuracy is 0.000000, time 56.474 samples/sec
+epoch 0, total_step 6980, total loss is 42.47 , inference loss is 28.53, weight deacy loss is 13.94, training accuracy is 0.000000, time 56.493 samples/sec
+epoch 0, total_step 7000, total loss is 39.58 , inference loss is 25.65, weight deacy loss is 13.94, training accuracy is 0.000000, time 56.227 samples/sec
+epoch 0, total_step 7020, total loss is 46.35 , inference loss is 32.42, weight deacy loss is 13.93, training accuracy is 0.000000, time 55.973 samples/sec
+epoch 0, total_step 7040, total loss is 44.07 , inference loss is 30.13, weight deacy loss is 13.93, training accuracy is 0.031250, time 56.587 samples/sec
+epoch 0, total_step 7060, total loss is 38.13 , inference loss is 24.20, weight deacy loss is 13.93, training accuracy is 0.000000, time 56.342 samples/sec
+epoch 0, total_step 7080, total loss is 40.13 , inference loss is 26.21, weight deacy loss is 13.93, training accuracy is 0.062500, time 55.673 samples/sec
+epoch 0, total_step 7100, total loss is 39.42 , inference loss is 25.49, weight deacy loss is 13.93, training accuracy is 0.000000, time 56.611 samples/sec
+epoch 0, total_step 7120, total loss is 48.25 , inference loss is 34.32, weight deacy loss is 13.93, training accuracy is 0.031250, time 55.773 samples/sec
+epoch 0, total_step 7140, total loss is 41.75 , inference loss is 27.82, weight deacy loss is 13.92, training accuracy is 0.031250, time 55.508 samples/sec
+epoch 0, total_step 7160, total loss is 38.21 , inference loss is 24.29, weight deacy loss is 13.92, training accuracy is 0.031250, time 56.386 samples/sec
+epoch 0, total_step 7180, total loss is 38.58 , inference loss is 24.66, weight deacy loss is 13.92, training accuracy is 0.000000, time 57.104 samples/sec
+epoch 0, total_step 7200, total loss is 41.17 , inference loss is 27.25, weight deacy loss is 13.92, training accuracy is 0.062500, time 55.715 samples/sec
+epoch 0, total_step 7220, total loss is 36.95 , inference loss is 23.03, weight deacy loss is 13.92, training accuracy is 0.093750, time 56.645 samples/sec
+epoch 0, total_step 7240, total loss is 45.37 , inference loss is 31.45, weight deacy loss is 13.92, training accuracy is 0.062500, time 55.865 samples/sec
+epoch 0, total_step 7260, total loss is 36.17 , inference loss is 22.26, weight deacy loss is 13.92, training accuracy is 0.031250, time 56.528 samples/sec
+epoch 0, total_step 7280, total loss is 35.62 , inference loss is 21.71, weight deacy loss is 13.91, training accuracy is 0.062500, time 55.243 samples/sec
+epoch 0, total_step 7300, total loss is 35.98 , inference loss is 22.07, weight deacy loss is 13.91, training accuracy is 0.031250, time 56.197 samples/sec
+epoch 0, total_step 7320, total loss is 39.59 , inference loss is 25.68, weight deacy loss is 13.91, training accuracy is 0.031250, time 55.235 samples/sec
+epoch 0, total_step 7340, total loss is 33.82 , inference loss is 19.91, weight deacy loss is 13.91, training accuracy is 0.062500, time 54.654 samples/sec
+epoch 0, total_step 7360, total loss is 36.90 , inference loss is 22.99, weight deacy loss is 13.91, training accuracy is 0.093750, time 56.384 samples/sec
+epoch 0, total_step 7380, total loss is 35.72 , inference loss is 21.82, weight deacy loss is 13.91, training accuracy is 0.093750, time 55.548 samples/sec
+epoch 0, total_step 7400, total loss is 40.27 , inference loss is 26.37, weight deacy loss is 13.90, training accuracy is 0.093750, time 56.130 samples/sec
+epoch 0, total_step 7420, total loss is 36.75 , inference loss is 22.85, weight deacy loss is 13.90, training accuracy is 0.062500, time 55.542 samples/sec
+epoch 0, total_step 7440, total loss is 37.11 , inference loss is 23.21, weight deacy loss is 13.90, training accuracy is 0.031250, time 56.961 samples/sec
+epoch 0, total_step 7460, total loss is 41.45 , inference loss is 27.55, weight deacy loss is 13.90, training accuracy is 0.062500, time 55.973 samples/sec
+epoch 0, total_step 7480, total loss is 38.80 , inference loss is 24.90, weight deacy loss is 13.90, training accuracy is 0.062500, time 55.400 samples/sec
+epoch 0, total_step 7500, total loss is 34.76 , inference loss is 20.86, weight deacy loss is 13.90, training accuracy is 0.031250, time 55.640 samples/sec
+epoch 0, total_step 7520, total loss is 33.00 , inference loss is 19.11, weight deacy loss is 13.90, training accuracy is 0.156250, time 56.485 samples/sec
+epoch 0, total_step 7540, total loss is 44.34 , inference loss is 30.44, weight deacy loss is 13.89, training accuracy is 0.093750, time 57.095 samples/sec
+epoch 0, total_step 7560, total loss is 38.49 , inference loss is 24.60, weight deacy loss is 13.89, training accuracy is 0.093750, time 56.467 samples/sec
+epoch 0, total_step 7580, total loss is 42.16 , inference loss is 28.27, weight deacy loss is 13.89, training accuracy is 0.031250, time 55.616 samples/sec
+epoch 0, total_step 7600, total loss is 37.24 , inference loss is 23.35, weight deacy loss is 13.89, training accuracy is 0.031250, time 54.044 samples/sec
+epoch 0, total_step 7620, total loss is 39.78 , inference loss is 25.89, weight deacy loss is 13.89, training accuracy is 0.062500, time 56.507 samples/sec
+epoch 0, total_step 7640, total loss is 37.35 , inference loss is 23.47, weight deacy loss is 13.89, training accuracy is 0.031250, time 56.321 samples/sec
+epoch 0, total_step 7660, total loss is 36.31 , inference loss is 22.43, weight deacy loss is 13.88, training accuracy is 0.062500, time 56.789 samples/sec
+epoch 0, total_step 7680, total loss is 38.11 , inference loss is 24.23, weight deacy loss is 13.88, training accuracy is 0.093750, time 57.159 samples/sec
+epoch 0, total_step 7700, total loss is 36.79 , inference loss is 22.91, weight deacy loss is 13.88, training accuracy is 0.000000, time 56.368 samples/sec
+epoch 0, total_step 7720, total loss is 38.27 , inference loss is 24.39, weight deacy loss is 13.88, training accuracy is 0.031250, time 55.772 samples/sec
+epoch 0, total_step 7740, total loss is 35.32 , inference loss is 21.44, weight deacy loss is 13.88, training accuracy is 0.093750, time 56.857 samples/sec
+epoch 0, total_step 7760, total loss is 31.57 , inference loss is 17.69, weight deacy loss is 13.88, training accuracy is 0.125000, time 56.468 samples/sec
+epoch 0, total_step 7780, total loss is 37.71 , inference loss is 23.84, weight deacy loss is 13.88, training accuracy is 0.031250, time 56.080 samples/sec
+epoch 0, total_step 7800, total loss is 36.55 , inference loss is 22.67, weight deacy loss is 13.87, training accuracy is 0.031250, time 56.075 samples/sec
+epoch 0, total_step 7820, total loss is 42.95 , inference loss is 29.07, weight deacy loss is 13.87, training accuracy is 0.000000, time 56.756 samples/sec
+epoch 0, total_step 7840, total loss is 42.42 , inference loss is 28.55, weight deacy loss is 13.87, training accuracy is 0.031250, time 56.640 samples/sec
+epoch 0, total_step 7860, total loss is 39.23 , inference loss is 25.36, weight deacy loss is 13.87, training accuracy is 0.156250, time 56.256 samples/sec
+epoch 0, total_step 7880, total loss is 39.28 , inference loss is 25.41, weight deacy loss is 13.87, training accuracy is 0.000000, time 56.464 samples/sec
+epoch 0, total_step 7900, total loss is 35.09 , inference loss is 21.23, weight deacy loss is 13.87, training accuracy is 0.031250, time 56.224 samples/sec
+epoch 0, total_step 7920, total loss is 43.32 , inference loss is 29.46, weight deacy loss is 13.87, training accuracy is 0.031250, time 56.107 samples/sec
+epoch 0, total_step 7940, total loss is 35.17 , inference loss is 21.31, weight deacy loss is 13.86, training accuracy is 0.031250, time 56.113 samples/sec
+epoch 0, total_step 7960, total loss is 44.53 , inference loss is 30.67, weight deacy loss is 13.86, training accuracy is 0.000000, time 56.335 samples/sec
+epoch 0, total_step 7980, total loss is 42.20 , inference loss is 28.34, weight deacy loss is 13.86, training accuracy is 0.031250, time 55.761 samples/sec
+testing verification..
+(12000, 512)
+infer time 75.00374600000008
+best_threshold_index 110 0.9401851851851852
+best_threshold_index 111 0.9409259259259259
+best_threshold_index 108 0.9422222222222222
+best_threshold_index 117 0.9407407407407408
+best_threshold_index 110 0.945
+best_threshold_index 117 0.9396296296296296
+best_threshold_index 111 0.9394444444444444
+best_threshold_index 117 0.9411111111111111
+best_threshold_index 111 0.9381481481481482
+best_threshold_index 117 0.9409259259259259
+[lfw][8000]XNorm: 22.766812
+[lfw][8000]Accuracy-Flip: 0.93783+-0.01623
+testing verification..
+(14000, 512)
+infer time 87.34959800000011
+best_threshold_index 111 0.937936507936508
+best_threshold_index 114 0.9342857142857143
+best_threshold_index 111 0.9350793650793651
+best_threshold_index 111 0.9342857142857143
+best_threshold_index 111 0.9355555555555556
+best_threshold_index 111 0.933968253968254
+best_threshold_index 111 0.9355555555555556
+best_threshold_index 110 0.9361904761904762
+best_threshold_index 114 0.9355555555555556
+best_threshold_index 111 0.9366666666666666
+[cfp_ff][8000]XNorm: 22.745294
+[cfp_ff][8000]Accuracy-Flip: 0.93357+-0.01025
+testing verification..
+(14000, 512)
+infer time 86.84563999999992
+best_threshold_index 141 0.78
+best_threshold_index 141 0.7774603174603174
+best_threshold_index 141 0.7749206349206349
+best_threshold_index 141 0.7771428571428571
+best_threshold_index 141 0.7801587301587302
+best_threshold_index 139 0.7698412698412699
+best_threshold_index 141 0.7753968253968254
+best_threshold_index 141 0.7746031746031746
+best_threshold_index 141 0.7749206349206349
+best_threshold_index 141 0.7785714285714286
+[cfp_fp][8000]XNorm: 20.570911
+[cfp_fp][8000]Accuracy-Flip: 0.77514+-0.02411
+testing verification..
+(12000, 512)
+infer time 74.07670599999994
+best_threshold_index 133 0.7622222222222222
+best_threshold_index 135 0.7603703703703704
+best_threshold_index 137 0.7625925925925926
+best_threshold_index 135 0.7642592592592593
+best_threshold_index 135 0.7688888888888888
+best_threshold_index 135 0.7644444444444445
+best_threshold_index 135 0.7638888888888888
+best_threshold_index 133 0.7657407407407407
+best_threshold_index 135 0.7616666666666667
+best_threshold_index 135 0.7638888888888888
+[agedb_30][8000]XNorm: 21.030835
+[agedb_30][8000]Accuracy-Flip: 0.76050+-0.02022
+test accuracy is: 0.9378333333333334
+epoch 0, total_step 8000, total loss is 34.30 , inference loss is 20.44, weight deacy loss is 13.86, training accuracy is 0.031250, time 52.881 samples/sec
+epoch 0, total_step 8020, total loss is 38.85 , inference loss is 25.00, weight deacy loss is 13.86, training accuracy is 0.031250, time 55.151 samples/sec
+epoch 0, total_step 8040, total loss is 36.66 , inference loss is 22.80, weight deacy loss is 13.86, training accuracy is 0.093750, time 56.065 samples/sec
+epoch 0, total_step 8060, total loss is 35.85 , inference loss is 22.00, weight deacy loss is 13.85, training accuracy is 0.031250, time 56.576 samples/sec
+epoch 0, total_step 8080, total loss is 37.25 , inference loss is 23.40, weight deacy loss is 13.85, training accuracy is 0.093750, time 55.296 samples/sec
+epoch 0, total_step 8100, total loss is 34.44 , inference loss is 20.58, weight deacy loss is 13.85, training accuracy is 0.062500, time 55.177 samples/sec
+epoch 0, total_step 8120, total loss is 36.15 , inference loss is 22.30, weight deacy loss is 13.85, training accuracy is 0.062500, time 56.703 samples/sec
+epoch 0, total_step 8140, total loss is 50.03 , inference loss is 36.19, weight deacy loss is 13.85, training accuracy is 0.000000, time 56.080 samples/sec
+epoch 0, total_step 8160, total loss is 42.52 , inference loss is 28.67, weight deacy loss is 13.85, training accuracy is 0.093750, time 55.704 samples/sec
+epoch 0, total_step 8180, total loss is 41.86 , inference loss is 28.01, weight deacy loss is 13.85, training accuracy is 0.062500, time 56.766 samples/sec
+epoch 0, total_step 8200, total loss is 36.37 , inference loss is 22.52, weight deacy loss is 13.84, training accuracy is 0.093750, time 55.642 samples/sec
+epoch 0, total_step 8220, total loss is 41.65 , inference loss is 27.81, weight deacy loss is 13.84, training accuracy is 0.000000, time 56.434 samples/sec
+epoch 0, total_step 8240, total loss is 33.58 , inference loss is 19.74, weight deacy loss is 13.84, training accuracy is 0.031250, time 55.874 samples/sec
+epoch 0, total_step 8260, total loss is 45.39 , inference loss is 31.56, weight deacy loss is 13.84, training accuracy is 0.062500, time 55.748 samples/sec
+epoch 0, total_step 8280, total loss is 40.10 , inference loss is 26.26, weight deacy loss is 13.84, training accuracy is 0.031250, time 57.199 samples/sec
+epoch 0, total_step 8300, total loss is 43.88 , inference loss is 30.04, weight deacy loss is 13.84, training accuracy is 0.093750, time 56.691 samples/sec
+epoch 0, total_step 8320, total loss is 36.93 , inference loss is 23.10, weight deacy loss is 13.84, training accuracy is 0.031250, time 55.936 samples/sec
+epoch 0, total_step 8340, total loss is 37.63 , inference loss is 23.80, weight deacy loss is 13.83, training accuracy is 0.031250, time 55.866 samples/sec
+epoch 0, total_step 8360, total loss is 41.13 , inference loss is 27.30, weight deacy loss is 13.83, training accuracy is 0.031250, time 56.810 samples/sec
+epoch 0, total_step 8380, total loss is 36.57 , inference loss is 22.74, weight deacy loss is 13.83, training accuracy is 0.000000, time 56.475 samples/sec
+epoch 0, total_step 8400, total loss is 39.38 , inference loss is 25.55, weight deacy loss is 13.83, training accuracy is 0.031250, time 55.505 samples/sec
+epoch 0, total_step 8420, total loss is 39.08 , inference loss is 25.25, weight deacy loss is 13.83, training accuracy is 0.031250, time 53.784 samples/sec
+epoch 0, total_step 8440, total loss is 36.68 , inference loss is 22.85, weight deacy loss is 13.83, training accuracy is 0.000000, time 56.136 samples/sec
+epoch 0, total_step 8460, total loss is 36.66 , inference loss is 22.84, weight deacy loss is 13.82, training accuracy is 0.125000, time 56.086 samples/sec
+epoch 0, total_step 8480, total loss is 39.93 , inference loss is 26.11, weight deacy loss is 13.82, training accuracy is 0.093750, time 56.023 samples/sec
+epoch 0, total_step 8500, total loss is 36.27 , inference loss is 22.44, weight deacy loss is 13.82, training accuracy is 0.000000, time 56.574 samples/sec
+epoch 0, total_step 8520, total loss is 37.71 , inference loss is 23.89, weight deacy loss is 13.82, training accuracy is 0.093750, time 56.136 samples/sec
+epoch 0, total_step 8540, total loss is 39.52 , inference loss is 25.70, weight deacy loss is 13.82, training accuracy is 0.031250, time 56.522 samples/sec
+epoch 0, total_step 8560, total loss is 36.32 , inference loss is 22.50, weight deacy loss is 13.82, training accuracy is 0.000000, time 56.393 samples/sec
+epoch 0, total_step 8580, total loss is 38.32 , inference loss is 24.50, weight deacy loss is 13.82, training accuracy is 0.031250, time 53.946 samples/sec
+epoch 0, total_step 8600, total loss is 38.67 , inference loss is 24.85, weight deacy loss is 13.81, training accuracy is 0.000000, time 56.578 samples/sec
+epoch 0, total_step 8620, total loss is 32.22 , inference loss is 18.41, weight deacy loss is 13.81, training accuracy is 0.062500, time 56.866 samples/sec
+epoch 0, total_step 8640, total loss is 33.68 , inference loss is 19.87, weight deacy loss is 13.81, training accuracy is 0.125000, time 56.778 samples/sec
+epoch 0, total_step 8660, total loss is 38.04 , inference loss is 24.24, weight deacy loss is 13.81, training accuracy is 0.062500, time 56.806 samples/sec
+epoch 0, total_step 8680, total loss is 41.14 , inference loss is 27.33, weight deacy loss is 13.81, training accuracy is 0.000000, time 55.699 samples/sec
+epoch 0, total_step 8700, total loss is 40.24 , inference loss is 26.43, weight deacy loss is 13.81, training accuracy is 0.000000, time 56.657 samples/sec
+epoch 0, total_step 8720, total loss is 43.61 , inference loss is 29.81, weight deacy loss is 13.81, training accuracy is 0.000000, time 56.386 samples/sec
+epoch 0, total_step 8740, total loss is 38.43 , inference loss is 24.63, weight deacy loss is 13.80, training accuracy is 0.062500, time 56.957 samples/sec
+epoch 0, total_step 8760, total loss is 40.85 , inference loss is 27.05, weight deacy loss is 13.80, training accuracy is 0.000000, time 56.594 samples/sec
+epoch 0, total_step 8780, total loss is 40.63 , inference loss is 26.83, weight deacy loss is 13.80, training accuracy is 0.000000, time 57.596 samples/sec
+epoch 0, total_step 8800, total loss is 36.96 , inference loss is 23.16, weight deacy loss is 13.80, training accuracy is 0.031250, time 56.557 samples/sec
+epoch 0, total_step 8820, total loss is 39.13 , inference loss is 25.34, weight deacy loss is 13.80, training accuracy is 0.000000, time 55.307 samples/sec
+epoch 0, total_step 8840, total loss is 32.36 , inference loss is 18.56, weight deacy loss is 13.80, training accuracy is 0.125000, time 56.953 samples/sec
+epoch 0, total_step 8860, total loss is 32.91 , inference loss is 19.11, weight deacy loss is 13.79, training accuracy is 0.031250, time 55.965 samples/sec
+epoch 0, total_step 8880, total loss is 34.15 , inference loss is 20.35, weight deacy loss is 13.79, training accuracy is 0.062500, time 56.114 samples/sec
+epoch 0, total_step 8900, total loss is 40.22 , inference loss is 26.43, weight deacy loss is 13.79, training accuracy is 0.062500, time 56.147 samples/sec
+epoch 0, total_step 8920, total loss is 40.67 , inference loss is 26.88, weight deacy loss is 13.79, training accuracy is 0.000000, time 56.155 samples/sec
+epoch 0, total_step 8940, total loss is 37.24 , inference loss is 23.45, weight deacy loss is 13.79, training accuracy is 0.000000, time 55.825 samples/sec
+epoch 0, total_step 8960, total loss is 36.65 , inference loss is 22.87, weight deacy loss is 13.79, training accuracy is 0.000000, time 56.368 samples/sec
+epoch 0, total_step 8980, total loss is 38.61 , inference loss is 24.82, weight deacy loss is 13.79, training accuracy is 0.093750, time 56.282 samples/sec
+epoch 0, total_step 9000, total loss is 36.90 , inference loss is 23.11, weight deacy loss is 13.78, training accuracy is 0.000000, time 56.519 samples/sec
+epoch 0, total_step 9020, total loss is 43.50 , inference loss is 29.72, weight deacy loss is 13.78, training accuracy is 0.000000, time 55.121 samples/sec
+epoch 0, total_step 9040, total loss is 38.08 , inference loss is 24.30, weight deacy loss is 13.78, training accuracy is 0.031250, time 57.097 samples/sec
+epoch 0, total_step 9060, total loss is 36.81 , inference loss is 23.03, weight deacy loss is 13.78, training accuracy is 0.031250, time 56.398 samples/sec
+epoch 0, total_step 9080, total loss is 36.26 , inference loss is 22.48, weight deacy loss is 13.78, training accuracy is 0.000000, time 55.562 samples/sec
+epoch 0, total_step 9100, total loss is 37.42 , inference loss is 23.64, weight deacy loss is 13.78, training accuracy is 0.000000, time 56.279 samples/sec
+epoch 0, total_step 9120, total loss is 39.93 , inference loss is 26.15, weight deacy loss is 13.78, training accuracy is 0.000000, time 57.800 samples/sec
+epoch 0, total_step 9140, total loss is 40.49 , inference loss is 26.71, weight deacy loss is 13.77, training accuracy is 0.062500, time 56.632 samples/sec
+epoch 0, total_step 9160, total loss is 39.48 , inference loss is 25.71, weight deacy loss is 13.77, training accuracy is 0.062500, time 56.363 samples/sec
+epoch 0, total_step 9180, total loss is 37.59 , inference loss is 23.82, weight deacy loss is 13.77, training accuracy is 0.031250, time 56.075 samples/sec
+epoch 0, total_step 9200, total loss is 44.27 , inference loss is 30.50, weight deacy loss is 13.77, training accuracy is 0.031250, time 56.639 samples/sec
+epoch 0, total_step 9220, total loss is 36.31 , inference loss is 22.54, weight deacy loss is 13.77, training accuracy is 0.062500, time 55.894 samples/sec
+epoch 0, total_step 9240, total loss is 35.13 , inference loss is 21.36, weight deacy loss is 13.77, training accuracy is 0.000000, time 56.377 samples/sec
+epoch 0, total_step 9260, total loss is 37.75 , inference loss is 23.98, weight deacy loss is 13.77, training accuracy is 0.062500, time 56.011 samples/sec
+epoch 0, total_step 9280, total loss is 39.39 , inference loss is 25.63, weight deacy loss is 13.76, training accuracy is 0.000000, time 56.591 samples/sec
+epoch 0, total_step 9300, total loss is 37.32 , inference loss is 23.56, weight deacy loss is 13.76, training accuracy is 0.062500, time 56.371 samples/sec
+epoch 0, total_step 9320, total loss is 38.64 , inference loss is 24.88, weight deacy loss is 13.76, training accuracy is 0.031250, time 56.187 samples/sec
+epoch 0, total_step 9340, total loss is 39.76 , inference loss is 26.00, weight deacy loss is 13.76, training accuracy is 0.031250, time 56.379 samples/sec
+epoch 0, total_step 9360, total loss is 38.35 , inference loss is 24.59, weight deacy loss is 13.76, training accuracy is 0.000000, time 56.554 samples/sec
+epoch 0, total_step 9380, total loss is 38.67 , inference loss is 24.92, weight deacy loss is 13.76, training accuracy is 0.125000, time 56.117 samples/sec
+epoch 0, total_step 9400, total loss is 39.81 , inference loss is 26.05, weight deacy loss is 13.75, training accuracy is 0.031250, time 56.225 samples/sec
+epoch 0, total_step 9420, total loss is 44.30 , inference loss is 30.55, weight deacy loss is 13.75, training accuracy is 0.093750, time 56.127 samples/sec
+epoch 0, total_step 9440, total loss is 37.84 , inference loss is 24.09, weight deacy loss is 13.75, training accuracy is 0.000000, time 55.457 samples/sec
+epoch 0, total_step 9460, total loss is 41.31 , inference loss is 27.56, weight deacy loss is 13.75, training accuracy is 0.031250, time 56.740 samples/sec
+epoch 0, total_step 9480, total loss is 41.53 , inference loss is 27.78, weight deacy loss is 13.75, training accuracy is 0.000000, time 55.925 samples/sec
+epoch 0, total_step 9500, total loss is 38.56 , inference loss is 24.81, weight deacy loss is 13.75, training accuracy is 0.031250, time 56.647 samples/sec
+epoch 0, total_step 9520, total loss is 39.73 , inference loss is 25.99, weight deacy loss is 13.75, training accuracy is 0.031250, time 56.682 samples/sec
+epoch 0, total_step 9540, total loss is 39.17 , inference loss is 25.43, weight deacy loss is 13.74, training accuracy is 0.000000, time 56.703 samples/sec
+epoch 0, total_step 9560, total loss is 37.07 , inference loss is 23.33, weight deacy loss is 13.74, training accuracy is 0.000000, time 56.905 samples/sec
+epoch 0, total_step 9580, total loss is 38.25 , inference loss is 24.51, weight deacy loss is 13.74, training accuracy is 0.062500, time 56.736 samples/sec
+epoch 0, total_step 9600, total loss is 36.19 , inference loss is 22.45, weight deacy loss is 13.74, training accuracy is 0.031250, time 55.615 samples/sec
+epoch 0, total_step 9620, total loss is 35.91 , inference loss is 22.17, weight deacy loss is 13.74, training accuracy is 0.031250, time 56.989 samples/sec
+epoch 0, total_step 9640, total loss is 35.70 , inference loss is 21.96, weight deacy loss is 13.74, training accuracy is 0.031250, time 55.350 samples/sec
+epoch 0, total_step 9660, total loss is 39.46 , inference loss is 25.72, weight deacy loss is 13.74, training accuracy is 0.000000, time 55.442 samples/sec
+epoch 0, total_step 9680, total loss is 41.29 , inference loss is 27.55, weight deacy loss is 13.73, training accuracy is 0.000000, time 56.278 samples/sec
+epoch 0, total_step 9700, total loss is 43.81 , inference loss is 30.07, weight deacy loss is 13.73, training accuracy is 0.031250, time 56.064 samples/sec
+epoch 0, total_step 9720, total loss is 35.71 , inference loss is 21.98, weight deacy loss is 13.73, training accuracy is 0.031250, time 55.793 samples/sec
+epoch 0, total_step 9740, total loss is 37.51 , inference loss is 23.78, weight deacy loss is 13.73, training accuracy is 0.031250, time 56.676 samples/sec
+epoch 0, total_step 9760, total loss is 37.05 , inference loss is 23.32, weight deacy loss is 13.73, training accuracy is 0.031250, time 56.678 samples/sec
+epoch 0, total_step 9780, total loss is 39.03 , inference loss is 25.30, weight deacy loss is 13.73, training accuracy is 0.093750, time 56.088 samples/sec
+epoch 0, total_step 9800, total loss is 37.61 , inference loss is 23.88, weight deacy loss is 13.73, training accuracy is 0.093750, time 56.613 samples/sec
+epoch 0, total_step 9820, total loss is 40.35 , inference loss is 26.63, weight deacy loss is 13.72, training accuracy is 0.000000, time 56.581 samples/sec
+epoch 0, total_step 9840, total loss is 35.33 , inference loss is 21.61, weight deacy loss is 13.72, training accuracy is 0.093750, time 55.830 samples/sec
+epoch 0, total_step 9860, total loss is 38.35 , inference loss is 24.63, weight deacy loss is 13.72, training accuracy is 0.062500, time 56.593 samples/sec
+epoch 0, total_step 9880, total loss is 41.77 , inference loss is 28.05, weight deacy loss is 13.72, training accuracy is 0.000000, time 56.471 samples/sec
+epoch 0, total_step 9900, total loss is 38.94 , inference loss is 25.22, weight deacy loss is 13.72, training accuracy is 0.093750, time 55.721 samples/sec
+epoch 0, total_step 9920, total loss is 36.35 , inference loss is 22.63, weight deacy loss is 13.72, training accuracy is 0.062500, time 55.538 samples/sec
+epoch 0, total_step 9940, total loss is 39.23 , inference loss is 25.52, weight deacy loss is 13.71, training accuracy is 0.000000, time 56.013 samples/sec
+epoch 0, total_step 9960, total loss is 40.03 , inference loss is 26.32, weight deacy loss is 13.71, training accuracy is 0.062500, time 56.933 samples/sec
+epoch 0, total_step 9980, total loss is 36.27 , inference loss is 22.56, weight deacy loss is 13.71, training accuracy is 0.031250, time 56.516 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.92542599999999
+best_threshold_index 98 0.9398148148148148
+best_threshold_index 100 0.9388888888888889
+best_threshold_index 100 0.9401851851851852
+best_threshold_index 100 0.9403703703703704
+best_threshold_index 98 0.9411111111111111
+best_threshold_index 98 0.9390740740740741
+best_threshold_index 100 0.9396296296296296
+best_threshold_index 98 0.9398148148148148
+best_threshold_index 100 0.9368518518518518
+best_threshold_index 100 0.9372222222222222
+[lfw][10000]XNorm: 24.557854
+[lfw][10000]Accuracy-Flip: 0.93733+-0.01188
+testing verification..
+(14000, 512)
+infer time 86.54144800000013
+best_threshold_index 108 0.9422222222222222
+best_threshold_index 108 0.9406349206349206
+best_threshold_index 108 0.9422222222222222
+best_threshold_index 104 0.9395238095238095
+best_threshold_index 108 0.9417460317460318
+best_threshold_index 105 0.9395238095238095
+best_threshold_index 108 0.9411111111111111
+best_threshold_index 105 0.9380952380952381
+best_threshold_index 104 0.9411111111111111
+best_threshold_index 105 0.9411111111111111
+[cfp_ff][10000]XNorm: 23.347114
+[cfp_ff][10000]Accuracy-Flip: 0.93843+-0.01083
+testing verification..
+(14000, 512)
+infer time 86.27496199999997
+best_threshold_index 127 0.7938095238095239
+best_threshold_index 126 0.7936507936507936
+best_threshold_index 128 0.792063492063492
+best_threshold_index 126 0.7926984126984127
+best_threshold_index 131 0.7931746031746032
+best_threshold_index 131 0.7887301587301587
+best_threshold_index 127 0.7904761904761904
+best_threshold_index 128 0.7915873015873016
+best_threshold_index 126 0.7923809523809524
+best_threshold_index 128 0.7926984126984127
+[cfp_fp][10000]XNorm: 21.072894
+[cfp_fp][10000]Accuracy-Flip: 0.78914+-0.01327
+testing verification..
+(12000, 512)
+infer time 73.60644100000002
+best_threshold_index 127 0.78
+best_threshold_index 127 0.7777777777777778
+best_threshold_index 127 0.7838888888888889
+best_threshold_index 127 0.7811111111111111
+best_threshold_index 127 0.7844444444444445
+best_threshold_index 127 0.7801851851851852
+best_threshold_index 127 0.7798148148148148
+best_threshold_index 127 0.7820370370370371
+best_threshold_index 127 0.7794444444444445
+best_threshold_index 127 0.782962962962963
+[agedb_30][10000]XNorm: 22.943766
+[agedb_30][10000]Accuracy-Flip: 0.78117+-0.01817
+test accuracy is: 0.9373333333333335
+epoch 0, total_step 10000, total loss is 40.92 , inference loss is 27.21, weight deacy loss is 13.71, training accuracy is 0.031250, time 53.178 samples/sec
+epoch 0, total_step 10020, total loss is 38.48 , inference loss is 24.77, weight deacy loss is 13.71, training accuracy is 0.031250, time 56.113 samples/sec
+epoch 0, total_step 10040, total loss is 33.76 , inference loss is 20.06, weight deacy loss is 13.71, training accuracy is 0.093750, time 55.532 samples/sec
+epoch 0, total_step 10060, total loss is 33.16 , inference loss is 19.45, weight deacy loss is 13.71, training accuracy is 0.125000, time 56.511 samples/sec
+epoch 0, total_step 10080, total loss is 33.87 , inference loss is 20.17, weight deacy loss is 13.70, training accuracy is 0.125000, time 56.197 samples/sec
+epoch 0, total_step 10100, total loss is 34.03 , inference loss is 20.33, weight deacy loss is 13.70, training accuracy is 0.093750, time 55.277 samples/sec
+epoch 0, total_step 10120, total loss is 39.34 , inference loss is 25.64, weight deacy loss is 13.70, training accuracy is 0.000000, time 56.335 samples/sec
+epoch 0, total_step 10140, total loss is 39.10 , inference loss is 25.40, weight deacy loss is 13.70, training accuracy is 0.031250, time 56.382 samples/sec
+epoch 0, total_step 10160, total loss is 36.45 , inference loss is 22.76, weight deacy loss is 13.70, training accuracy is 0.062500, time 56.346 samples/sec
+epoch 0, total_step 10180, total loss is 38.27 , inference loss is 24.58, weight deacy loss is 13.70, training accuracy is 0.062500, time 56.615 samples/sec
+epoch 0, total_step 10200, total loss is 38.10 , inference loss is 24.40, weight deacy loss is 13.70, training accuracy is 0.000000, time 56.448 samples/sec
+epoch 0, total_step 10220, total loss is 38.96 , inference loss is 25.26, weight deacy loss is 13.69, training accuracy is 0.187500, time 55.573 samples/sec
+epoch 0, total_step 10240, total loss is 38.83 , inference loss is 25.13, weight deacy loss is 13.69, training accuracy is 0.093750, time 55.037 samples/sec
+epoch 0, total_step 10260, total loss is 41.51 , inference loss is 27.82, weight deacy loss is 13.69, training accuracy is 0.093750, time 57.171 samples/sec
+epoch 0, total_step 10280, total loss is 38.37 , inference loss is 24.68, weight deacy loss is 13.69, training accuracy is 0.062500, time 56.285 samples/sec
+epoch 0, total_step 10300, total loss is 42.05 , inference loss is 28.36, weight deacy loss is 13.69, training accuracy is 0.093750, time 56.654 samples/sec
+epoch 0, total_step 10320, total loss is 39.07 , inference loss is 25.38, weight deacy loss is 13.69, training accuracy is 0.031250, time 56.396 samples/sec
+epoch 0, total_step 10340, total loss is 36.22 , inference loss is 22.53, weight deacy loss is 13.69, training accuracy is 0.062500, time 56.346 samples/sec
+epoch 0, total_step 10360, total loss is 37.47 , inference loss is 23.79, weight deacy loss is 13.68, training accuracy is 0.062500, time 56.351 samples/sec
+epoch 0, total_step 10380, total loss is 40.24 , inference loss is 26.56, weight deacy loss is 13.68, training accuracy is 0.031250, time 53.317 samples/sec
+epoch 0, total_step 10400, total loss is 33.47 , inference loss is 19.79, weight deacy loss is 13.68, training accuracy is 0.062500, time 56.060 samples/sec
+epoch 0, total_step 10420, total loss is 36.51 , inference loss is 22.83, weight deacy loss is 13.68, training accuracy is 0.031250, time 57.276 samples/sec
+epoch 0, total_step 10440, total loss is 39.55 , inference loss is 25.88, weight deacy loss is 13.68, training accuracy is 0.062500, time 55.670 samples/sec
+epoch 0, total_step 10460, total loss is 39.47 , inference loss is 25.79, weight deacy loss is 13.68, training accuracy is 0.062500, time 56.468 samples/sec
+epoch 0, total_step 10480, total loss is 31.30 , inference loss is 17.63, weight deacy loss is 13.67, training accuracy is 0.031250, time 56.110 samples/sec
+epoch 0, total_step 10500, total loss is 41.60 , inference loss is 27.93, weight deacy loss is 13.67, training accuracy is 0.031250, time 56.062 samples/sec
+epoch 0, total_step 10520, total loss is 35.04 , inference loss is 21.36, weight deacy loss is 13.67, training accuracy is 0.062500, time 55.699 samples/sec
+epoch 0, total_step 10540, total loss is 39.15 , inference loss is 25.48, weight deacy loss is 13.67, training accuracy is 0.062500, time 56.177 samples/sec
+epoch 0, total_step 10560, total loss is 34.80 , inference loss is 21.13, weight deacy loss is 13.67, training accuracy is 0.093750, time 57.285 samples/sec
+epoch 0, total_step 10580, total loss is 39.27 , inference loss is 25.60, weight deacy loss is 13.67, training accuracy is 0.000000, time 55.228 samples/sec
+epoch 0, total_step 10600, total loss is 38.66 , inference loss is 25.00, weight deacy loss is 13.67, training accuracy is 0.093750, time 55.280 samples/sec
+epoch 0, total_step 10620, total loss is 40.14 , inference loss is 26.47, weight deacy loss is 13.66, training accuracy is 0.000000, time 56.683 samples/sec
+epoch 0, total_step 10640, total loss is 39.02 , inference loss is 25.36, weight deacy loss is 13.66, training accuracy is 0.062500, time 56.387 samples/sec
+epoch 0, total_step 10660, total loss is 42.03 , inference loss is 28.37, weight deacy loss is 13.66, training accuracy is 0.000000, time 56.076 samples/sec
+epoch 0, total_step 10680, total loss is 39.38 , inference loss is 25.72, weight deacy loss is 13.66, training accuracy is 0.062500, time 55.896 samples/sec
+epoch 0, total_step 10700, total loss is 36.83 , inference loss is 23.17, weight deacy loss is 13.66, training accuracy is 0.062500, time 56.086 samples/sec
+epoch 0, total_step 10720, total loss is 40.72 , inference loss is 27.06, weight deacy loss is 13.66, training accuracy is 0.031250, time 56.141 samples/sec
+epoch 0, total_step 10740, total loss is 35.99 , inference loss is 22.33, weight deacy loss is 13.66, training accuracy is 0.093750, time 55.781 samples/sec
+epoch 0, total_step 10760, total loss is 39.20 , inference loss is 25.55, weight deacy loss is 13.65, training accuracy is 0.000000, time 56.724 samples/sec
+epoch 0, total_step 10780, total loss is 38.40 , inference loss is 24.75, weight deacy loss is 13.65, training accuracy is 0.062500, time 56.084 samples/sec
+epoch 0, total_step 10800, total loss is 42.86 , inference loss is 29.21, weight deacy loss is 13.65, training accuracy is 0.000000, time 56.391 samples/sec
+epoch 0, total_step 10820, total loss is 38.22 , inference loss is 24.57, weight deacy loss is 13.65, training accuracy is 0.093750, time 55.848 samples/sec
+epoch 0, total_step 10840, total loss is 40.60 , inference loss is 26.96, weight deacy loss is 13.65, training accuracy is 0.062500, time 56.327 samples/sec
+epoch 0, total_step 10860, total loss is 35.66 , inference loss is 22.01, weight deacy loss is 13.65, training accuracy is 0.031250, time 56.535 samples/sec
+epoch 0, total_step 10880, total loss is 37.29 , inference loss is 23.64, weight deacy loss is 13.65, training accuracy is 0.031250, time 54.127 samples/sec
+epoch 0, total_step 10900, total loss is 36.72 , inference loss is 23.08, weight deacy loss is 13.64, training accuracy is 0.031250, time 56.842 samples/sec
+epoch 0, total_step 10920, total loss is 40.18 , inference loss is 26.54, weight deacy loss is 13.64, training accuracy is 0.000000, time 57.153 samples/sec
+epoch 0, total_step 10940, total loss is 38.43 , inference loss is 24.79, weight deacy loss is 13.64, training accuracy is 0.031250, time 55.969 samples/sec
+epoch 0, total_step 10960, total loss is 39.74 , inference loss is 26.11, weight deacy loss is 13.64, training accuracy is 0.031250, time 55.537 samples/sec
+epoch 0, total_step 10980, total loss is 33.89 , inference loss is 20.25, weight deacy loss is 13.64, training accuracy is 0.031250, time 53.468 samples/sec
+epoch 0, total_step 11000, total loss is 35.17 , inference loss is 21.53, weight deacy loss is 13.64, training accuracy is 0.093750, time 56.428 samples/sec
+epoch 0, total_step 11020, total loss is 37.31 , inference loss is 23.68, weight deacy loss is 13.64, training accuracy is 0.031250, time 56.499 samples/sec
+epoch 0, total_step 11040, total loss is 39.98 , inference loss is 26.35, weight deacy loss is 13.63, training accuracy is 0.062500, time 57.128 samples/sec
+epoch 0, total_step 11060, total loss is 35.88 , inference loss is 22.25, weight deacy loss is 13.63, training accuracy is 0.031250, time 56.799 samples/sec
+epoch 0, total_step 11080, total loss is 35.30 , inference loss is 21.67, weight deacy loss is 13.63, training accuracy is 0.062500, time 55.783 samples/sec
+epoch 0, total_step 11100, total loss is 40.00 , inference loss is 26.37, weight deacy loss is 13.63, training accuracy is 0.093750, time 56.412 samples/sec
+epoch 0, total_step 11120, total loss is 38.44 , inference loss is 24.81, weight deacy loss is 13.63, training accuracy is 0.062500, time 55.087 samples/sec
+epoch 0, total_step 11140, total loss is 35.01 , inference loss is 21.38, weight deacy loss is 13.63, training accuracy is 0.031250, time 54.945 samples/sec
+epoch 0, total_step 11160, total loss is 33.90 , inference loss is 20.27, weight deacy loss is 13.63, training accuracy is 0.062500, time 56.161 samples/sec
+epoch 0, total_step 11180, total loss is 36.78 , inference loss is 23.16, weight deacy loss is 13.62, training accuracy is 0.031250, time 56.317 samples/sec
+epoch 0, total_step 11200, total loss is 42.85 , inference loss is 29.23, weight deacy loss is 13.62, training accuracy is 0.125000, time 56.105 samples/sec
+epoch 0, total_step 11220, total loss is 37.49 , inference loss is 23.87, weight deacy loss is 13.62, training accuracy is 0.031250, time 55.881 samples/sec
+epoch 0, total_step 11240, total loss is 32.00 , inference loss is 18.38, weight deacy loss is 13.62, training accuracy is 0.031250, time 56.427 samples/sec
+epoch 0, total_step 11260, total loss is 33.67 , inference loss is 20.05, weight deacy loss is 13.62, training accuracy is 0.000000, time 55.791 samples/sec
+epoch 0, total_step 11280, total loss is 37.54 , inference loss is 23.92, weight deacy loss is 13.62, training accuracy is 0.000000, time 54.104 samples/sec
+epoch 0, total_step 11300, total loss is 36.38 , inference loss is 22.76, weight deacy loss is 13.62, training accuracy is 0.062500, time 56.244 samples/sec
+epoch 0, total_step 11320, total loss is 39.47 , inference loss is 25.85, weight deacy loss is 13.61, training accuracy is 0.062500, time 55.934 samples/sec
+epoch 0, total_step 11340, total loss is 42.00 , inference loss is 28.39, weight deacy loss is 13.61, training accuracy is 0.000000, time 56.757 samples/sec
+epoch 0, total_step 11360, total loss is 35.79 , inference loss is 22.18, weight deacy loss is 13.61, training accuracy is 0.062500, time 55.278 samples/sec
+epoch 0, total_step 11380, total loss is 39.94 , inference loss is 26.33, weight deacy loss is 13.61, training accuracy is 0.062500, time 56.037 samples/sec
+epoch 0, total_step 11400, total loss is 39.66 , inference loss is 26.05, weight deacy loss is 13.61, training accuracy is 0.062500, time 56.618 samples/sec
+epoch 0, total_step 11420, total loss is 39.93 , inference loss is 26.32, weight deacy loss is 13.61, training accuracy is 0.000000, time 52.604 samples/sec
+epoch 0, total_step 11440, total loss is 35.21 , inference loss is 21.60, weight deacy loss is 13.60, training accuracy is 0.125000, time 56.462 samples/sec
+epoch 0, total_step 11460, total loss is 36.60 , inference loss is 23.00, weight deacy loss is 13.60, training accuracy is 0.062500, time 55.298 samples/sec
+epoch 0, total_step 11480, total loss is 37.79 , inference loss is 24.19, weight deacy loss is 13.60, training accuracy is 0.000000, time 57.060 samples/sec
+epoch 0, total_step 11500, total loss is 40.44 , inference loss is 26.84, weight deacy loss is 13.60, training accuracy is 0.062500, time 55.998 samples/sec
+epoch 0, total_step 11520, total loss is 39.73 , inference loss is 26.13, weight deacy loss is 13.60, training accuracy is 0.031250, time 56.881 samples/sec
+epoch 0, total_step 11540, total loss is 40.75 , inference loss is 27.15, weight deacy loss is 13.60, training accuracy is 0.000000, time 56.938 samples/sec
+epoch 0, total_step 11560, total loss is 38.57 , inference loss is 24.97, weight deacy loss is 13.60, training accuracy is 0.000000, time 56.082 samples/sec
+epoch 0, total_step 11580, total loss is 37.67 , inference loss is 24.07, weight deacy loss is 13.59, training accuracy is 0.031250, time 56.840 samples/sec
+epoch 0, total_step 11600, total loss is 42.16 , inference loss is 28.57, weight deacy loss is 13.59, training accuracy is 0.000000, time 55.974 samples/sec
+epoch 0, total_step 11620, total loss is 37.26 , inference loss is 23.66, weight deacy loss is 13.59, training accuracy is 0.031250, time 56.239 samples/sec
+epoch 0, total_step 11640, total loss is 32.83 , inference loss is 19.24, weight deacy loss is 13.59, training accuracy is 0.156250, time 55.645 samples/sec
+epoch 0, total_step 11660, total loss is 41.52 , inference loss is 27.93, weight deacy loss is 13.59, training accuracy is 0.031250, time 56.417 samples/sec
+epoch 0, total_step 11680, total loss is 34.75 , inference loss is 21.16, weight deacy loss is 13.59, training accuracy is 0.031250, time 56.254 samples/sec
+epoch 0, total_step 11700, total loss is 38.32 , inference loss is 24.73, weight deacy loss is 13.59, training accuracy is 0.031250, time 55.546 samples/sec
+epoch 0, total_step 11720, total loss is 38.29 , inference loss is 24.70, weight deacy loss is 13.59, training accuracy is 0.031250, time 56.014 samples/sec
+epoch 0, total_step 11740, total loss is 38.73 , inference loss is 25.14, weight deacy loss is 13.58, training accuracy is 0.062500, time 57.244 samples/sec
+epoch 0, total_step 11760, total loss is 33.06 , inference loss is 19.48, weight deacy loss is 13.58, training accuracy is 0.000000, time 56.253 samples/sec
+epoch 0, total_step 11780, total loss is 38.59 , inference loss is 25.01, weight deacy loss is 13.58, training accuracy is 0.125000, time 56.303 samples/sec
+epoch 0, total_step 11800, total loss is 38.25 , inference loss is 24.67, weight deacy loss is 13.58, training accuracy is 0.062500, time 57.035 samples/sec
+epoch 0, total_step 11820, total loss is 42.47 , inference loss is 28.89, weight deacy loss is 13.58, training accuracy is 0.000000, time 56.451 samples/sec
+epoch 0, total_step 11840, total loss is 42.06 , inference loss is 28.48, weight deacy loss is 13.58, training accuracy is 0.000000, time 56.360 samples/sec
+epoch 0, total_step 11860, total loss is 38.58 , inference loss is 25.00, weight deacy loss is 13.58, training accuracy is 0.093750, time 56.664 samples/sec
+epoch 0, total_step 11880, total loss is 40.63 , inference loss is 27.06, weight deacy loss is 13.57, training accuracy is 0.000000, time 55.721 samples/sec
+epoch 0, total_step 11900, total loss is 43.24 , inference loss is 29.67, weight deacy loss is 13.57, training accuracy is 0.000000, time 56.690 samples/sec
+epoch 0, total_step 11920, total loss is 42.42 , inference loss is 28.85, weight deacy loss is 13.57, training accuracy is 0.000000, time 56.902 samples/sec
+epoch 0, total_step 11940, total loss is 40.09 , inference loss is 26.52, weight deacy loss is 13.57, training accuracy is 0.000000, time 56.491 samples/sec
+epoch 0, total_step 11960, total loss is 35.91 , inference loss is 22.34, weight deacy loss is 13.57, training accuracy is 0.031250, time 55.860 samples/sec
+epoch 0, total_step 11980, total loss is 38.22 , inference loss is 24.65, weight deacy loss is 13.57, training accuracy is 0.000000, time 55.174 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.775719
+best_threshold_index 100 0.9472222222222222
+best_threshold_index 100 0.9472222222222222
+best_threshold_index 100 0.9488888888888889
+best_threshold_index 100 0.9483333333333334
+best_threshold_index 100 0.9494444444444444
+best_threshold_index 100 0.9464814814814815
+best_threshold_index 100 0.947962962962963
+best_threshold_index 100 0.9477777777777778
+best_threshold_index 100 0.9448148148148148
+best_threshold_index 100 0.9451851851851852
+[lfw][12000]XNorm: 25.516436
+[lfw][12000]Accuracy-Flip: 0.94733+-0.01278
+testing verification..
+(14000, 512)
+infer time 86.19727399999992
+best_threshold_index 104 0.9522222222222222
+best_threshold_index 104 0.9514285714285714
+best_threshold_index 104 0.9511111111111111
+best_threshold_index 104 0.9493650793650794
+best_threshold_index 104 0.9514285714285714
+best_threshold_index 104 0.9503174603174603
+best_threshold_index 104 0.9514285714285714
+best_threshold_index 104 0.9477777777777778
+best_threshold_index 101 0.9501587301587302
+best_threshold_index 104 0.9533333333333334
+[cfp_ff][12000]XNorm: 25.261931
+[cfp_ff][12000]Accuracy-Flip: 0.94986+-0.01307
+testing verification..
+(14000, 512)
+infer time 86.54471900000007
+best_threshold_index 127 0.8103174603174603
+best_threshold_index 127 0.8092063492063492
+best_threshold_index 127 0.8053968253968254
+best_threshold_index 127 0.8087301587301587
+best_threshold_index 127 0.8117460317460318
+best_threshold_index 127 0.8033333333333333
+best_threshold_index 127 0.8066666666666666
+best_threshold_index 127 0.8065079365079365
+best_threshold_index 127 0.8087301587301587
+best_threshold_index 127 0.807936507936508
+[cfp_fp][12000]XNorm: 22.910408
+[cfp_fp][12000]Accuracy-Flip: 0.80786+-0.02097
+testing verification..
+(12000, 512)
+infer time 74.15823599999995
+best_threshold_index 123 0.7946296296296296
+best_threshold_index 122 0.7966666666666666
+best_threshold_index 122 0.7996296296296296
+best_threshold_index 123 0.7974074074074075
+best_threshold_index 122 0.8022222222222222
+best_threshold_index 122 0.7942592592592592
+best_threshold_index 122 0.7983333333333333
+best_threshold_index 122 0.8012962962962963
+best_threshold_index 122 0.7972222222222223
+best_threshold_index 122 0.7977777777777778
+[agedb_30][12000]XNorm: 24.324147
+[agedb_30][12000]Accuracy-Flip: 0.79650+-0.02171
+test accuracy is: 0.9473333333333335
+epoch 0, total_step 12000, total loss is 37.47 , inference loss is 23.91, weight deacy loss is 13.57, training accuracy is 0.000000, time 53.379 samples/sec
+epoch 0, total_step 12020, total loss is 37.03 , inference loss is 23.47, weight deacy loss is 13.56, training accuracy is 0.125000, time 55.901 samples/sec
+epoch 0, total_step 12040, total loss is 37.60 , inference loss is 24.04, weight deacy loss is 13.56, training accuracy is 0.000000, time 56.661 samples/sec
+epoch 0, total_step 12060, total loss is 39.36 , inference loss is 25.80, weight deacy loss is 13.56, training accuracy is 0.000000, time 56.168 samples/sec
+epoch 0, total_step 12080, total loss is 35.79 , inference loss is 22.23, weight deacy loss is 13.56, training accuracy is 0.031250, time 56.104 samples/sec
+epoch 0, total_step 12100, total loss is 39.48 , inference loss is 25.92, weight deacy loss is 13.56, training accuracy is 0.062500, time 56.230 samples/sec
+epoch 0, total_step 12120, total loss is 39.63 , inference loss is 26.07, weight deacy loss is 13.56, training accuracy is 0.000000, time 55.486 samples/sec
+epoch 0, total_step 12140, total loss is 40.38 , inference loss is 26.83, weight deacy loss is 13.56, training accuracy is 0.000000, time 55.624 samples/sec
+epoch 0, total_step 12160, total loss is 38.95 , inference loss is 25.40, weight deacy loss is 13.55, training accuracy is 0.031250, time 56.393 samples/sec
+epoch 0, total_step 12180, total loss is 38.77 , inference loss is 25.22, weight deacy loss is 13.55, training accuracy is 0.031250, time 56.646 samples/sec
+epoch 0, total_step 12200, total loss is 43.89 , inference loss is 30.34, weight deacy loss is 13.55, training accuracy is 0.031250, time 55.796 samples/sec
+epoch 0, total_step 12220, total loss is 34.64 , inference loss is 21.09, weight deacy loss is 13.55, training accuracy is 0.000000, time 56.771 samples/sec
+epoch 0, total_step 12240, total loss is 35.77 , inference loss is 22.22, weight deacy loss is 13.55, training accuracy is 0.000000, time 54.519 samples/sec
+epoch 0, total_step 12260, total loss is 37.67 , inference loss is 24.12, weight deacy loss is 13.55, training accuracy is 0.062500, time 57.119 samples/sec
+epoch 0, total_step 12280, total loss is 39.90 , inference loss is 26.35, weight deacy loss is 13.55, training accuracy is 0.062500, time 56.331 samples/sec
+epoch 0, total_step 12300, total loss is 41.29 , inference loss is 27.74, weight deacy loss is 13.55, training accuracy is 0.000000, time 55.609 samples/sec
+epoch 0, total_step 12320, total loss is 34.22 , inference loss is 20.68, weight deacy loss is 13.54, training accuracy is 0.062500, time 57.084 samples/sec
+epoch 0, total_step 12340, total loss is 35.95 , inference loss is 22.41, weight deacy loss is 13.54, training accuracy is 0.031250, time 55.669 samples/sec
+epoch 0, total_step 12360, total loss is 33.77 , inference loss is 20.23, weight deacy loss is 13.54, training accuracy is 0.000000, time 56.914 samples/sec
+epoch 0, total_step 12380, total loss is 35.81 , inference loss is 22.27, weight deacy loss is 13.54, training accuracy is 0.062500, time 56.196 samples/sec
+epoch 0, total_step 12400, total loss is 35.13 , inference loss is 21.59, weight deacy loss is 13.54, training accuracy is 0.093750, time 55.679 samples/sec
+epoch 0, total_step 12420, total loss is 34.17 , inference loss is 20.63, weight deacy loss is 13.54, training accuracy is 0.062500, time 55.867 samples/sec
+epoch 0, total_step 12440, total loss is 38.12 , inference loss is 24.58, weight deacy loss is 13.54, training accuracy is 0.031250, time 55.884 samples/sec
+epoch 0, total_step 12460, total loss is 40.73 , inference loss is 27.19, weight deacy loss is 13.53, training accuracy is 0.000000, time 56.071 samples/sec
+epoch 0, total_step 12480, total loss is 36.62 , inference loss is 23.09, weight deacy loss is 13.53, training accuracy is 0.000000, time 56.934 samples/sec
+epoch 0, total_step 12500, total loss is 43.68 , inference loss is 30.15, weight deacy loss is 13.53, training accuracy is 0.000000, time 56.225 samples/sec
+epoch 0, total_step 12520, total loss is 32.38 , inference loss is 18.85, weight deacy loss is 13.53, training accuracy is 0.062500, time 55.906 samples/sec
+epoch 0, total_step 12540, total loss is 43.39 , inference loss is 29.86, weight deacy loss is 13.53, training accuracy is 0.000000, time 55.844 samples/sec
+epoch 0, total_step 12560, total loss is 35.14 , inference loss is 21.61, weight deacy loss is 13.53, training accuracy is 0.031250, time 56.844 samples/sec
+epoch 0, total_step 12580, total loss is 39.87 , inference loss is 26.34, weight deacy loss is 13.53, training accuracy is 0.000000, time 56.360 samples/sec
+epoch 0, total_step 12600, total loss is 38.80 , inference loss is 25.28, weight deacy loss is 13.52, training accuracy is 0.031250, time 54.195 samples/sec
+epoch 0, total_step 12620, total loss is 37.60 , inference loss is 24.08, weight deacy loss is 13.52, training accuracy is 0.031250, time 55.521 samples/sec
+epoch 0, total_step 12640, total loss is 37.25 , inference loss is 23.73, weight deacy loss is 13.52, training accuracy is 0.031250, time 56.704 samples/sec
+epoch 0, total_step 12660, total loss is 39.10 , inference loss is 25.58, weight deacy loss is 13.52, training accuracy is 0.000000, time 56.968 samples/sec
+epoch 0, total_step 12680, total loss is 37.96 , inference loss is 24.44, weight deacy loss is 13.52, training accuracy is 0.000000, time 56.049 samples/sec
+epoch 0, total_step 12700, total loss is 42.42 , inference loss is 28.90, weight deacy loss is 13.52, training accuracy is 0.000000, time 56.234 samples/sec
+epoch 0, total_step 12720, total loss is 33.47 , inference loss is 19.95, weight deacy loss is 13.52, training accuracy is 0.031250, time 57.293 samples/sec
+epoch 0, total_step 12740, total loss is 39.55 , inference loss is 26.04, weight deacy loss is 13.51, training accuracy is 0.031250, time 56.262 samples/sec
+epoch 0, total_step 12760, total loss is 39.55 , inference loss is 26.04, weight deacy loss is 13.51, training accuracy is 0.000000, time 57.210 samples/sec
+epoch 0, total_step 12780, total loss is 44.91 , inference loss is 31.39, weight deacy loss is 13.51, training accuracy is 0.000000, time 56.381 samples/sec
+epoch 0, total_step 12800, total loss is 41.33 , inference loss is 27.82, weight deacy loss is 13.51, training accuracy is 0.000000, time 57.147 samples/sec
+epoch 0, total_step 12820, total loss is 38.54 , inference loss is 25.03, weight deacy loss is 13.51, training accuracy is 0.031250, time 55.719 samples/sec
+epoch 0, total_step 12840, total loss is 38.77 , inference loss is 25.26, weight deacy loss is 13.51, training accuracy is 0.031250, time 55.502 samples/sec
+epoch 0, total_step 12860, total loss is 41.86 , inference loss is 28.35, weight deacy loss is 13.51, training accuracy is 0.000000, time 56.487 samples/sec
+epoch 0, total_step 12880, total loss is 38.22 , inference loss is 24.71, weight deacy loss is 13.51, training accuracy is 0.031250, time 56.410 samples/sec
+epoch 0, total_step 12900, total loss is 36.76 , inference loss is 23.26, weight deacy loss is 13.50, training accuracy is 0.000000, time 56.880 samples/sec
+epoch 0, total_step 12920, total loss is 36.79 , inference loss is 23.29, weight deacy loss is 13.50, training accuracy is 0.062500, time 56.342 samples/sec
+epoch 0, total_step 12940, total loss is 41.31 , inference loss is 27.81, weight deacy loss is 13.50, training accuracy is 0.062500, time 56.388 samples/sec
+epoch 0, total_step 12960, total loss is 41.24 , inference loss is 27.74, weight deacy loss is 13.50, training accuracy is 0.031250, time 55.731 samples/sec
+epoch 0, total_step 12980, total loss is 40.40 , inference loss is 26.91, weight deacy loss is 13.50, training accuracy is 0.031250, time 56.698 samples/sec
+epoch 0, total_step 13000, total loss is 40.18 , inference loss is 26.68, weight deacy loss is 13.50, training accuracy is 0.000000, time 55.445 samples/sec
+epoch 0, total_step 13020, total loss is 41.16 , inference loss is 27.66, weight deacy loss is 13.50, training accuracy is 0.000000, time 56.664 samples/sec
+epoch 0, total_step 13040, total loss is 35.00 , inference loss is 21.51, weight deacy loss is 13.49, training accuracy is 0.031250, time 56.885 samples/sec
+epoch 0, total_step 13060, total loss is 39.01 , inference loss is 25.52, weight deacy loss is 13.49, training accuracy is 0.000000, time 57.395 samples/sec
+epoch 0, total_step 13080, total loss is 37.45 , inference loss is 23.96, weight deacy loss is 13.49, training accuracy is 0.031250, time 56.068 samples/sec
+epoch 0, total_step 13100, total loss is 38.94 , inference loss is 25.45, weight deacy loss is 13.49, training accuracy is 0.000000, time 56.997 samples/sec
+epoch 0, total_step 13120, total loss is 35.51 , inference loss is 22.02, weight deacy loss is 13.49, training accuracy is 0.031250, time 56.560 samples/sec
+epoch 0, total_step 13140, total loss is 41.16 , inference loss is 27.67, weight deacy loss is 13.49, training accuracy is 0.000000, time 56.939 samples/sec
+epoch 0, total_step 13160, total loss is 38.15 , inference loss is 24.66, weight deacy loss is 13.49, training accuracy is 0.062500, time 56.007 samples/sec
+epoch 0, total_step 13180, total loss is 41.83 , inference loss is 28.34, weight deacy loss is 13.48, training accuracy is 0.000000, time 56.784 samples/sec
+epoch 0, total_step 13200, total loss is 35.43 , inference loss is 21.94, weight deacy loss is 13.48, training accuracy is 0.031250, time 56.953 samples/sec
+epoch 0, total_step 13220, total loss is 40.78 , inference loss is 27.29, weight deacy loss is 13.48, training accuracy is 0.000000, time 56.606 samples/sec
+epoch 0, total_step 13240, total loss is 39.34 , inference loss is 25.86, weight deacy loss is 13.48, training accuracy is 0.062500, time 56.451 samples/sec
+epoch 0, total_step 13260, total loss is 46.75 , inference loss is 33.27, weight deacy loss is 13.48, training accuracy is 0.031250, time 56.655 samples/sec
+epoch 0, total_step 13280, total loss is 37.23 , inference loss is 23.75, weight deacy loss is 13.48, training accuracy is 0.093750, time 56.396 samples/sec
+epoch 0, total_step 13300, total loss is 37.27 , inference loss is 23.79, weight deacy loss is 13.48, training accuracy is 0.000000, time 56.506 samples/sec
+epoch 0, total_step 13320, total loss is 37.38 , inference loss is 23.91, weight deacy loss is 13.47, training accuracy is 0.031250, time 57.007 samples/sec
+epoch 0, total_step 13340, total loss is 38.16 , inference loss is 24.69, weight deacy loss is 13.47, training accuracy is 0.031250, time 56.766 samples/sec
+epoch 0, total_step 13360, total loss is 41.01 , inference loss is 27.53, weight deacy loss is 13.47, training accuracy is 0.000000, time 56.592 samples/sec
+epoch 0, total_step 13380, total loss is 36.77 , inference loss is 23.30, weight deacy loss is 13.47, training accuracy is 0.062500, time 57.084 samples/sec
+epoch 0, total_step 13400, total loss is 36.78 , inference loss is 23.31, weight deacy loss is 13.47, training accuracy is 0.000000, time 55.428 samples/sec
+epoch 0, total_step 13420, total loss is 36.44 , inference loss is 22.97, weight deacy loss is 13.47, training accuracy is 0.031250, time 57.038 samples/sec
+epoch 0, total_step 13440, total loss is 37.43 , inference loss is 23.97, weight deacy loss is 13.47, training accuracy is 0.062500, time 56.738 samples/sec
+epoch 0, total_step 13460, total loss is 42.29 , inference loss is 28.83, weight deacy loss is 13.47, training accuracy is 0.000000, time 55.254 samples/sec
+epoch 0, total_step 13480, total loss is 38.88 , inference loss is 25.42, weight deacy loss is 13.46, training accuracy is 0.062500, time 55.514 samples/sec
+epoch 0, total_step 13500, total loss is 34.78 , inference loss is 21.32, weight deacy loss is 13.46, training accuracy is 0.031250, time 55.901 samples/sec
+epoch 0, total_step 13520, total loss is 44.88 , inference loss is 31.42, weight deacy loss is 13.46, training accuracy is 0.000000, time 56.234 samples/sec
+epoch 0, total_step 13540, total loss is 36.42 , inference loss is 22.96, weight deacy loss is 13.46, training accuracy is 0.031250, time 56.682 samples/sec
+epoch 0, total_step 13560, total loss is 41.27 , inference loss is 27.82, weight deacy loss is 13.46, training accuracy is 0.000000, time 56.791 samples/sec
+epoch 0, total_step 13580, total loss is 39.13 , inference loss is 25.67, weight deacy loss is 13.46, training accuracy is 0.000000, time 56.622 samples/sec
+epoch 0, total_step 13600, total loss is 39.82 , inference loss is 26.37, weight deacy loss is 13.46, training accuracy is 0.062500, time 56.690 samples/sec
+epoch 0, total_step 13620, total loss is 41.13 , inference loss is 27.68, weight deacy loss is 13.45, training accuracy is 0.031250, time 56.766 samples/sec
+epoch 0, total_step 13640, total loss is 39.77 , inference loss is 26.31, weight deacy loss is 13.45, training accuracy is 0.031250, time 56.165 samples/sec
+epoch 0, total_step 13660, total loss is 36.96 , inference loss is 23.50, weight deacy loss is 13.45, training accuracy is 0.031250, time 56.330 samples/sec
+epoch 0, total_step 13680, total loss is 42.78 , inference loss is 29.33, weight deacy loss is 13.45, training accuracy is 0.000000, time 56.084 samples/sec
+epoch 0, total_step 13700, total loss is 41.30 , inference loss is 27.85, weight deacy loss is 13.45, training accuracy is 0.000000, time 57.446 samples/sec
+epoch 0, total_step 13720, total loss is 35.04 , inference loss is 21.59, weight deacy loss is 13.45, training accuracy is 0.062500, time 56.505 samples/sec
+epoch 0, total_step 13740, total loss is 39.18 , inference loss is 25.73, weight deacy loss is 13.45, training accuracy is 0.000000, time 56.632 samples/sec
+epoch 0, total_step 13760, total loss is 31.56 , inference loss is 18.12, weight deacy loss is 13.44, training accuracy is 0.093750, time 56.243 samples/sec
+epoch 0, total_step 13780, total loss is 33.85 , inference loss is 20.41, weight deacy loss is 13.44, training accuracy is 0.031250, time 56.435 samples/sec
+epoch 0, total_step 13800, total loss is 37.50 , inference loss is 24.06, weight deacy loss is 13.44, training accuracy is 0.031250, time 55.996 samples/sec
+epoch 0, total_step 13820, total loss is 37.93 , inference loss is 24.48, weight deacy loss is 13.44, training accuracy is 0.093750, time 56.850 samples/sec
+epoch 0, total_step 13840, total loss is 36.67 , inference loss is 23.23, weight deacy loss is 13.44, training accuracy is 0.031250, time 56.327 samples/sec
+epoch 0, total_step 13860, total loss is 33.03 , inference loss is 19.59, weight deacy loss is 13.44, training accuracy is 0.031250, time 55.291 samples/sec
+epoch 0, total_step 13880, total loss is 36.94 , inference loss is 23.50, weight deacy loss is 13.44, training accuracy is 0.000000, time 55.628 samples/sec
+epoch 0, total_step 13900, total loss is 38.90 , inference loss is 25.47, weight deacy loss is 13.43, training accuracy is 0.000000, time 56.071 samples/sec
+epoch 0, total_step 13920, total loss is 41.50 , inference loss is 28.07, weight deacy loss is 13.43, training accuracy is 0.031250, time 56.415 samples/sec
+epoch 0, total_step 13940, total loss is 32.82 , inference loss is 19.38, weight deacy loss is 13.43, training accuracy is 0.062500, time 57.263 samples/sec
+epoch 0, total_step 13960, total loss is 38.83 , inference loss is 25.40, weight deacy loss is 13.43, training accuracy is 0.000000, time 56.379 samples/sec
+epoch 0, total_step 13980, total loss is 41.82 , inference loss is 28.39, weight deacy loss is 13.43, training accuracy is 0.031250, time 54.784 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.56741299999996
+best_threshold_index 95 0.9522222222222222
+best_threshold_index 95 0.9516666666666667
+best_threshold_index 95 0.9538888888888889
+best_threshold_index 95 0.9540740740740741
+best_threshold_index 95 0.9535185185185185
+best_threshold_index 95 0.9503703703703704
+best_threshold_index 95 0.9524074074074074
+best_threshold_index 95 0.9522222222222222
+best_threshold_index 95 0.95
+best_threshold_index 95 0.9512962962962963
+[lfw][14000]XNorm: 26.565265
+[lfw][14000]Accuracy-Flip: 0.95217+-0.01188
+testing verification..
+(14000, 512)
+infer time 86.62059800000013
+best_threshold_index 101 0.9493650793650794
+best_threshold_index 100 0.9450793650793651
+best_threshold_index 101 0.9473015873015873
+best_threshold_index 101 0.9455555555555556
+best_threshold_index 101 0.9477777777777778
+best_threshold_index 101 0.9446031746031746
+best_threshold_index 101 0.946984126984127
+best_threshold_index 101 0.9453968253968253
+best_threshold_index 101 0.9463492063492064
+best_threshold_index 101 0.9473015873015873
+[cfp_ff][14000]XNorm: 25.964787
+[cfp_ff][14000]Accuracy-Flip: 0.94614+-0.01202
+testing verification..
+(14000, 512)
+infer time 86.05347900000004
+best_threshold_index 115 0.806031746031746
+best_threshold_index 114 0.8050793650793651
+best_threshold_index 117 0.806031746031746
+best_threshold_index 115 0.8065079365079365
+best_threshold_index 115 0.8084126984126984
+best_threshold_index 115 0.8019047619047619
+best_threshold_index 115 0.8033333333333333
+best_threshold_index 115 0.8053968253968254
+best_threshold_index 115 0.8073015873015873
+best_threshold_index 115 0.8047619047619048
+[cfp_fp][14000]XNorm: 23.668343
+[cfp_fp][14000]Accuracy-Flip: 0.80371+-0.01632
+testing verification..
+(12000, 512)
+infer time 74.12428799999999
+best_threshold_index 116 0.8024074074074075
+best_threshold_index 116 0.8031481481481482
+best_threshold_index 116 0.8066666666666666
+best_threshold_index 116 0.8033333333333333
+best_threshold_index 116 0.8072222222222222
+best_threshold_index 116 0.8022222222222222
+best_threshold_index 116 0.8072222222222222
+best_threshold_index 119 0.8098148148148148
+best_threshold_index 116 0.8037037037037037
+best_threshold_index 116 0.8048148148148148
+[agedb_30][14000]XNorm: 25.114798
+[agedb_30][14000]Accuracy-Flip: 0.80367+-0.02343
+test accuracy is: 0.9521666666666668
+epoch 0, total_step 14000, total loss is 38.92 , inference loss is 25.49, weight deacy loss is 13.43, training accuracy is 0.000000, time 53.487 samples/sec
+epoch 0, total_step 14020, total loss is 43.02 , inference loss is 29.59, weight deacy loss is 13.43, training accuracy is 0.000000, time 56.712 samples/sec
+epoch 0, total_step 14040, total loss is 39.87 , inference loss is 26.44, weight deacy loss is 13.43, training accuracy is 0.000000, time 56.106 samples/sec
+epoch 0, total_step 14060, total loss is 37.12 , inference loss is 23.69, weight deacy loss is 13.42, training accuracy is 0.000000, time 56.740 samples/sec
+epoch 0, total_step 14080, total loss is 37.67 , inference loss is 24.24, weight deacy loss is 13.42, training accuracy is 0.000000, time 54.994 samples/sec
+epoch 0, total_step 14100, total loss is 36.98 , inference loss is 23.55, weight deacy loss is 13.42, training accuracy is 0.031250, time 56.247 samples/sec
+epoch 0, total_step 14120, total loss is 37.90 , inference loss is 24.47, weight deacy loss is 13.42, training accuracy is 0.000000, time 56.171 samples/sec
+epoch 0, total_step 14140, total loss is 42.02 , inference loss is 28.60, weight deacy loss is 13.42, training accuracy is 0.000000, time 56.165 samples/sec
+epoch 0, total_step 14160, total loss is 36.07 , inference loss is 22.65, weight deacy loss is 13.42, training accuracy is 0.031250, time 56.180 samples/sec
+epoch 0, total_step 14180, total loss is 38.38 , inference loss is 24.96, weight deacy loss is 13.42, training accuracy is 0.093750, time 56.292 samples/sec
+epoch 0, total_step 14200, total loss is 36.52 , inference loss is 23.10, weight deacy loss is 13.42, training accuracy is 0.062500, time 56.481 samples/sec
+epoch 0, total_step 14220, total loss is 39.51 , inference loss is 26.09, weight deacy loss is 13.41, training accuracy is 0.031250, time 54.083 samples/sec
+epoch 0, total_step 14240, total loss is 41.16 , inference loss is 27.75, weight deacy loss is 13.41, training accuracy is 0.000000, time 56.611 samples/sec
+epoch 0, total_step 14260, total loss is 37.26 , inference loss is 23.85, weight deacy loss is 13.41, training accuracy is 0.031250, time 55.771 samples/sec
+epoch 0, total_step 14280, total loss is 35.27 , inference loss is 21.86, weight deacy loss is 13.41, training accuracy is 0.000000, time 56.665 samples/sec
+epoch 0, total_step 14300, total loss is 35.81 , inference loss is 22.40, weight deacy loss is 13.41, training accuracy is 0.000000, time 56.588 samples/sec
+epoch 0, total_step 14320, total loss is 39.99 , inference loss is 26.59, weight deacy loss is 13.41, training accuracy is 0.000000, time 56.210 samples/sec
+epoch 0, total_step 14340, total loss is 39.60 , inference loss is 26.19, weight deacy loss is 13.41, training accuracy is 0.000000, time 55.921 samples/sec
+epoch 0, total_step 14360, total loss is 37.74 , inference loss is 24.33, weight deacy loss is 13.40, training accuracy is 0.000000, time 56.106 samples/sec
+epoch 0, total_step 14380, total loss is 34.81 , inference loss is 21.40, weight deacy loss is 13.40, training accuracy is 0.000000, time 56.282 samples/sec
+epoch 0, total_step 14400, total loss is 39.31 , inference loss is 25.91, weight deacy loss is 13.40, training accuracy is 0.000000, time 56.622 samples/sec
+epoch 0, total_step 14420, total loss is 40.02 , inference loss is 26.62, weight deacy loss is 13.40, training accuracy is 0.000000, time 55.710 samples/sec
+epoch 0, total_step 14440, total loss is 35.40 , inference loss is 22.01, weight deacy loss is 13.40, training accuracy is 0.031250, time 56.316 samples/sec
+epoch 0, total_step 14460, total loss is 36.10 , inference loss is 22.70, weight deacy loss is 13.40, training accuracy is 0.000000, time 55.912 samples/sec
+epoch 0, total_step 14480, total loss is 37.18 , inference loss is 23.78, weight deacy loss is 13.40, training accuracy is 0.000000, time 56.341 samples/sec
+epoch 0, total_step 14500, total loss is 38.79 , inference loss is 25.39, weight deacy loss is 13.40, training accuracy is 0.000000, time 56.246 samples/sec
+epoch 0, total_step 14520, total loss is 41.25 , inference loss is 27.85, weight deacy loss is 13.39, training accuracy is 0.000000, time 56.318 samples/sec
+epoch 0, total_step 14540, total loss is 33.56 , inference loss is 20.17, weight deacy loss is 13.39, training accuracy is 0.062500, time 56.071 samples/sec
+epoch 0, total_step 14560, total loss is 38.63 , inference loss is 25.24, weight deacy loss is 13.39, training accuracy is 0.000000, time 55.778 samples/sec
+epoch 0, total_step 14580, total loss is 36.78 , inference loss is 23.39, weight deacy loss is 13.39, training accuracy is 0.000000, time 57.017 samples/sec
+epoch 0, total_step 14600, total loss is 36.98 , inference loss is 23.59, weight deacy loss is 13.39, training accuracy is 0.000000, time 56.341 samples/sec
+epoch 0, total_step 14620, total loss is 36.15 , inference loss is 22.76, weight deacy loss is 13.39, training accuracy is 0.062500, time 55.881 samples/sec
+epoch 0, total_step 14640, total loss is 43.66 , inference loss is 30.28, weight deacy loss is 13.39, training accuracy is 0.000000, time 56.041 samples/sec
+epoch 0, total_step 14660, total loss is 39.65 , inference loss is 26.26, weight deacy loss is 13.38, training accuracy is 0.031250, time 56.304 samples/sec
+epoch 0, total_step 14680, total loss is 42.81 , inference loss is 29.42, weight deacy loss is 13.38, training accuracy is 0.062500, time 54.271 samples/sec
+epoch 0, total_step 14700, total loss is 39.73 , inference loss is 26.35, weight deacy loss is 13.38, training accuracy is 0.062500, time 56.914 samples/sec
+epoch 0, total_step 14720, total loss is 35.50 , inference loss is 22.11, weight deacy loss is 13.38, training accuracy is 0.062500, time 56.000 samples/sec
+epoch 0, total_step 14740, total loss is 40.70 , inference loss is 27.32, weight deacy loss is 13.38, training accuracy is 0.031250, time 56.777 samples/sec
+epoch 0, total_step 14760, total loss is 39.14 , inference loss is 25.76, weight deacy loss is 13.38, training accuracy is 0.031250, time 57.450 samples/sec
+epoch 0, total_step 14780, total loss is 33.53 , inference loss is 20.15, weight deacy loss is 13.38, training accuracy is 0.000000, time 56.940 samples/sec
+epoch 0, total_step 14800, total loss is 39.38 , inference loss is 26.00, weight deacy loss is 13.38, training accuracy is 0.000000, time 54.516 samples/sec
+epoch 0, total_step 14820, total loss is 40.09 , inference loss is 26.71, weight deacy loss is 13.37, training accuracy is 0.000000, time 56.319 samples/sec
+epoch 0, total_step 14840, total loss is 40.44 , inference loss is 27.07, weight deacy loss is 13.37, training accuracy is 0.000000, time 56.032 samples/sec
+epoch 0, total_step 14860, total loss is 36.39 , inference loss is 23.02, weight deacy loss is 13.37, training accuracy is 0.000000, time 56.110 samples/sec
+epoch 0, total_step 14880, total loss is 44.87 , inference loss is 31.50, weight deacy loss is 13.37, training accuracy is 0.000000, time 56.967 samples/sec
+epoch 0, total_step 14900, total loss is 40.41 , inference loss is 27.04, weight deacy loss is 13.37, training accuracy is 0.031250, time 56.327 samples/sec
+epoch 0, total_step 14920, total loss is 40.64 , inference loss is 27.28, weight deacy loss is 13.37, training accuracy is 0.000000, time 56.165 samples/sec
+epoch 0, total_step 14940, total loss is 42.57 , inference loss is 29.21, weight deacy loss is 13.37, training accuracy is 0.000000, time 55.089 samples/sec
+epoch 0, total_step 14960, total loss is 42.40 , inference loss is 29.03, weight deacy loss is 13.37, training accuracy is 0.031250, time 55.940 samples/sec
+epoch 0, total_step 14980, total loss is 39.90 , inference loss is 26.54, weight deacy loss is 13.36, training accuracy is 0.031250, time 55.002 samples/sec
+epoch 0, total_step 15000, total loss is 43.00 , inference loss is 29.64, weight deacy loss is 13.36, training accuracy is 0.000000, time 56.615 samples/sec
+epoch 0, total_step 15020, total loss is 38.24 , inference loss is 24.88, weight deacy loss is 13.36, training accuracy is 0.000000, time 57.004 samples/sec
+epoch 0, total_step 15040, total loss is 37.63 , inference loss is 24.27, weight deacy loss is 13.36, training accuracy is 0.031250, time 56.287 samples/sec
+epoch 0, total_step 15060, total loss is 35.75 , inference loss is 22.39, weight deacy loss is 13.36, training accuracy is 0.000000, time 56.171 samples/sec
+epoch 0, total_step 15080, total loss is 38.38 , inference loss is 25.02, weight deacy loss is 13.36, training accuracy is 0.031250, time 56.727 samples/sec
+epoch 0, total_step 15100, total loss is 42.13 , inference loss is 28.78, weight deacy loss is 13.36, training accuracy is 0.000000, time 56.584 samples/sec
+epoch 0, total_step 15120, total loss is 33.36 , inference loss is 20.00, weight deacy loss is 13.35, training accuracy is 0.031250, time 57.274 samples/sec
+epoch 0, total_step 15140, total loss is 35.55 , inference loss is 22.19, weight deacy loss is 13.35, training accuracy is 0.062500, time 55.393 samples/sec
+epoch 0, total_step 15160, total loss is 35.44 , inference loss is 22.09, weight deacy loss is 13.35, training accuracy is 0.000000, time 56.236 samples/sec
+epoch 0, total_step 15180, total loss is 37.37 , inference loss is 24.02, weight deacy loss is 13.35, training accuracy is 0.000000, time 54.231 samples/sec
+epoch 0, total_step 15200, total loss is 37.12 , inference loss is 23.77, weight deacy loss is 13.35, training accuracy is 0.031250, time 55.067 samples/sec
+epoch 0, total_step 15220, total loss is 36.11 , inference loss is 22.77, weight deacy loss is 13.34, training accuracy is 0.000000, time 54.223 samples/sec
+epoch 0, total_step 15240, total loss is 33.81 , inference loss is 20.47, weight deacy loss is 13.34, training accuracy is 0.000000, time 55.944 samples/sec
+epoch 0, total_step 15260, total loss is 35.16 , inference loss is 21.82, weight deacy loss is 13.34, training accuracy is 0.031250, time 56.126 samples/sec
+epoch 0, total_step 15280, total loss is 34.67 , inference loss is 21.33, weight deacy loss is 13.34, training accuracy is 0.000000, time 54.596 samples/sec
+epoch 0, total_step 15300, total loss is 30.87 , inference loss is 17.53, weight deacy loss is 13.34, training accuracy is 0.031250, time 55.557 samples/sec
+epoch 0, total_step 15320, total loss is 36.39 , inference loss is 23.06, weight deacy loss is 13.34, training accuracy is 0.000000, time 55.473 samples/sec
+
+
+
+
+End of epoch 0
+epoch 1, total_step 15340, total loss is 65.26 , inference loss is 51.92, weight deacy loss is 13.33, training accuracy is 0.000000, time 55.935 samples/sec
+epoch 1, total_step 15360, total loss is 47.37 , inference loss is 34.04, weight deacy loss is 13.33, training accuracy is 0.000000, time 56.110 samples/sec
+epoch 1, total_step 15380, total loss is 40.18 , inference loss is 26.85, weight deacy loss is 13.33, training accuracy is 0.062500, time 55.653 samples/sec
+epoch 1, total_step 15400, total loss is 31.85 , inference loss is 18.52, weight deacy loss is 13.33, training accuracy is 0.093750, time 56.028 samples/sec
+epoch 1, total_step 15420, total loss is 36.51 , inference loss is 23.19, weight deacy loss is 13.33, training accuracy is 0.031250, time 56.014 samples/sec
+epoch 1, total_step 15440, total loss is 29.47 , inference loss is 16.14, weight deacy loss is 13.33, training accuracy is 0.093750, time 55.645 samples/sec
+epoch 1, total_step 15460, total loss is 32.71 , inference loss is 19.39, weight deacy loss is 13.32, training accuracy is 0.156250, time 56.164 samples/sec
+epoch 1, total_step 15480, total loss is 31.92 , inference loss is 18.60, weight deacy loss is 13.32, training accuracy is 0.156250, time 55.844 samples/sec
+epoch 1, total_step 15500, total loss is 31.03 , inference loss is 17.71, weight deacy loss is 13.32, training accuracy is 0.187500, time 54.900 samples/sec
+epoch 1, total_step 15520, total loss is 31.04 , inference loss is 17.72, weight deacy loss is 13.32, training accuracy is 0.125000, time 56.868 samples/sec
+epoch 1, total_step 15540, total loss is 35.65 , inference loss is 22.34, weight deacy loss is 13.32, training accuracy is 0.093750, time 56.794 samples/sec
+epoch 1, total_step 15560, total loss is 41.31 , inference loss is 27.99, weight deacy loss is 13.31, training accuracy is 0.125000, time 56.735 samples/sec
+epoch 1, total_step 15580, total loss is 36.79 , inference loss is 23.48, weight deacy loss is 13.31, training accuracy is 0.062500, time 53.146 samples/sec
+epoch 1, total_step 15600, total loss is 32.76 , inference loss is 19.45, weight deacy loss is 13.31, training accuracy is 0.062500, time 56.016 samples/sec
+epoch 1, total_step 15620, total loss is 36.27 , inference loss is 22.96, weight deacy loss is 13.31, training accuracy is 0.062500, time 56.392 samples/sec
+epoch 1, total_step 15640, total loss is 38.56 , inference loss is 25.26, weight deacy loss is 13.31, training accuracy is 0.062500, time 57.170 samples/sec
+epoch 1, total_step 15660, total loss is 35.29 , inference loss is 21.99, weight deacy loss is 13.30, training accuracy is 0.125000, time 56.743 samples/sec
+epoch 1, total_step 15680, total loss is 33.09 , inference loss is 19.79, weight deacy loss is 13.30, training accuracy is 0.156250, time 56.561 samples/sec
+epoch 1, total_step 15700, total loss is 32.89 , inference loss is 19.59, weight deacy loss is 13.30, training accuracy is 0.093750, time 55.815 samples/sec
+epoch 1, total_step 15720, total loss is 31.62 , inference loss is 18.32, weight deacy loss is 13.30, training accuracy is 0.062500, time 56.791 samples/sec
+epoch 1, total_step 15740, total loss is 33.48 , inference loss is 20.18, weight deacy loss is 13.30, training accuracy is 0.125000, time 55.705 samples/sec
+epoch 1, total_step 15760, total loss is 36.67 , inference loss is 23.38, weight deacy loss is 13.29, training accuracy is 0.031250, time 56.553 samples/sec
+epoch 1, total_step 15780, total loss is 31.41 , inference loss is 18.12, weight deacy loss is 13.29, training accuracy is 0.156250, time 56.188 samples/sec
+epoch 1, total_step 15800, total loss is 33.89 , inference loss is 20.60, weight deacy loss is 13.29, training accuracy is 0.062500, time 56.002 samples/sec
+epoch 1, total_step 15820, total loss is 31.08 , inference loss is 17.79, weight deacy loss is 13.29, training accuracy is 0.125000, time 55.587 samples/sec
+epoch 1, total_step 15840, total loss is 29.21 , inference loss is 15.92, weight deacy loss is 13.29, training accuracy is 0.218750, time 56.387 samples/sec
+epoch 1, total_step 15860, total loss is 30.17 , inference loss is 16.89, weight deacy loss is 13.28, training accuracy is 0.125000, time 56.847 samples/sec
+epoch 1, total_step 15880, total loss is 31.46 , inference loss is 18.18, weight deacy loss is 13.28, training accuracy is 0.093750, time 56.419 samples/sec
+epoch 1, total_step 15900, total loss is 36.24 , inference loss is 22.96, weight deacy loss is 13.28, training accuracy is 0.125000, time 56.162 samples/sec
+epoch 1, total_step 15920, total loss is 35.33 , inference loss is 22.05, weight deacy loss is 13.28, training accuracy is 0.156250, time 56.530 samples/sec
+epoch 1, total_step 15940, total loss is 29.35 , inference loss is 16.08, weight deacy loss is 13.28, training accuracy is 0.187500, time 55.563 samples/sec
+epoch 1, total_step 15960, total loss is 36.17 , inference loss is 22.90, weight deacy loss is 13.27, training accuracy is 0.062500, time 56.449 samples/sec
+epoch 1, total_step 15980, total loss is 34.53 , inference loss is 21.26, weight deacy loss is 13.27, training accuracy is 0.000000, time 56.373 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.12825600000006
+best_threshold_index 94 0.9551851851851851
+best_threshold_index 95 0.9555555555555556
+best_threshold_index 93 0.9557407407407408
+best_threshold_index 94 0.9551851851851851
+best_threshold_index 93 0.9575925925925926
+best_threshold_index 94 0.9542592592592593
+best_threshold_index 94 0.9555555555555556
+best_threshold_index 94 0.955925925925926
+best_threshold_index 93 0.9542592592592593
+best_threshold_index 94 0.9546296296296296
+[lfw][16000]XNorm: 25.147690
+[lfw][16000]Accuracy-Flip: 0.95417+-0.00864
+testing verification..
+(14000, 512)
+infer time 86.22780500000009
+best_threshold_index 94 0.9423809523809524
+best_threshold_index 94 0.9396825396825397
+best_threshold_index 94 0.940952380952381
+best_threshold_index 94 0.9401587301587302
+best_threshold_index 94 0.9396825396825397
+best_threshold_index 94 0.9385714285714286
+best_threshold_index 94 0.9403174603174603
+best_threshold_index 94 0.9412698412698413
+best_threshold_index 94 0.9393650793650794
+best_threshold_index 94 0.9433333333333334
+[cfp_ff][16000]XNorm: 25.132904
+[cfp_ff][16000]Accuracy-Flip: 0.94057+-0.01234
+testing verification..
+(14000, 512)
+infer time 85.772344
+best_threshold_index 106 0.8023809523809524
+best_threshold_index 104 0.8025396825396826
+best_threshold_index 104 0.8001587301587302
+best_threshold_index 104 0.802063492063492
+best_threshold_index 104 0.8052380952380952
+best_threshold_index 104 0.7984126984126985
+best_threshold_index 104 0.7998412698412698
+best_threshold_index 104 0.8019047619047619
+best_threshold_index 104 0.8026984126984127
+best_threshold_index 104 0.8007936507936508
+[cfp_fp][16000]XNorm: 24.046680
+[cfp_fp][16000]Accuracy-Flip: 0.80029+-0.01685
+testing verification..
+(12000, 512)
+infer time 73.99148700000002
+best_threshold_index 109 0.7844444444444445
+best_threshold_index 109 0.7838888888888889
+best_threshold_index 112 0.7875925925925926
+best_threshold_index 109 0.7868518518518518
+best_threshold_index 112 0.7927777777777778
+best_threshold_index 109 0.7844444444444445
+best_threshold_index 109 0.7887037037037037
+best_threshold_index 109 0.7866666666666666
+best_threshold_index 109 0.787962962962963
+best_threshold_index 112 0.7881481481481482
+[agedb_30][16000]XNorm: 24.528756
+[agedb_30][16000]Accuracy-Flip: 0.78517+-0.02281
+test accuracy is: 0.9541666666666666
+epoch 1, total_step 16000, total loss is 33.73 , inference loss is 20.46, weight deacy loss is 13.27, training accuracy is 0.062500, time 53.488 samples/sec
+epoch 1, total_step 16020, total loss is 31.29 , inference loss is 18.02, weight deacy loss is 13.27, training accuracy is 0.062500, time 56.801 samples/sec
+epoch 1, total_step 16040, total loss is 35.19 , inference loss is 21.92, weight deacy loss is 13.27, training accuracy is 0.125000, time 56.030 samples/sec
+epoch 1, total_step 16060, total loss is 33.54 , inference loss is 20.27, weight deacy loss is 13.26, training accuracy is 0.093750, time 56.110 samples/sec
+epoch 1, total_step 16080, total loss is 37.86 , inference loss is 24.60, weight deacy loss is 13.26, training accuracy is 0.062500, time 56.826 samples/sec
+epoch 1, total_step 16100, total loss is 33.08 , inference loss is 19.82, weight deacy loss is 13.26, training accuracy is 0.093750, time 56.875 samples/sec
+epoch 1, total_step 16120, total loss is 38.90 , inference loss is 25.64, weight deacy loss is 13.26, training accuracy is 0.031250, time 57.098 samples/sec
+epoch 1, total_step 16140, total loss is 35.64 , inference loss is 22.38, weight deacy loss is 13.26, training accuracy is 0.125000, time 55.804 samples/sec
+epoch 1, total_step 16160, total loss is 36.59 , inference loss is 23.34, weight deacy loss is 13.26, training accuracy is 0.000000, time 56.620 samples/sec
+epoch 1, total_step 16180, total loss is 31.79 , inference loss is 18.53, weight deacy loss is 13.25, training accuracy is 0.093750, time 55.734 samples/sec
+epoch 1, total_step 16200, total loss is 35.22 , inference loss is 21.97, weight deacy loss is 13.25, training accuracy is 0.093750, time 56.295 samples/sec
+epoch 1, total_step 16220, total loss is 32.41 , inference loss is 19.16, weight deacy loss is 13.25, training accuracy is 0.125000, time 56.170 samples/sec
+epoch 1, total_step 16240, total loss is 31.02 , inference loss is 17.77, weight deacy loss is 13.25, training accuracy is 0.125000, time 55.445 samples/sec
+epoch 1, total_step 16260, total loss is 37.09 , inference loss is 23.84, weight deacy loss is 13.25, training accuracy is 0.031250, time 56.148 samples/sec
+epoch 1, total_step 16280, total loss is 37.27 , inference loss is 24.02, weight deacy loss is 13.24, training accuracy is 0.125000, time 56.592 samples/sec
+epoch 1, total_step 16300, total loss is 31.00 , inference loss is 17.76, weight deacy loss is 13.24, training accuracy is 0.093750, time 55.820 samples/sec
+epoch 1, total_step 16320, total loss is 33.44 , inference loss is 20.20, weight deacy loss is 13.24, training accuracy is 0.093750, time 57.369 samples/sec
+epoch 1, total_step 16340, total loss is 40.31 , inference loss is 27.07, weight deacy loss is 13.24, training accuracy is 0.031250, time 55.936 samples/sec
+epoch 1, total_step 16360, total loss is 29.68 , inference loss is 16.45, weight deacy loss is 13.24, training accuracy is 0.125000, time 53.142 samples/sec
+epoch 1, total_step 16380, total loss is 34.93 , inference loss is 21.70, weight deacy loss is 13.23, training accuracy is 0.000000, time 56.523 samples/sec
+epoch 1, total_step 16400, total loss is 31.81 , inference loss is 18.58, weight deacy loss is 13.23, training accuracy is 0.062500, time 56.025 samples/sec
+epoch 1, total_step 16420, total loss is 33.42 , inference loss is 20.19, weight deacy loss is 13.23, training accuracy is 0.062500, time 56.089 samples/sec
+epoch 1, total_step 16440, total loss is 39.29 , inference loss is 26.06, weight deacy loss is 13.23, training accuracy is 0.031250, time 55.633 samples/sec
+epoch 1, total_step 16460, total loss is 33.29 , inference loss is 20.06, weight deacy loss is 13.23, training accuracy is 0.156250, time 56.527 samples/sec
+epoch 1, total_step 16480, total loss is 37.66 , inference loss is 24.44, weight deacy loss is 13.22, training accuracy is 0.031250, time 56.188 samples/sec
+epoch 1, total_step 16500, total loss is 39.17 , inference loss is 25.94, weight deacy loss is 13.22, training accuracy is 0.031250, time 56.275 samples/sec
+epoch 1, total_step 16520, total loss is 29.41 , inference loss is 16.19, weight deacy loss is 13.22, training accuracy is 0.062500, time 54.842 samples/sec
+epoch 1, total_step 16540, total loss is 33.56 , inference loss is 20.34, weight deacy loss is 13.22, training accuracy is 0.125000, time 56.215 samples/sec
+epoch 1, total_step 16560, total loss is 34.74 , inference loss is 21.52, weight deacy loss is 13.22, training accuracy is 0.125000, time 55.881 samples/sec
+epoch 1, total_step 16580, total loss is 32.94 , inference loss is 19.72, weight deacy loss is 13.21, training accuracy is 0.062500, time 55.865 samples/sec
+epoch 1, total_step 16600, total loss is 33.50 , inference loss is 20.28, weight deacy loss is 13.21, training accuracy is 0.093750, time 56.052 samples/sec
+epoch 1, total_step 16620, total loss is 30.94 , inference loss is 17.73, weight deacy loss is 13.21, training accuracy is 0.156250, time 55.206 samples/sec
+epoch 1, total_step 16640, total loss is 35.23 , inference loss is 22.02, weight deacy loss is 13.21, training accuracy is 0.062500, time 56.210 samples/sec
+epoch 1, total_step 16660, total loss is 38.78 , inference loss is 25.57, weight deacy loss is 13.21, training accuracy is 0.093750, time 55.094 samples/sec
+epoch 1, total_step 16680, total loss is 35.47 , inference loss is 22.27, weight deacy loss is 13.21, training accuracy is 0.125000, time 55.913 samples/sec
+epoch 1, total_step 16700, total loss is 40.06 , inference loss is 26.86, weight deacy loss is 13.20, training accuracy is 0.000000, time 56.196 samples/sec
+epoch 1, total_step 16720, total loss is 33.35 , inference loss is 20.15, weight deacy loss is 13.20, training accuracy is 0.031250, time 55.833 samples/sec
+epoch 1, total_step 16740, total loss is 35.10 , inference loss is 21.90, weight deacy loss is 13.20, training accuracy is 0.062500, time 56.038 samples/sec
+epoch 1, total_step 16760, total loss is 30.91 , inference loss is 17.71, weight deacy loss is 13.20, training accuracy is 0.062500, time 54.514 samples/sec
+epoch 1, total_step 16780, total loss is 32.04 , inference loss is 18.85, weight deacy loss is 13.20, training accuracy is 0.093750, time 56.364 samples/sec
+epoch 1, total_step 16800, total loss is 29.61 , inference loss is 16.42, weight deacy loss is 13.19, training accuracy is 0.125000, time 53.801 samples/sec
+epoch 1, total_step 16820, total loss is 32.17 , inference loss is 18.98, weight deacy loss is 13.19, training accuracy is 0.031250, time 55.683 samples/sec
+epoch 1, total_step 16840, total loss is 37.14 , inference loss is 23.95, weight deacy loss is 13.19, training accuracy is 0.031250, time 56.433 samples/sec
+epoch 1, total_step 16860, total loss is 35.73 , inference loss is 22.54, weight deacy loss is 13.19, training accuracy is 0.062500, time 56.192 samples/sec
+epoch 1, total_step 16880, total loss is 33.22 , inference loss is 20.03, weight deacy loss is 13.19, training accuracy is 0.156250, time 56.349 samples/sec
+epoch 1, total_step 16900, total loss is 34.69 , inference loss is 21.50, weight deacy loss is 13.18, training accuracy is 0.062500, time 56.003 samples/sec
+epoch 1, total_step 16920, total loss is 32.98 , inference loss is 19.80, weight deacy loss is 13.18, training accuracy is 0.093750, time 57.352 samples/sec
+epoch 1, total_step 16940, total loss is 36.38 , inference loss is 23.20, weight deacy loss is 13.18, training accuracy is 0.031250, time 56.803 samples/sec
+epoch 1, total_step 16960, total loss is 37.17 , inference loss is 23.99, weight deacy loss is 13.18, training accuracy is 0.125000, time 57.376 samples/sec
+epoch 1, total_step 16980, total loss is 36.86 , inference loss is 23.68, weight deacy loss is 13.18, training accuracy is 0.062500, time 55.509 samples/sec
+epoch 1, total_step 17000, total loss is 33.99 , inference loss is 20.81, weight deacy loss is 13.17, training accuracy is 0.093750, time 52.919 samples/sec
+epoch 1, total_step 17020, total loss is 35.25 , inference loss is 22.07, weight deacy loss is 13.17, training accuracy is 0.062500, time 55.633 samples/sec
+epoch 1, total_step 17040, total loss is 33.97 , inference loss is 20.80, weight deacy loss is 13.17, training accuracy is 0.062500, time 56.245 samples/sec
+epoch 1, total_step 17060, total loss is 33.32 , inference loss is 20.15, weight deacy loss is 13.17, training accuracy is 0.156250, time 56.409 samples/sec
+epoch 1, total_step 17080, total loss is 33.46 , inference loss is 20.29, weight deacy loss is 13.17, training accuracy is 0.093750, time 53.092 samples/sec
+epoch 1, total_step 17100, total loss is 36.79 , inference loss is 23.62, weight deacy loss is 13.17, training accuracy is 0.000000, time 54.361 samples/sec
+epoch 1, total_step 17120, total loss is 37.48 , inference loss is 24.32, weight deacy loss is 13.16, training accuracy is 0.062500, time 56.133 samples/sec
+epoch 1, total_step 17140, total loss is 32.51 , inference loss is 19.35, weight deacy loss is 13.16, training accuracy is 0.093750, time 57.552 samples/sec
+epoch 1, total_step 17160, total loss is 32.77 , inference loss is 19.61, weight deacy loss is 13.16, training accuracy is 0.093750, time 56.761 samples/sec
+epoch 1, total_step 17180, total loss is 36.23 , inference loss is 23.07, weight deacy loss is 13.16, training accuracy is 0.031250, time 55.880 samples/sec
+epoch 1, total_step 17200, total loss is 25.27 , inference loss is 12.11, weight deacy loss is 13.16, training accuracy is 0.250000, time 55.364 samples/sec
+epoch 1, total_step 17220, total loss is 30.14 , inference loss is 16.98, weight deacy loss is 13.15, training accuracy is 0.187500, time 56.554 samples/sec
+epoch 1, total_step 17240, total loss is 31.22 , inference loss is 18.07, weight deacy loss is 13.15, training accuracy is 0.125000, time 56.256 samples/sec
+epoch 1, total_step 17260, total loss is 33.21 , inference loss is 20.06, weight deacy loss is 13.15, training accuracy is 0.156250, time 54.842 samples/sec
+epoch 1, total_step 17280, total loss is 33.06 , inference loss is 19.92, weight deacy loss is 13.15, training accuracy is 0.125000, time 56.392 samples/sec
+epoch 1, total_step 17300, total loss is 31.01 , inference loss is 17.87, weight deacy loss is 13.15, training accuracy is 0.218750, time 52.407 samples/sec
+epoch 1, total_step 17320, total loss is 30.25 , inference loss is 17.11, weight deacy loss is 13.14, training accuracy is 0.218750, time 56.446 samples/sec
+epoch 1, total_step 17340, total loss is 29.92 , inference loss is 16.78, weight deacy loss is 13.14, training accuracy is 0.250000, time 56.503 samples/sec
+epoch 1, total_step 17360, total loss is 32.65 , inference loss is 19.51, weight deacy loss is 13.14, training accuracy is 0.093750, time 53.805 samples/sec
+epoch 1, total_step 17380, total loss is 36.10 , inference loss is 22.96, weight deacy loss is 13.14, training accuracy is 0.062500, time 56.824 samples/sec
+epoch 1, total_step 17400, total loss is 29.63 , inference loss is 16.49, weight deacy loss is 13.14, training accuracy is 0.125000, time 56.099 samples/sec
+epoch 1, total_step 17420, total loss is 32.21 , inference loss is 19.07, weight deacy loss is 13.13, training accuracy is 0.093750, time 56.625 samples/sec
+epoch 1, total_step 17440, total loss is 32.19 , inference loss is 19.05, weight deacy loss is 13.13, training accuracy is 0.125000, time 55.654 samples/sec
+epoch 1, total_step 17460, total loss is 37.13 , inference loss is 24.00, weight deacy loss is 13.13, training accuracy is 0.062500, time 55.968 samples/sec
+epoch 1, total_step 17480, total loss is 32.20 , inference loss is 19.07, weight deacy loss is 13.13, training accuracy is 0.062500, time 56.438 samples/sec
+epoch 1, total_step 17500, total loss is 31.24 , inference loss is 18.11, weight deacy loss is 13.13, training accuracy is 0.218750, time 56.397 samples/sec
+epoch 1, total_step 17520, total loss is 30.91 , inference loss is 17.79, weight deacy loss is 13.12, training accuracy is 0.187500, time 55.881 samples/sec
+epoch 1, total_step 17540, total loss is 26.46 , inference loss is 13.34, weight deacy loss is 13.12, training accuracy is 0.187500, time 56.311 samples/sec
+epoch 1, total_step 17560, total loss is 30.49 , inference loss is 17.37, weight deacy loss is 13.12, training accuracy is 0.187500, time 55.821 samples/sec
+epoch 1, total_step 17580, total loss is 29.21 , inference loss is 16.09, weight deacy loss is 13.12, training accuracy is 0.125000, time 56.457 samples/sec
+epoch 1, total_step 17600, total loss is 28.48 , inference loss is 15.36, weight deacy loss is 13.12, training accuracy is 0.250000, time 55.977 samples/sec
+epoch 1, total_step 17620, total loss is 31.95 , inference loss is 18.84, weight deacy loss is 13.12, training accuracy is 0.031250, time 55.922 samples/sec
+epoch 1, total_step 17640, total loss is 27.17 , inference loss is 14.06, weight deacy loss is 13.11, training accuracy is 0.187500, time 56.160 samples/sec
+epoch 1, total_step 17660, total loss is 28.31 , inference loss is 15.20, weight deacy loss is 13.11, training accuracy is 0.125000, time 56.232 samples/sec
+epoch 1, total_step 17680, total loss is 31.16 , inference loss is 18.05, weight deacy loss is 13.11, training accuracy is 0.031250, time 56.147 samples/sec
+epoch 1, total_step 17700, total loss is 29.03 , inference loss is 15.92, weight deacy loss is 13.11, training accuracy is 0.093750, time 54.741 samples/sec
+epoch 1, total_step 17720, total loss is 31.53 , inference loss is 18.42, weight deacy loss is 13.11, training accuracy is 0.187500, time 56.331 samples/sec
+epoch 1, total_step 17740, total loss is 30.89 , inference loss is 17.79, weight deacy loss is 13.10, training accuracy is 0.093750, time 55.589 samples/sec
+epoch 1, total_step 17760, total loss is 26.62 , inference loss is 13.52, weight deacy loss is 13.10, training accuracy is 0.062500, time 56.089 samples/sec
+epoch 1, total_step 17780, total loss is 32.23 , inference loss is 19.13, weight deacy loss is 13.10, training accuracy is 0.125000, time 55.291 samples/sec
+epoch 1, total_step 17800, total loss is 32.24 , inference loss is 19.14, weight deacy loss is 13.10, training accuracy is 0.062500, time 56.397 samples/sec
+epoch 1, total_step 17820, total loss is 28.87 , inference loss is 15.77, weight deacy loss is 13.10, training accuracy is 0.187500, time 56.415 samples/sec
+epoch 1, total_step 17840, total loss is 30.41 , inference loss is 17.32, weight deacy loss is 13.09, training accuracy is 0.125000, time 56.092 samples/sec
+epoch 1, total_step 17860, total loss is 28.03 , inference loss is 14.94, weight deacy loss is 13.09, training accuracy is 0.187500, time 55.967 samples/sec
+epoch 1, total_step 17880, total loss is 27.42 , inference loss is 14.33, weight deacy loss is 13.09, training accuracy is 0.156250, time 56.574 samples/sec
+epoch 1, total_step 17900, total loss is 33.01 , inference loss is 19.92, weight deacy loss is 13.09, training accuracy is 0.156250, time 56.353 samples/sec
+epoch 1, total_step 17920, total loss is 26.48 , inference loss is 13.39, weight deacy loss is 13.09, training accuracy is 0.156250, time 56.370 samples/sec
+epoch 1, total_step 17940, total loss is 27.80 , inference loss is 14.71, weight deacy loss is 13.08, training accuracy is 0.218750, time 54.905 samples/sec
+epoch 1, total_step 17960, total loss is 29.99 , inference loss is 16.91, weight deacy loss is 13.08, training accuracy is 0.156250, time 56.599 samples/sec
+epoch 1, total_step 17980, total loss is 28.35 , inference loss is 15.27, weight deacy loss is 13.08, training accuracy is 0.093750, time 56.275 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.80504300000008
+best_threshold_index 106 0.9592592592592593
+best_threshold_index 102 0.9612962962962963
+best_threshold_index 102 0.9618518518518518
+best_threshold_index 102 0.9622222222222222
+best_threshold_index 102 0.9637037037037037
+best_threshold_index 102 0.960925925925926
+best_threshold_index 102 0.9607407407407408
+best_threshold_index 102 0.9607407407407408
+best_threshold_index 102 0.9594444444444444
+best_threshold_index 102 0.96
+[lfw][18000]XNorm: 23.991722
+[lfw][18000]Accuracy-Flip: 0.95983+-0.01042
+testing verification..
+(14000, 512)
+infer time 85.96094800000006
+best_threshold_index 108 0.9568253968253968
+best_threshold_index 108 0.954920634920635
+best_threshold_index 110 0.9534920634920635
+best_threshold_index 107 0.9534920634920635
+best_threshold_index 107 0.9542857142857143
+best_threshold_index 107 0.9534920634920635
+best_threshold_index 107 0.9544444444444444
+best_threshold_index 108 0.9547619047619048
+best_threshold_index 107 0.9546031746031746
+best_threshold_index 107 0.9547619047619048
+[cfp_ff][18000]XNorm: 23.107365
+[cfp_ff][18000]Accuracy-Flip: 0.95357+-0.00831
+testing verification..
+(14000, 512)
+infer time 85.79858100000004
+best_threshold_index 119 0.8119047619047619
+best_threshold_index 122 0.8128571428571428
+best_threshold_index 121 0.8111111111111111
+best_threshold_index 119 0.812063492063492
+best_threshold_index 124 0.816984126984127
+best_threshold_index 119 0.8076190476190476
+best_threshold_index 122 0.8092063492063492
+best_threshold_index 119 0.8096825396825397
+best_threshold_index 122 0.8152380952380952
+best_threshold_index 122 0.8138095238095238
+[cfp_fp][18000]XNorm: 24.738692
+[cfp_fp][18000]Accuracy-Flip: 0.80743+-0.02563
+testing verification..
+(12000, 512)
+infer time 74.40662399999994
+best_threshold_index 129 0.8022222222222222
+best_threshold_index 127 0.8027777777777778
+best_threshold_index 127 0.8059259259259259
+best_threshold_index 125 0.8027777777777778
+best_threshold_index 126 0.8062962962962963
+best_threshold_index 131 0.8005555555555556
+best_threshold_index 125 0.8051851851851852
+best_threshold_index 127 0.8096296296296296
+best_threshold_index 129 0.8055555555555556
+best_threshold_index 129 0.8057407407407408
+[agedb_30][18000]XNorm: 23.503063
+[agedb_30][18000]Accuracy-Flip: 0.79933+-0.02102
+test accuracy is: 0.9598333333333334
+epoch 1, total_step 18000, total loss is 29.88 , inference loss is 16.80, weight deacy loss is 13.08, training accuracy is 0.093750, time 52.370 samples/sec
+epoch 1, total_step 18020, total loss is 29.03 , inference loss is 15.95, weight deacy loss is 13.08, training accuracy is 0.093750, time 55.795 samples/sec
+epoch 1, total_step 18040, total loss is 30.36 , inference loss is 17.28, weight deacy loss is 13.07, training accuracy is 0.250000, time 57.394 samples/sec
+epoch 1, total_step 18060, total loss is 25.53 , inference loss is 12.46, weight deacy loss is 13.07, training accuracy is 0.250000, time 56.979 samples/sec
+epoch 1, total_step 18080, total loss is 35.58 , inference loss is 22.51, weight deacy loss is 13.07, training accuracy is 0.062500, time 55.831 samples/sec
+epoch 1, total_step 18100, total loss is 27.69 , inference loss is 14.62, weight deacy loss is 13.07, training accuracy is 0.156250, time 56.574 samples/sec
+epoch 1, total_step 18120, total loss is 30.10 , inference loss is 17.03, weight deacy loss is 13.07, training accuracy is 0.125000, time 55.498 samples/sec
+epoch 1, total_step 18140, total loss is 32.23 , inference loss is 19.17, weight deacy loss is 13.07, training accuracy is 0.156250, time 56.894 samples/sec
+epoch 1, total_step 18160, total loss is 30.97 , inference loss is 17.91, weight deacy loss is 13.06, training accuracy is 0.093750, time 56.242 samples/sec
+epoch 1, total_step 18180, total loss is 29.82 , inference loss is 16.75, weight deacy loss is 13.06, training accuracy is 0.031250, time 56.167 samples/sec
+epoch 1, total_step 18200, total loss is 31.66 , inference loss is 18.60, weight deacy loss is 13.06, training accuracy is 0.156250, time 56.077 samples/sec
+epoch 1, total_step 18220, total loss is 31.81 , inference loss is 18.75, weight deacy loss is 13.06, training accuracy is 0.093750, time 56.071 samples/sec
+epoch 1, total_step 18240, total loss is 28.02 , inference loss is 14.96, weight deacy loss is 13.06, training accuracy is 0.250000, time 56.367 samples/sec
+epoch 1, total_step 18260, total loss is 30.33 , inference loss is 17.27, weight deacy loss is 13.05, training accuracy is 0.156250, time 56.138 samples/sec
+epoch 1, total_step 18280, total loss is 29.11 , inference loss is 16.05, weight deacy loss is 13.05, training accuracy is 0.125000, time 56.676 samples/sec
+epoch 1, total_step 18300, total loss is 32.48 , inference loss is 19.43, weight deacy loss is 13.05, training accuracy is 0.125000, time 55.521 samples/sec
+epoch 1, total_step 18320, total loss is 30.49 , inference loss is 17.44, weight deacy loss is 13.05, training accuracy is 0.218750, time 55.940 samples/sec
+epoch 1, total_step 18340, total loss is 27.43 , inference loss is 14.38, weight deacy loss is 13.05, training accuracy is 0.218750, time 56.484 samples/sec
+epoch 1, total_step 18360, total loss is 27.43 , inference loss is 14.38, weight deacy loss is 13.04, training accuracy is 0.187500, time 55.993 samples/sec
+epoch 1, total_step 18380, total loss is 30.90 , inference loss is 17.86, weight deacy loss is 13.04, training accuracy is 0.125000, time 56.134 samples/sec
+epoch 1, total_step 18400, total loss is 26.26 , inference loss is 13.22, weight deacy loss is 13.04, training accuracy is 0.218750, time 56.685 samples/sec
+epoch 1, total_step 18420, total loss is 27.89 , inference loss is 14.85, weight deacy loss is 13.04, training accuracy is 0.187500, time 56.430 samples/sec
+epoch 1, total_step 18440, total loss is 35.17 , inference loss is 22.13, weight deacy loss is 13.04, training accuracy is 0.125000, time 56.421 samples/sec
+epoch 1, total_step 18460, total loss is 27.12 , inference loss is 14.09, weight deacy loss is 13.04, training accuracy is 0.312500, time 56.290 samples/sec
+epoch 1, total_step 18480, total loss is 29.41 , inference loss is 16.37, weight deacy loss is 13.03, training accuracy is 0.187500, time 56.135 samples/sec
+epoch 1, total_step 18500, total loss is 31.49 , inference loss is 18.45, weight deacy loss is 13.03, training accuracy is 0.281250, time 56.491 samples/sec
+epoch 1, total_step 18520, total loss is 30.53 , inference loss is 17.50, weight deacy loss is 13.03, training accuracy is 0.187500, time 56.129 samples/sec
+epoch 1, total_step 18540, total loss is 35.58 , inference loss is 22.56, weight deacy loss is 13.03, training accuracy is 0.093750, time 56.320 samples/sec
+epoch 1, total_step 18560, total loss is 24.05 , inference loss is 11.02, weight deacy loss is 13.03, training accuracy is 0.250000, time 56.341 samples/sec
+epoch 1, total_step 18580, total loss is 30.97 , inference loss is 17.94, weight deacy loss is 13.02, training accuracy is 0.187500, time 56.175 samples/sec
+epoch 1, total_step 18600, total loss is 34.28 , inference loss is 21.26, weight deacy loss is 13.02, training accuracy is 0.125000, time 56.286 samples/sec
+epoch 1, total_step 18620, total loss is 28.46 , inference loss is 15.44, weight deacy loss is 13.02, training accuracy is 0.156250, time 55.824 samples/sec
+epoch 1, total_step 18640, total loss is 30.17 , inference loss is 17.15, weight deacy loss is 13.02, training accuracy is 0.093750, time 56.366 samples/sec
+epoch 1, total_step 18660, total loss is 33.10 , inference loss is 20.08, weight deacy loss is 13.02, training accuracy is 0.125000, time 55.786 samples/sec
+epoch 1, total_step 18680, total loss is 33.27 , inference loss is 20.26, weight deacy loss is 13.01, training accuracy is 0.125000, time 56.011 samples/sec
+epoch 1, total_step 18700, total loss is 31.14 , inference loss is 18.13, weight deacy loss is 13.01, training accuracy is 0.093750, time 57.230 samples/sec
+epoch 1, total_step 18720, total loss is 30.89 , inference loss is 17.87, weight deacy loss is 13.01, training accuracy is 0.093750, time 56.622 samples/sec
+epoch 1, total_step 18740, total loss is 29.75 , inference loss is 16.74, weight deacy loss is 13.01, training accuracy is 0.187500, time 56.071 samples/sec
+epoch 1, total_step 18760, total loss is 29.74 , inference loss is 16.74, weight deacy loss is 13.01, training accuracy is 0.125000, time 56.266 samples/sec
+epoch 1, total_step 18780, total loss is 28.63 , inference loss is 15.62, weight deacy loss is 13.01, training accuracy is 0.156250, time 56.130 samples/sec
+epoch 1, total_step 18800, total loss is 30.35 , inference loss is 17.35, weight deacy loss is 13.00, training accuracy is 0.093750, time 56.370 samples/sec
+epoch 1, total_step 18820, total loss is 34.52 , inference loss is 21.52, weight deacy loss is 13.00, training accuracy is 0.031250, time 56.195 samples/sec
+epoch 1, total_step 18840, total loss is 32.11 , inference loss is 19.11, weight deacy loss is 13.00, training accuracy is 0.093750, time 55.924 samples/sec
+epoch 1, total_step 18860, total loss is 28.66 , inference loss is 15.67, weight deacy loss is 13.00, training accuracy is 0.156250, time 56.080 samples/sec
+epoch 1, total_step 18880, total loss is 31.60 , inference loss is 18.60, weight deacy loss is 13.00, training accuracy is 0.062500, time 56.150 samples/sec
+epoch 1, total_step 18900, total loss is 30.80 , inference loss is 17.80, weight deacy loss is 12.99, training accuracy is 0.156250, time 56.910 samples/sec
+epoch 1, total_step 18920, total loss is 29.58 , inference loss is 16.58, weight deacy loss is 12.99, training accuracy is 0.093750, time 55.854 samples/sec
+epoch 1, total_step 18940, total loss is 34.64 , inference loss is 21.65, weight deacy loss is 12.99, training accuracy is 0.125000, time 55.274 samples/sec
+epoch 1, total_step 18960, total loss is 30.05 , inference loss is 17.06, weight deacy loss is 12.99, training accuracy is 0.187500, time 55.409 samples/sec
+epoch 1, total_step 18980, total loss is 31.94 , inference loss is 18.95, weight deacy loss is 12.99, training accuracy is 0.093750, time 55.684 samples/sec
+epoch 1, total_step 19000, total loss is 32.46 , inference loss is 19.47, weight deacy loss is 12.99, training accuracy is 0.062500, time 55.724 samples/sec
+epoch 1, total_step 19020, total loss is 30.41 , inference loss is 17.42, weight deacy loss is 12.98, training accuracy is 0.156250, time 56.128 samples/sec
+epoch 1, total_step 19040, total loss is 33.86 , inference loss is 20.88, weight deacy loss is 12.98, training accuracy is 0.093750, time 56.337 samples/sec
+epoch 1, total_step 19060, total loss is 33.82 , inference loss is 20.84, weight deacy loss is 12.98, training accuracy is 0.125000, time 55.776 samples/sec
+epoch 1, total_step 19080, total loss is 30.58 , inference loss is 17.60, weight deacy loss is 12.98, training accuracy is 0.156250, time 56.160 samples/sec
+epoch 1, total_step 19100, total loss is 30.48 , inference loss is 17.51, weight deacy loss is 12.98, training accuracy is 0.125000, time 55.448 samples/sec
+epoch 1, total_step 19120, total loss is 33.96 , inference loss is 20.98, weight deacy loss is 12.97, training accuracy is 0.031250, time 55.184 samples/sec
+epoch 1, total_step 19140, total loss is 33.73 , inference loss is 20.76, weight deacy loss is 12.97, training accuracy is 0.062500, time 57.199 samples/sec
+epoch 1, total_step 19160, total loss is 31.99 , inference loss is 19.02, weight deacy loss is 12.97, training accuracy is 0.093750, time 56.124 samples/sec
+epoch 1, total_step 19180, total loss is 31.02 , inference loss is 18.05, weight deacy loss is 12.97, training accuracy is 0.156250, time 57.228 samples/sec
+epoch 1, total_step 19200, total loss is 32.61 , inference loss is 19.64, weight deacy loss is 12.97, training accuracy is 0.156250, time 56.175 samples/sec
+epoch 1, total_step 19220, total loss is 32.53 , inference loss is 19.56, weight deacy loss is 12.97, training accuracy is 0.062500, time 56.728 samples/sec
+epoch 1, total_step 19240, total loss is 33.39 , inference loss is 20.43, weight deacy loss is 12.96, training accuracy is 0.093750, time 55.393 samples/sec
+epoch 1, total_step 19260, total loss is 35.14 , inference loss is 22.18, weight deacy loss is 12.96, training accuracy is 0.031250, time 55.847 samples/sec
+epoch 1, total_step 19280, total loss is 35.45 , inference loss is 22.49, weight deacy loss is 12.96, training accuracy is 0.062500, time 56.435 samples/sec
+epoch 1, total_step 19300, total loss is 35.92 , inference loss is 22.96, weight deacy loss is 12.96, training accuracy is 0.062500, time 56.271 samples/sec
+epoch 1, total_step 19320, total loss is 33.24 , inference loss is 20.28, weight deacy loss is 12.96, training accuracy is 0.062500, time 55.998 samples/sec
+epoch 1, total_step 19340, total loss is 30.56 , inference loss is 17.61, weight deacy loss is 12.95, training accuracy is 0.062500, time 56.052 samples/sec
+epoch 1, total_step 19360, total loss is 37.02 , inference loss is 24.07, weight deacy loss is 12.95, training accuracy is 0.000000, time 56.448 samples/sec
+epoch 1, total_step 19380, total loss is 30.43 , inference loss is 17.48, weight deacy loss is 12.95, training accuracy is 0.031250, time 56.458 samples/sec
+epoch 1, total_step 19400, total loss is 34.45 , inference loss is 21.50, weight deacy loss is 12.95, training accuracy is 0.031250, time 52.961 samples/sec
+epoch 1, total_step 19420, total loss is 36.36 , inference loss is 23.41, weight deacy loss is 12.95, training accuracy is 0.093750, time 56.076 samples/sec
+epoch 1, total_step 19440, total loss is 29.14 , inference loss is 16.19, weight deacy loss is 12.95, training accuracy is 0.125000, time 56.678 samples/sec
+epoch 1, total_step 19460, total loss is 32.20 , inference loss is 19.25, weight deacy loss is 12.94, training accuracy is 0.031250, time 56.509 samples/sec
+epoch 1, total_step 19480, total loss is 28.64 , inference loss is 15.70, weight deacy loss is 12.94, training accuracy is 0.093750, time 56.950 samples/sec
+epoch 1, total_step 19500, total loss is 33.36 , inference loss is 20.42, weight deacy loss is 12.94, training accuracy is 0.187500, time 56.300 samples/sec
+epoch 1, total_step 19520, total loss is 33.07 , inference loss is 20.13, weight deacy loss is 12.94, training accuracy is 0.062500, time 55.492 samples/sec
+epoch 1, total_step 19540, total loss is 34.84 , inference loss is 21.91, weight deacy loss is 12.94, training accuracy is 0.031250, time 57.197 samples/sec
+epoch 1, total_step 19560, total loss is 28.90 , inference loss is 15.96, weight deacy loss is 12.93, training accuracy is 0.125000, time 56.724 samples/sec
+epoch 1, total_step 19580, total loss is 34.17 , inference loss is 21.23, weight deacy loss is 12.93, training accuracy is 0.031250, time 56.858 samples/sec
+epoch 1, total_step 19600, total loss is 36.82 , inference loss is 23.89, weight deacy loss is 12.93, training accuracy is 0.031250, time 55.612 samples/sec
+epoch 1, total_step 19620, total loss is 32.58 , inference loss is 19.65, weight deacy loss is 12.93, training accuracy is 0.093750, time 56.654 samples/sec
+epoch 1, total_step 19640, total loss is 31.64 , inference loss is 18.72, weight deacy loss is 12.93, training accuracy is 0.093750, time 56.613 samples/sec
+epoch 1, total_step 19660, total loss is 33.11 , inference loss is 20.19, weight deacy loss is 12.93, training accuracy is 0.093750, time 56.242 samples/sec
+epoch 1, total_step 19680, total loss is 33.06 , inference loss is 20.14, weight deacy loss is 12.92, training accuracy is 0.031250, time 56.434 samples/sec
+epoch 1, total_step 19700, total loss is 30.08 , inference loss is 17.16, weight deacy loss is 12.92, training accuracy is 0.125000, time 52.815 samples/sec
+epoch 1, total_step 19720, total loss is 30.99 , inference loss is 18.07, weight deacy loss is 12.92, training accuracy is 0.062500, time 57.173 samples/sec
+epoch 1, total_step 19740, total loss is 30.77 , inference loss is 17.85, weight deacy loss is 12.92, training accuracy is 0.093750, time 56.219 samples/sec
+epoch 1, total_step 19760, total loss is 33.70 , inference loss is 20.78, weight deacy loss is 12.92, training accuracy is 0.093750, time 56.027 samples/sec
+epoch 1, total_step 19780, total loss is 32.18 , inference loss is 19.27, weight deacy loss is 12.91, training accuracy is 0.125000, time 56.616 samples/sec
+epoch 1, total_step 19800, total loss is 34.67 , inference loss is 21.76, weight deacy loss is 12.91, training accuracy is 0.125000, time 56.230 samples/sec
+epoch 1, total_step 19820, total loss is 29.09 , inference loss is 16.18, weight deacy loss is 12.91, training accuracy is 0.156250, time 55.606 samples/sec
+epoch 1, total_step 19840, total loss is 31.63 , inference loss is 18.72, weight deacy loss is 12.91, training accuracy is 0.062500, time 56.848 samples/sec
+epoch 1, total_step 19860, total loss is 37.79 , inference loss is 24.88, weight deacy loss is 12.91, training accuracy is 0.093750, time 56.231 samples/sec
+epoch 1, total_step 19880, total loss is 27.60 , inference loss is 14.69, weight deacy loss is 12.91, training accuracy is 0.125000, time 56.277 samples/sec
+epoch 1, total_step 19900, total loss is 27.61 , inference loss is 14.70, weight deacy loss is 12.90, training accuracy is 0.093750, time 56.047 samples/sec
+epoch 1, total_step 19920, total loss is 30.70 , inference loss is 17.80, weight deacy loss is 12.90, training accuracy is 0.093750, time 55.750 samples/sec
+epoch 1, total_step 19940, total loss is 31.13 , inference loss is 18.23, weight deacy loss is 12.90, training accuracy is 0.093750, time 56.234 samples/sec
+epoch 1, total_step 19960, total loss is 25.31 , inference loss is 12.41, weight deacy loss is 12.90, training accuracy is 0.125000, time 56.843 samples/sec
+epoch 1, total_step 19980, total loss is 31.16 , inference loss is 18.26, weight deacy loss is 12.90, training accuracy is 0.093750, time 56.919 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.57952099999997
+best_threshold_index 111 0.9664814814814815
+best_threshold_index 111 0.9666666666666667
+best_threshold_index 111 0.9675925925925926
+best_threshold_index 111 0.9679629629629629
+best_threshold_index 111 0.9688888888888889
+best_threshold_index 111 0.9662962962962963
+best_threshold_index 111 0.9661111111111111
+best_threshold_index 111 0.9668518518518519
+best_threshold_index 111 0.9653703703703703
+best_threshold_index 111 0.9661111111111111
+[lfw][20000]XNorm: 22.940827
+[lfw][20000]Accuracy-Flip: 0.96683+-0.00886
+testing verification..
+(14000, 512)
+infer time 86.24205499999998
+best_threshold_index 115 0.9647619047619047
+best_threshold_index 115 0.9620634920634921
+best_threshold_index 115 0.9636507936507936
+best_threshold_index 115 0.963015873015873
+best_threshold_index 115 0.9622222222222222
+best_threshold_index 115 0.960952380952381
+best_threshold_index 115 0.9625396825396826
+best_threshold_index 115 0.9628571428571429
+best_threshold_index 115 0.9626984126984127
+best_threshold_index 115 0.9623809523809523
+[cfp_ff][20000]XNorm: 22.412570
+[cfp_ff][20000]Accuracy-Flip: 0.96271+-0.00858
+testing verification..
+(14000, 512)
+infer time 86.37985699999997
+best_threshold_index 127 0.8374603174603175
+best_threshold_index 126 0.8373015873015873
+best_threshold_index 128 0.8353968253968254
+best_threshold_index 126 0.839047619047619
+best_threshold_index 128 0.8401587301587301
+best_threshold_index 126 0.8301587301587302
+best_threshold_index 126 0.8358730158730159
+best_threshold_index 127 0.8401587301587301
+best_threshold_index 128 0.8369841269841269
+best_threshold_index 128 0.8392063492063492
+[cfp_fp][20000]XNorm: 22.049591
+[cfp_fp][20000]Accuracy-Flip: 0.83457+-0.02526
+testing verification..
+(12000, 512)
+infer time 74.73598299999985
+best_threshold_index 133 0.8072222222222222
+best_threshold_index 133 0.8077777777777778
+best_threshold_index 133 0.8087037037037037
+best_threshold_index 133 0.8083333333333333
+best_threshold_index 133 0.8112962962962963
+best_threshold_index 133 0.8085185185185185
+best_threshold_index 133 0.8114814814814815
+best_threshold_index 133 0.8114814814814815
+best_threshold_index 133 0.8107407407407408
+best_threshold_index 133 0.8077777777777778
+[agedb_30][20000]XNorm: 22.340433
+[agedb_30][20000]Accuracy-Flip: 0.80933+-0.01463
+test accuracy is: 0.9668333333333333
+epoch 1, total_step 20000, total loss is 28.51 , inference loss is 15.61, weight deacy loss is 12.89, training accuracy is 0.125000, time 52.870 samples/sec
+epoch 1, total_step 20020, total loss is 30.88 , inference loss is 17.99, weight deacy loss is 12.89, training accuracy is 0.218750, time 56.548 samples/sec
+epoch 1, total_step 20040, total loss is 33.34 , inference loss is 20.45, weight deacy loss is 12.89, training accuracy is 0.031250, time 55.687 samples/sec
+epoch 1, total_step 20060, total loss is 32.48 , inference loss is 19.59, weight deacy loss is 12.89, training accuracy is 0.125000, time 56.551 samples/sec
+epoch 1, total_step 20080, total loss is 33.83 , inference loss is 20.94, weight deacy loss is 12.89, training accuracy is 0.125000, time 56.738 samples/sec
+epoch 1, total_step 20100, total loss is 33.43 , inference loss is 20.54, weight deacy loss is 12.89, training accuracy is 0.062500, time 56.048 samples/sec
+epoch 1, total_step 20120, total loss is 29.29 , inference loss is 16.40, weight deacy loss is 12.88, training accuracy is 0.187500, time 56.457 samples/sec
+epoch 1, total_step 20140, total loss is 31.14 , inference loss is 18.26, weight deacy loss is 12.88, training accuracy is 0.156250, time 56.670 samples/sec
+epoch 1, total_step 20160, total loss is 27.43 , inference loss is 14.55, weight deacy loss is 12.88, training accuracy is 0.125000, time 56.437 samples/sec
+epoch 1, total_step 20180, total loss is 34.85 , inference loss is 21.97, weight deacy loss is 12.88, training accuracy is 0.000000, time 55.760 samples/sec
+epoch 1, total_step 20200, total loss is 31.00 , inference loss is 18.13, weight deacy loss is 12.88, training accuracy is 0.093750, time 56.674 samples/sec
+epoch 1, total_step 20220, total loss is 32.60 , inference loss is 19.72, weight deacy loss is 12.88, training accuracy is 0.000000, time 55.804 samples/sec
+epoch 1, total_step 20240, total loss is 31.79 , inference loss is 18.92, weight deacy loss is 12.87, training accuracy is 0.031250, time 56.624 samples/sec
+epoch 1, total_step 20260, total loss is 31.74 , inference loss is 18.87, weight deacy loss is 12.87, training accuracy is 0.062500, time 56.272 samples/sec
+epoch 1, total_step 20280, total loss is 27.93 , inference loss is 15.06, weight deacy loss is 12.87, training accuracy is 0.125000, time 54.697 samples/sec
+epoch 1, total_step 20300, total loss is 33.26 , inference loss is 20.39, weight deacy loss is 12.87, training accuracy is 0.125000, time 55.963 samples/sec
+epoch 1, total_step 20320, total loss is 28.56 , inference loss is 15.69, weight deacy loss is 12.87, training accuracy is 0.125000, time 56.072 samples/sec
+epoch 1, total_step 20340, total loss is 35.53 , inference loss is 22.66, weight deacy loss is 12.86, training accuracy is 0.031250, time 56.367 samples/sec
+epoch 1, total_step 20360, total loss is 33.98 , inference loss is 21.12, weight deacy loss is 12.86, training accuracy is 0.062500, time 56.794 samples/sec
+epoch 1, total_step 20380, total loss is 30.00 , inference loss is 17.14, weight deacy loss is 12.86, training accuracy is 0.187500, time 56.198 samples/sec
+epoch 1, total_step 20400, total loss is 38.25 , inference loss is 25.39, weight deacy loss is 12.86, training accuracy is 0.125000, time 56.442 samples/sec
+epoch 1, total_step 20420, total loss is 27.21 , inference loss is 14.35, weight deacy loss is 12.86, training accuracy is 0.062500, time 56.399 samples/sec
+epoch 1, total_step 20440, total loss is 34.87 , inference loss is 22.01, weight deacy loss is 12.86, training accuracy is 0.093750, time 55.719 samples/sec
+epoch 1, total_step 20460, total loss is 36.51 , inference loss is 23.66, weight deacy loss is 12.85, training accuracy is 0.093750, time 55.671 samples/sec
+epoch 1, total_step 20480, total loss is 31.28 , inference loss is 18.42, weight deacy loss is 12.85, training accuracy is 0.062500, time 56.123 samples/sec
+epoch 1, total_step 20500, total loss is 29.02 , inference loss is 16.17, weight deacy loss is 12.85, training accuracy is 0.093750, time 55.842 samples/sec
+epoch 1, total_step 20520, total loss is 31.49 , inference loss is 18.64, weight deacy loss is 12.85, training accuracy is 0.156250, time 55.402 samples/sec
+epoch 1, total_step 20540, total loss is 30.70 , inference loss is 17.86, weight deacy loss is 12.85, training accuracy is 0.062500, time 56.753 samples/sec
+epoch 1, total_step 20560, total loss is 33.89 , inference loss is 21.04, weight deacy loss is 12.84, training accuracy is 0.062500, time 55.627 samples/sec
+epoch 1, total_step 20580, total loss is 31.29 , inference loss is 18.45, weight deacy loss is 12.84, training accuracy is 0.093750, time 56.245 samples/sec
+epoch 1, total_step 20600, total loss is 33.71 , inference loss is 20.87, weight deacy loss is 12.84, training accuracy is 0.125000, time 56.634 samples/sec
+epoch 1, total_step 20620, total loss is 35.05 , inference loss is 22.21, weight deacy loss is 12.84, training accuracy is 0.156250, time 55.788 samples/sec
+epoch 1, total_step 20640, total loss is 30.91 , inference loss is 18.08, weight deacy loss is 12.84, training accuracy is 0.125000, time 56.892 samples/sec
+epoch 1, total_step 20660, total loss is 31.12 , inference loss is 18.29, weight deacy loss is 12.84, training accuracy is 0.093750, time 54.414 samples/sec
+epoch 1, total_step 20680, total loss is 33.38 , inference loss is 20.54, weight deacy loss is 12.83, training accuracy is 0.031250, time 56.694 samples/sec
+epoch 1, total_step 20700, total loss is 27.37 , inference loss is 14.53, weight deacy loss is 12.83, training accuracy is 0.125000, time 56.344 samples/sec
+epoch 1, total_step 20720, total loss is 29.69 , inference loss is 16.86, weight deacy loss is 12.83, training accuracy is 0.156250, time 56.286 samples/sec
+epoch 1, total_step 20740, total loss is 31.53 , inference loss is 18.70, weight deacy loss is 12.83, training accuracy is 0.125000, time 55.754 samples/sec
+epoch 1, total_step 20760, total loss is 32.87 , inference loss is 20.05, weight deacy loss is 12.83, training accuracy is 0.031250, time 56.127 samples/sec
+epoch 1, total_step 20780, total loss is 26.19 , inference loss is 13.37, weight deacy loss is 12.82, training accuracy is 0.187500, time 56.924 samples/sec
+epoch 1, total_step 20800, total loss is 32.31 , inference loss is 19.49, weight deacy loss is 12.82, training accuracy is 0.156250, time 55.867 samples/sec
+epoch 1, total_step 20820, total loss is 28.62 , inference loss is 15.80, weight deacy loss is 12.82, training accuracy is 0.125000, time 55.851 samples/sec
+epoch 1, total_step 20840, total loss is 32.60 , inference loss is 19.78, weight deacy loss is 12.82, training accuracy is 0.093750, time 56.744 samples/sec
+epoch 1, total_step 20860, total loss is 35.23 , inference loss is 22.41, weight deacy loss is 12.82, training accuracy is 0.125000, time 56.067 samples/sec
+epoch 1, total_step 20880, total loss is 32.61 , inference loss is 19.80, weight deacy loss is 12.82, training accuracy is 0.062500, time 56.962 samples/sec
+epoch 1, total_step 20900, total loss is 33.68 , inference loss is 20.87, weight deacy loss is 12.81, training accuracy is 0.062500, time 55.623 samples/sec
+epoch 1, total_step 20920, total loss is 32.88 , inference loss is 20.06, weight deacy loss is 12.81, training accuracy is 0.187500, time 55.545 samples/sec
+epoch 1, total_step 20940, total loss is 34.21 , inference loss is 21.40, weight deacy loss is 12.81, training accuracy is 0.125000, time 56.261 samples/sec
+epoch 1, total_step 20960, total loss is 34.09 , inference loss is 21.28, weight deacy loss is 12.81, training accuracy is 0.031250, time 52.328 samples/sec
+epoch 1, total_step 20980, total loss is 30.10 , inference loss is 17.30, weight deacy loss is 12.81, training accuracy is 0.125000, time 55.674 samples/sec
+epoch 1, total_step 21000, total loss is 37.09 , inference loss is 24.28, weight deacy loss is 12.81, training accuracy is 0.125000, time 55.333 samples/sec
+epoch 1, total_step 21020, total loss is 33.76 , inference loss is 20.96, weight deacy loss is 12.80, training accuracy is 0.093750, time 54.825 samples/sec
+epoch 1, total_step 21040, total loss is 31.04 , inference loss is 18.23, weight deacy loss is 12.80, training accuracy is 0.093750, time 56.404 samples/sec
+epoch 1, total_step 21060, total loss is 32.04 , inference loss is 19.24, weight deacy loss is 12.80, training accuracy is 0.125000, time 56.477 samples/sec
+epoch 1, total_step 21080, total loss is 30.27 , inference loss is 17.48, weight deacy loss is 12.80, training accuracy is 0.156250, time 55.774 samples/sec
+epoch 1, total_step 21100, total loss is 34.67 , inference loss is 21.88, weight deacy loss is 12.80, training accuracy is 0.093750, time 56.311 samples/sec
+epoch 1, total_step 21120, total loss is 28.85 , inference loss is 16.06, weight deacy loss is 12.79, training accuracy is 0.093750, time 56.306 samples/sec
+epoch 1, total_step 21140, total loss is 36.28 , inference loss is 23.48, weight deacy loss is 12.79, training accuracy is 0.062500, time 56.498 samples/sec
+epoch 1, total_step 21160, total loss is 32.65 , inference loss is 19.86, weight deacy loss is 12.79, training accuracy is 0.062500, time 56.999 samples/sec
+epoch 1, total_step 21180, total loss is 33.73 , inference loss is 20.94, weight deacy loss is 12.79, training accuracy is 0.062500, time 56.538 samples/sec
+epoch 1, total_step 21200, total loss is 31.48 , inference loss is 18.70, weight deacy loss is 12.79, training accuracy is 0.156250, time 56.143 samples/sec
+epoch 1, total_step 21220, total loss is 30.71 , inference loss is 17.92, weight deacy loss is 12.79, training accuracy is 0.156250, time 55.529 samples/sec
+epoch 1, total_step 21240, total loss is 28.70 , inference loss is 15.91, weight deacy loss is 12.78, training accuracy is 0.156250, time 56.369 samples/sec
+epoch 1, total_step 21260, total loss is 26.51 , inference loss is 13.73, weight deacy loss is 12.78, training accuracy is 0.093750, time 56.295 samples/sec
+epoch 1, total_step 21280, total loss is 26.83 , inference loss is 14.05, weight deacy loss is 12.78, training accuracy is 0.156250, time 56.206 samples/sec
+epoch 1, total_step 21300, total loss is 27.74 , inference loss is 14.96, weight deacy loss is 12.78, training accuracy is 0.218750, time 56.057 samples/sec
+epoch 1, total_step 21320, total loss is 30.67 , inference loss is 17.90, weight deacy loss is 12.78, training accuracy is 0.156250, time 55.284 samples/sec
+epoch 1, total_step 21340, total loss is 35.14 , inference loss is 22.37, weight deacy loss is 12.77, training accuracy is 0.062500, time 56.878 samples/sec
+epoch 1, total_step 21360, total loss is 33.73 , inference loss is 20.96, weight deacy loss is 12.77, training accuracy is 0.062500, time 55.330 samples/sec
+epoch 1, total_step 21380, total loss is 33.15 , inference loss is 20.38, weight deacy loss is 12.77, training accuracy is 0.156250, time 56.106 samples/sec
+epoch 1, total_step 21400, total loss is 35.75 , inference loss is 22.98, weight deacy loss is 12.77, training accuracy is 0.000000, time 57.193 samples/sec
+epoch 1, total_step 21420, total loss is 28.74 , inference loss is 15.97, weight deacy loss is 12.77, training accuracy is 0.125000, time 55.523 samples/sec
+epoch 1, total_step 21440, total loss is 34.75 , inference loss is 21.98, weight deacy loss is 12.77, training accuracy is 0.000000, time 56.488 samples/sec
+epoch 1, total_step 21460, total loss is 32.89 , inference loss is 20.12, weight deacy loss is 12.76, training accuracy is 0.062500, time 55.218 samples/sec
+epoch 1, total_step 21480, total loss is 35.00 , inference loss is 22.23, weight deacy loss is 12.76, training accuracy is 0.062500, time 56.185 samples/sec
+epoch 1, total_step 21500, total loss is 30.15 , inference loss is 17.39, weight deacy loss is 12.76, training accuracy is 0.093750, time 56.276 samples/sec
+epoch 1, total_step 21520, total loss is 32.81 , inference loss is 20.05, weight deacy loss is 12.76, training accuracy is 0.062500, time 56.161 samples/sec
+epoch 1, total_step 21540, total loss is 31.48 , inference loss is 18.72, weight deacy loss is 12.76, training accuracy is 0.062500, time 55.264 samples/sec
+epoch 1, total_step 21560, total loss is 27.72 , inference loss is 14.96, weight deacy loss is 12.76, training accuracy is 0.187500, time 56.575 samples/sec
+epoch 1, total_step 21580, total loss is 31.36 , inference loss is 18.61, weight deacy loss is 12.75, training accuracy is 0.093750, time 57.033 samples/sec
+epoch 1, total_step 21600, total loss is 31.50 , inference loss is 18.75, weight deacy loss is 12.75, training accuracy is 0.093750, time 55.782 samples/sec
+epoch 1, total_step 21620, total loss is 26.35 , inference loss is 13.60, weight deacy loss is 12.75, training accuracy is 0.093750, time 57.011 samples/sec
+epoch 1, total_step 21640, total loss is 30.10 , inference loss is 17.35, weight deacy loss is 12.75, training accuracy is 0.125000, time 55.771 samples/sec
+epoch 1, total_step 21660, total loss is 34.25 , inference loss is 21.51, weight deacy loss is 12.75, training accuracy is 0.093750, time 56.744 samples/sec
+epoch 1, total_step 21680, total loss is 29.99 , inference loss is 17.25, weight deacy loss is 12.74, training accuracy is 0.031250, time 56.279 samples/sec
+epoch 1, total_step 21700, total loss is 31.39 , inference loss is 18.65, weight deacy loss is 12.74, training accuracy is 0.031250, time 54.950 samples/sec
+epoch 1, total_step 21720, total loss is 27.91 , inference loss is 15.17, weight deacy loss is 12.74, training accuracy is 0.031250, time 55.574 samples/sec
+epoch 1, total_step 21740, total loss is 33.97 , inference loss is 21.23, weight deacy loss is 12.74, training accuracy is 0.031250, time 55.671 samples/sec
+epoch 1, total_step 21760, total loss is 30.46 , inference loss is 17.72, weight deacy loss is 12.74, training accuracy is 0.125000, time 56.378 samples/sec
+epoch 1, total_step 21780, total loss is 29.11 , inference loss is 16.37, weight deacy loss is 12.74, training accuracy is 0.156250, time 56.118 samples/sec
+epoch 1, total_step 21800, total loss is 28.48 , inference loss is 15.74, weight deacy loss is 12.73, training accuracy is 0.156250, time 56.294 samples/sec
+epoch 1, total_step 21820, total loss is 37.09 , inference loss is 24.35, weight deacy loss is 12.73, training accuracy is 0.031250, time 56.636 samples/sec
+epoch 1, total_step 21840, total loss is 33.81 , inference loss is 21.08, weight deacy loss is 12.73, training accuracy is 0.125000, time 56.346 samples/sec
+epoch 1, total_step 21860, total loss is 30.10 , inference loss is 17.37, weight deacy loss is 12.73, training accuracy is 0.156250, time 56.406 samples/sec
+epoch 1, total_step 21880, total loss is 32.63 , inference loss is 19.91, weight deacy loss is 12.73, training accuracy is 0.187500, time 55.622 samples/sec
+epoch 1, total_step 21900, total loss is 28.92 , inference loss is 16.20, weight deacy loss is 12.73, training accuracy is 0.125000, time 54.591 samples/sec
+epoch 1, total_step 21920, total loss is 28.27 , inference loss is 15.54, weight deacy loss is 12.72, training accuracy is 0.125000, time 56.290 samples/sec
+epoch 1, total_step 21940, total loss is 35.14 , inference loss is 22.42, weight deacy loss is 12.72, training accuracy is 0.093750, time 55.873 samples/sec
+epoch 1, total_step 21960, total loss is 31.90 , inference loss is 19.18, weight deacy loss is 12.72, training accuracy is 0.000000, time 56.123 samples/sec
+epoch 1, total_step 21980, total loss is 31.53 , inference loss is 18.81, weight deacy loss is 12.72, training accuracy is 0.062500, time 54.712 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.80803300000001
+best_threshold_index 100 0.9638888888888889
+best_threshold_index 104 0.9642592592592593
+best_threshold_index 100 0.9640740740740741
+best_threshold_index 101 0.9655555555555555
+best_threshold_index 100 0.9646296296296296
+best_threshold_index 101 0.9627777777777777
+best_threshold_index 100 0.9638888888888889
+best_threshold_index 100 0.9640740740740741
+best_threshold_index 100 0.9622222222222222
+best_threshold_index 100 0.962037037037037
+[lfw][22000]XNorm: 24.449271
+[lfw][22000]Accuracy-Flip: 0.96250+-0.00970
+testing verification..
+(14000, 512)
+infer time 87.3313910000002
+best_threshold_index 105 0.9636507936507936
+best_threshold_index 106 0.9606349206349206
+best_threshold_index 105 0.9615873015873015
+best_threshold_index 105 0.9601587301587302
+best_threshold_index 105 0.9598412698412698
+best_threshold_index 106 0.9596825396825397
+best_threshold_index 106 0.9606349206349206
+best_threshold_index 105 0.9596825396825397
+best_threshold_index 105 0.9615873015873015
+best_threshold_index 105 0.9619047619047619
+[cfp_ff][22000]XNorm: 23.937119
+[cfp_ff][22000]Accuracy-Flip: 0.96014+-0.01076
+testing verification..
+(14000, 512)
+infer time 87.19210100000001
+best_threshold_index 133 0.8231746031746032
+best_threshold_index 124 0.8196825396825397
+best_threshold_index 124 0.8219047619047619
+best_threshold_index 131 0.8214285714285714
+best_threshold_index 130 0.8242857142857143
+best_threshold_index 124 0.8155555555555556
+best_threshold_index 129 0.8198412698412698
+best_threshold_index 124 0.8185714285714286
+best_threshold_index 131 0.8219047619047619
+best_threshold_index 130 0.8238095238095238
+[cfp_fp][22000]XNorm: 23.420402
+[cfp_fp][22000]Accuracy-Flip: 0.81529+-0.02168
+testing verification..
+(12000, 512)
+infer time 74.90750200000001
+best_threshold_index 123 0.8048148148148148
+best_threshold_index 123 0.8044444444444444
+best_threshold_index 123 0.8087037037037037
+best_threshold_index 123 0.8053703703703704
+best_threshold_index 123 0.8122222222222222
+best_threshold_index 123 0.8053703703703704
+best_threshold_index 123 0.8087037037037037
+best_threshold_index 123 0.8114814814814815
+best_threshold_index 123 0.8083333333333333
+best_threshold_index 123 0.8088888888888889
+[agedb_30][22000]XNorm: 23.224413
+[agedb_30][22000]Accuracy-Flip: 0.80783+-0.02350
+test accuracy is: 0.9625
+epoch 1, total_step 22000, total loss is 37.31 , inference loss is 24.59, weight deacy loss is 12.72, training accuracy is 0.000000, time 52.252 samples/sec
+epoch 1, total_step 22020, total loss is 32.86 , inference loss is 20.15, weight deacy loss is 12.71, training accuracy is 0.093750, time 55.850 samples/sec
+epoch 1, total_step 22040, total loss is 30.19 , inference loss is 17.48, weight deacy loss is 12.71, training accuracy is 0.093750, time 56.289 samples/sec
+epoch 1, total_step 22060, total loss is 30.78 , inference loss is 18.07, weight deacy loss is 12.71, training accuracy is 0.031250, time 55.245 samples/sec
+epoch 1, total_step 22080, total loss is 28.85 , inference loss is 16.14, weight deacy loss is 12.71, training accuracy is 0.125000, time 57.176 samples/sec
+epoch 1, total_step 22100, total loss is 27.33 , inference loss is 14.63, weight deacy loss is 12.71, training accuracy is 0.125000, time 57.403 samples/sec
+epoch 1, total_step 22120, total loss is 29.58 , inference loss is 16.88, weight deacy loss is 12.71, training accuracy is 0.156250, time 57.211 samples/sec
+epoch 1, total_step 22140, total loss is 33.14 , inference loss is 20.44, weight deacy loss is 12.70, training accuracy is 0.031250, time 56.211 samples/sec
+epoch 1, total_step 22160, total loss is 32.98 , inference loss is 20.28, weight deacy loss is 12.70, training accuracy is 0.062500, time 55.986 samples/sec
+epoch 1, total_step 22180, total loss is 28.99 , inference loss is 16.29, weight deacy loss is 12.70, training accuracy is 0.093750, time 56.884 samples/sec
+epoch 1, total_step 22200, total loss is 29.06 , inference loss is 16.36, weight deacy loss is 12.70, training accuracy is 0.093750, time 55.791 samples/sec
+epoch 1, total_step 22220, total loss is 32.65 , inference loss is 19.95, weight deacy loss is 12.70, training accuracy is 0.093750, time 55.518 samples/sec
+epoch 1, total_step 22240, total loss is 32.79 , inference loss is 20.09, weight deacy loss is 12.70, training accuracy is 0.093750, time 56.430 samples/sec
+epoch 1, total_step 22260, total loss is 31.66 , inference loss is 18.97, weight deacy loss is 12.69, training accuracy is 0.062500, time 56.087 samples/sec
+epoch 1, total_step 22280, total loss is 28.36 , inference loss is 15.66, weight deacy loss is 12.69, training accuracy is 0.125000, time 55.574 samples/sec
+epoch 1, total_step 22300, total loss is 30.23 , inference loss is 17.54, weight deacy loss is 12.69, training accuracy is 0.125000, time 55.432 samples/sec
+epoch 1, total_step 22320, total loss is 29.51 , inference loss is 16.83, weight deacy loss is 12.69, training accuracy is 0.093750, time 56.775 samples/sec
+epoch 1, total_step 22340, total loss is 30.92 , inference loss is 18.23, weight deacy loss is 12.69, training accuracy is 0.125000, time 56.637 samples/sec
+epoch 1, total_step 22360, total loss is 32.24 , inference loss is 19.55, weight deacy loss is 12.68, training accuracy is 0.062500, time 56.435 samples/sec
+epoch 1, total_step 22380, total loss is 32.08 , inference loss is 19.40, weight deacy loss is 12.68, training accuracy is 0.062500, time 55.501 samples/sec
+epoch 1, total_step 22400, total loss is 36.35 , inference loss is 23.67, weight deacy loss is 12.68, training accuracy is 0.031250, time 56.291 samples/sec
+epoch 1, total_step 22420, total loss is 32.41 , inference loss is 19.73, weight deacy loss is 12.68, training accuracy is 0.000000, time 56.439 samples/sec
+epoch 1, total_step 22440, total loss is 33.80 , inference loss is 21.12, weight deacy loss is 12.68, training accuracy is 0.125000, time 56.285 samples/sec
+epoch 1, total_step 22460, total loss is 26.83 , inference loss is 14.16, weight deacy loss is 12.68, training accuracy is 0.156250, time 56.104 samples/sec
+epoch 1, total_step 22480, total loss is 30.76 , inference loss is 18.09, weight deacy loss is 12.67, training accuracy is 0.062500, time 56.603 samples/sec
+epoch 1, total_step 22500, total loss is 33.62 , inference loss is 20.94, weight deacy loss is 12.67, training accuracy is 0.093750, time 55.363 samples/sec
+epoch 1, total_step 22520, total loss is 31.72 , inference loss is 19.05, weight deacy loss is 12.67, training accuracy is 0.062500, time 55.643 samples/sec
+epoch 1, total_step 22540, total loss is 30.41 , inference loss is 17.74, weight deacy loss is 12.67, training accuracy is 0.093750, time 56.142 samples/sec
+epoch 1, total_step 22560, total loss is 35.42 , inference loss is 22.76, weight deacy loss is 12.67, training accuracy is 0.062500, time 56.136 samples/sec
+epoch 1, total_step 22580, total loss is 31.32 , inference loss is 18.65, weight deacy loss is 12.67, training accuracy is 0.031250, time 56.322 samples/sec
+epoch 1, total_step 22600, total loss is 28.89 , inference loss is 16.22, weight deacy loss is 12.66, training accuracy is 0.187500, time 55.794 samples/sec
+epoch 1, total_step 22620, total loss is 32.87 , inference loss is 20.21, weight deacy loss is 12.66, training accuracy is 0.093750, time 56.074 samples/sec
+epoch 1, total_step 22640, total loss is 29.47 , inference loss is 16.81, weight deacy loss is 12.66, training accuracy is 0.062500, time 55.702 samples/sec
+epoch 1, total_step 22660, total loss is 34.56 , inference loss is 21.90, weight deacy loss is 12.66, training accuracy is 0.062500, time 56.667 samples/sec
+epoch 1, total_step 22680, total loss is 30.17 , inference loss is 17.51, weight deacy loss is 12.66, training accuracy is 0.093750, time 55.930 samples/sec
+epoch 1, total_step 22700, total loss is 36.22 , inference loss is 23.56, weight deacy loss is 12.66, training accuracy is 0.062500, time 55.181 samples/sec
+epoch 1, total_step 22720, total loss is 31.05 , inference loss is 18.40, weight deacy loss is 12.65, training accuracy is 0.062500, time 56.014 samples/sec
+epoch 1, total_step 22740, total loss is 27.39 , inference loss is 14.73, weight deacy loss is 12.65, training accuracy is 0.218750, time 56.377 samples/sec
+epoch 1, total_step 22760, total loss is 31.19 , inference loss is 18.54, weight deacy loss is 12.65, training accuracy is 0.093750, time 56.132 samples/sec
+epoch 1, total_step 22780, total loss is 28.93 , inference loss is 16.28, weight deacy loss is 12.65, training accuracy is 0.156250, time 55.829 samples/sec
+epoch 1, total_step 22800, total loss is 35.74 , inference loss is 23.09, weight deacy loss is 12.65, training accuracy is 0.031250, time 56.476 samples/sec
+epoch 1, total_step 22820, total loss is 26.30 , inference loss is 13.66, weight deacy loss is 12.64, training accuracy is 0.156250, time 56.138 samples/sec
+epoch 1, total_step 22840, total loss is 31.10 , inference loss is 18.46, weight deacy loss is 12.64, training accuracy is 0.156250, time 55.724 samples/sec
+epoch 1, total_step 22860, total loss is 35.21 , inference loss is 22.57, weight deacy loss is 12.64, training accuracy is 0.125000, time 56.109 samples/sec
+epoch 1, total_step 22880, total loss is 29.63 , inference loss is 16.99, weight deacy loss is 12.64, training accuracy is 0.156250, time 54.128 samples/sec
+epoch 1, total_step 22900, total loss is 30.76 , inference loss is 18.12, weight deacy loss is 12.64, training accuracy is 0.031250, time 56.362 samples/sec
+epoch 1, total_step 22920, total loss is 32.43 , inference loss is 19.79, weight deacy loss is 12.64, training accuracy is 0.093750, time 55.822 samples/sec
+epoch 1, total_step 22940, total loss is 31.72 , inference loss is 19.09, weight deacy loss is 12.63, training accuracy is 0.031250, time 56.263 samples/sec
+epoch 1, total_step 22960, total loss is 31.20 , inference loss is 18.57, weight deacy loss is 12.63, training accuracy is 0.093750, time 55.879 samples/sec
+epoch 1, total_step 22980, total loss is 26.23 , inference loss is 13.60, weight deacy loss is 12.63, training accuracy is 0.125000, time 55.097 samples/sec
+epoch 1, total_step 23000, total loss is 29.56 , inference loss is 16.93, weight deacy loss is 12.63, training accuracy is 0.156250, time 56.765 samples/sec
+epoch 1, total_step 23020, total loss is 29.00 , inference loss is 16.37, weight deacy loss is 12.63, training accuracy is 0.187500, time 55.842 samples/sec
+epoch 1, total_step 23040, total loss is 28.48 , inference loss is 15.86, weight deacy loss is 12.63, training accuracy is 0.125000, time 56.003 samples/sec
+epoch 1, total_step 23060, total loss is 32.23 , inference loss is 19.61, weight deacy loss is 12.62, training accuracy is 0.125000, time 55.759 samples/sec
+epoch 1, total_step 23080, total loss is 30.19 , inference loss is 17.57, weight deacy loss is 12.62, training accuracy is 0.218750, time 55.688 samples/sec
+epoch 1, total_step 23100, total loss is 30.13 , inference loss is 17.51, weight deacy loss is 12.62, training accuracy is 0.093750, time 56.522 samples/sec
+epoch 1, total_step 23120, total loss is 26.95 , inference loss is 14.33, weight deacy loss is 12.62, training accuracy is 0.093750, time 56.763 samples/sec
+epoch 1, total_step 23140, total loss is 31.26 , inference loss is 18.64, weight deacy loss is 12.62, training accuracy is 0.000000, time 55.650 samples/sec
+epoch 1, total_step 23160, total loss is 32.05 , inference loss is 19.43, weight deacy loss is 12.62, training accuracy is 0.031250, time 56.337 samples/sec
+epoch 1, total_step 23180, total loss is 29.37 , inference loss is 16.75, weight deacy loss is 12.61, training accuracy is 0.156250, time 55.694 samples/sec
+epoch 1, total_step 23200, total loss is 27.99 , inference loss is 15.38, weight deacy loss is 12.61, training accuracy is 0.218750, time 56.041 samples/sec
+epoch 1, total_step 23220, total loss is 31.55 , inference loss is 18.94, weight deacy loss is 12.61, training accuracy is 0.125000, time 55.966 samples/sec
+epoch 1, total_step 23240, total loss is 35.00 , inference loss is 22.40, weight deacy loss is 12.61, training accuracy is 0.125000, time 55.358 samples/sec
+epoch 1, total_step 23260, total loss is 29.03 , inference loss is 16.42, weight deacy loss is 12.61, training accuracy is 0.156250, time 56.774 samples/sec
+epoch 1, total_step 23280, total loss is 32.33 , inference loss is 19.73, weight deacy loss is 12.60, training accuracy is 0.062500, time 55.605 samples/sec
+epoch 1, total_step 23300, total loss is 32.71 , inference loss is 20.11, weight deacy loss is 12.60, training accuracy is 0.093750, time 56.281 samples/sec
+epoch 1, total_step 23320, total loss is 29.80 , inference loss is 17.19, weight deacy loss is 12.60, training accuracy is 0.125000, time 56.202 samples/sec
+epoch 1, total_step 23340, total loss is 29.42 , inference loss is 16.82, weight deacy loss is 12.60, training accuracy is 0.187500, time 55.889 samples/sec
+epoch 1, total_step 23360, total loss is 33.53 , inference loss is 20.93, weight deacy loss is 12.60, training accuracy is 0.031250, time 55.777 samples/sec
+epoch 1, total_step 23380, total loss is 24.79 , inference loss is 12.20, weight deacy loss is 12.60, training accuracy is 0.156250, time 56.012 samples/sec
+epoch 1, total_step 23400, total loss is 29.89 , inference loss is 17.29, weight deacy loss is 12.59, training accuracy is 0.031250, time 56.436 samples/sec
+epoch 1, total_step 23420, total loss is 29.26 , inference loss is 16.67, weight deacy loss is 12.59, training accuracy is 0.093750, time 54.839 samples/sec
+epoch 1, total_step 23440, total loss is 33.98 , inference loss is 21.39, weight deacy loss is 12.59, training accuracy is 0.062500, time 55.910 samples/sec
+epoch 1, total_step 23460, total loss is 29.10 , inference loss is 16.51, weight deacy loss is 12.59, training accuracy is 0.062500, time 53.715 samples/sec
+epoch 1, total_step 23480, total loss is 29.22 , inference loss is 16.63, weight deacy loss is 12.59, training accuracy is 0.125000, time 56.931 samples/sec
+epoch 1, total_step 23500, total loss is 32.75 , inference loss is 20.16, weight deacy loss is 12.59, training accuracy is 0.125000, time 56.543 samples/sec
+epoch 1, total_step 23520, total loss is 28.88 , inference loss is 16.29, weight deacy loss is 12.58, training accuracy is 0.156250, time 55.737 samples/sec
+epoch 1, total_step 23540, total loss is 32.28 , inference loss is 19.70, weight deacy loss is 12.58, training accuracy is 0.125000, time 56.567 samples/sec
+epoch 1, total_step 23560, total loss is 35.09 , inference loss is 22.51, weight deacy loss is 12.58, training accuracy is 0.031250, time 56.953 samples/sec
+epoch 1, total_step 23580, total loss is 27.86 , inference loss is 15.28, weight deacy loss is 12.58, training accuracy is 0.156250, time 55.688 samples/sec
+epoch 1, total_step 23600, total loss is 33.20 , inference loss is 20.62, weight deacy loss is 12.58, training accuracy is 0.125000, time 55.882 samples/sec
+epoch 1, total_step 23620, total loss is 31.05 , inference loss is 18.47, weight deacy loss is 12.58, training accuracy is 0.156250, time 56.722 samples/sec
+epoch 1, total_step 23640, total loss is 30.35 , inference loss is 17.78, weight deacy loss is 12.57, training accuracy is 0.093750, time 55.853 samples/sec
+epoch 1, total_step 23660, total loss is 32.14 , inference loss is 19.57, weight deacy loss is 12.57, training accuracy is 0.093750, time 56.706 samples/sec
+epoch 1, total_step 23680, total loss is 29.94 , inference loss is 17.37, weight deacy loss is 12.57, training accuracy is 0.062500, time 55.481 samples/sec
+epoch 1, total_step 23700, total loss is 32.30 , inference loss is 19.73, weight deacy loss is 12.57, training accuracy is 0.156250, time 56.508 samples/sec
+epoch 1, total_step 23720, total loss is 30.06 , inference loss is 17.50, weight deacy loss is 12.57, training accuracy is 0.062500, time 55.719 samples/sec
+epoch 1, total_step 23740, total loss is 28.62 , inference loss is 16.05, weight deacy loss is 12.56, training accuracy is 0.093750, time 56.606 samples/sec
+epoch 1, total_step 23760, total loss is 30.70 , inference loss is 18.13, weight deacy loss is 12.56, training accuracy is 0.125000, time 55.571 samples/sec
+epoch 1, total_step 23780, total loss is 32.62 , inference loss is 20.06, weight deacy loss is 12.56, training accuracy is 0.093750, time 56.609 samples/sec
+epoch 1, total_step 23800, total loss is 29.26 , inference loss is 16.70, weight deacy loss is 12.56, training accuracy is 0.218750, time 55.686 samples/sec
+epoch 1, total_step 23820, total loss is 33.37 , inference loss is 20.81, weight deacy loss is 12.56, training accuracy is 0.093750, time 56.392 samples/sec
+epoch 1, total_step 23840, total loss is 32.13 , inference loss is 19.57, weight deacy loss is 12.56, training accuracy is 0.156250, time 55.357 samples/sec
+epoch 1, total_step 23860, total loss is 31.36 , inference loss is 18.81, weight deacy loss is 12.55, training accuracy is 0.062500, time 55.993 samples/sec
+epoch 1, total_step 23880, total loss is 28.76 , inference loss is 16.21, weight deacy loss is 12.55, training accuracy is 0.125000, time 55.399 samples/sec
+epoch 1, total_step 23900, total loss is 26.97 , inference loss is 14.42, weight deacy loss is 12.55, training accuracy is 0.125000, time 56.182 samples/sec
+epoch 1, total_step 23920, total loss is 31.00 , inference loss is 18.45, weight deacy loss is 12.55, training accuracy is 0.187500, time 56.236 samples/sec
+epoch 1, total_step 23940, total loss is 32.25 , inference loss is 19.70, weight deacy loss is 12.55, training accuracy is 0.062500, time 56.085 samples/sec
+epoch 1, total_step 23960, total loss is 31.11 , inference loss is 18.57, weight deacy loss is 12.55, training accuracy is 0.218750, time 57.041 samples/sec
+epoch 1, total_step 23980, total loss is 29.79 , inference loss is 17.25, weight deacy loss is 12.54, training accuracy is 0.093750, time 56.073 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.65746300000004
+best_threshold_index 114 0.9727777777777777
+best_threshold_index 113 0.9729629629629629
+best_threshold_index 114 0.9738888888888889
+best_threshold_index 113 0.9737037037037037
+best_threshold_index 118 0.9740740740740741
+best_threshold_index 113 0.9724074074074074
+best_threshold_index 114 0.9724074074074074
+best_threshold_index 114 0.9725925925925926
+best_threshold_index 118 0.970925925925926
+best_threshold_index 113 0.9714814814814815
+[lfw][24000]XNorm: 23.265991
+[lfw][24000]Accuracy-Flip: 0.97083+-0.00851
+testing verification..
+(14000, 512)
+infer time 87.2520730000001
+best_threshold_index 116 0.9698412698412698
+best_threshold_index 116 0.966984126984127
+best_threshold_index 116 0.9680952380952381
+best_threshold_index 116 0.9663492063492064
+best_threshold_index 116 0.9668253968253968
+best_threshold_index 116 0.9665079365079365
+best_threshold_index 116 0.9668253968253968
+best_threshold_index 114 0.9658730158730159
+best_threshold_index 116 0.9674603174603175
+best_threshold_index 116 0.9680952380952381
+[cfp_ff][24000]XNorm: 23.382801
+[cfp_ff][24000]Accuracy-Flip: 0.96700+-0.00946
+testing verification..
+(14000, 512)
+infer time 87.283989
+best_threshold_index 133 0.8542857142857143
+best_threshold_index 133 0.8538095238095238
+best_threshold_index 133 0.8533333333333334
+best_threshold_index 133 0.8546031746031746
+best_threshold_index 133 0.856984126984127
+best_threshold_index 133 0.8493650793650793
+best_threshold_index 133 0.8517460317460317
+best_threshold_index 133 0.8517460317460317
+best_threshold_index 133 0.8534920634920635
+best_threshold_index 133 0.854920634920635
+[cfp_fp][24000]XNorm: 22.747041
+[cfp_fp][24000]Accuracy-Flip: 0.85343+-0.01785
+testing verification..
+(12000, 512)
+infer time 74.91884999999988
+best_threshold_index 136 0.8175925925925925
+best_threshold_index 130 0.8157407407407408
+best_threshold_index 137 0.8207407407407408
+best_threshold_index 136 0.8188888888888889
+best_threshold_index 136 0.8233333333333334
+best_threshold_index 136 0.8174074074074074
+best_threshold_index 136 0.8205555555555556
+best_threshold_index 137 0.8218518518518518
+best_threshold_index 136 0.8198148148148148
+best_threshold_index 136 0.8216666666666667
+[agedb_30][24000]XNorm: 22.322557
+[agedb_30][24000]Accuracy-Flip: 0.81683+-0.01842
+test accuracy is: 0.9708333333333334
+epoch 1, total_step 24000, total loss is 27.72 , inference loss is 15.17, weight deacy loss is 12.54, training accuracy is 0.093750, time 55.363 samples/sec
+epoch 1, total_step 24020, total loss is 29.27 , inference loss is 16.73, weight deacy loss is 12.54, training accuracy is 0.093750, time 56.829 samples/sec
+epoch 1, total_step 24040, total loss is 34.06 , inference loss is 21.52, weight deacy loss is 12.54, training accuracy is 0.125000, time 55.167 samples/sec
+epoch 1, total_step 24060, total loss is 34.05 , inference loss is 21.51, weight deacy loss is 12.54, training accuracy is 0.031250, time 55.662 samples/sec
+epoch 1, total_step 24080, total loss is 30.12 , inference loss is 17.58, weight deacy loss is 12.54, training accuracy is 0.093750, time 56.108 samples/sec
+epoch 1, total_step 24100, total loss is 29.07 , inference loss is 16.54, weight deacy loss is 12.53, training accuracy is 0.093750, time 55.762 samples/sec
+epoch 1, total_step 24120, total loss is 27.54 , inference loss is 15.01, weight deacy loss is 12.53, training accuracy is 0.062500, time 55.469 samples/sec
+epoch 1, total_step 24140, total loss is 30.73 , inference loss is 18.20, weight deacy loss is 12.53, training accuracy is 0.031250, time 56.951 samples/sec
+epoch 1, total_step 24160, total loss is 30.57 , inference loss is 18.04, weight deacy loss is 12.53, training accuracy is 0.125000, time 55.295 samples/sec
+epoch 1, total_step 24180, total loss is 29.38 , inference loss is 16.85, weight deacy loss is 12.53, training accuracy is 0.156250, time 56.467 samples/sec
+epoch 1, total_step 24200, total loss is 31.09 , inference loss is 18.56, weight deacy loss is 12.53, training accuracy is 0.093750, time 54.124 samples/sec
+epoch 1, total_step 24220, total loss is 29.52 , inference loss is 17.00, weight deacy loss is 12.52, training accuracy is 0.093750, time 55.552 samples/sec
+epoch 1, total_step 24240, total loss is 31.05 , inference loss is 18.53, weight deacy loss is 12.52, training accuracy is 0.062500, time 56.286 samples/sec
+epoch 1, total_step 24260, total loss is 30.20 , inference loss is 17.68, weight deacy loss is 12.52, training accuracy is 0.031250, time 56.140 samples/sec
+epoch 1, total_step 24280, total loss is 28.75 , inference loss is 16.23, weight deacy loss is 12.52, training accuracy is 0.156250, time 56.045 samples/sec
+epoch 1, total_step 24300, total loss is 34.32 , inference loss is 21.80, weight deacy loss is 12.52, training accuracy is 0.093750, time 55.854 samples/sec
+epoch 1, total_step 24320, total loss is 31.55 , inference loss is 19.03, weight deacy loss is 12.52, training accuracy is 0.062500, time 55.340 samples/sec
+epoch 1, total_step 24340, total loss is 34.00 , inference loss is 21.48, weight deacy loss is 12.51, training accuracy is 0.093750, time 53.101 samples/sec
+epoch 1, total_step 24360, total loss is 29.21 , inference loss is 16.70, weight deacy loss is 12.51, training accuracy is 0.156250, time 56.730 samples/sec
+epoch 1, total_step 24380, total loss is 28.32 , inference loss is 15.81, weight deacy loss is 12.51, training accuracy is 0.125000, time 56.357 samples/sec
+epoch 1, total_step 24400, total loss is 29.38 , inference loss is 16.87, weight deacy loss is 12.51, training accuracy is 0.125000, time 55.420 samples/sec
+epoch 1, total_step 24420, total loss is 27.55 , inference loss is 15.05, weight deacy loss is 12.51, training accuracy is 0.125000, time 55.955 samples/sec
+epoch 1, total_step 24440, total loss is 28.60 , inference loss is 16.09, weight deacy loss is 12.50, training accuracy is 0.156250, time 56.423 samples/sec
+epoch 1, total_step 24460, total loss is 34.38 , inference loss is 21.87, weight deacy loss is 12.50, training accuracy is 0.031250, time 57.062 samples/sec
+epoch 1, total_step 24480, total loss is 29.99 , inference loss is 17.49, weight deacy loss is 12.50, training accuracy is 0.156250, time 55.913 samples/sec
+epoch 1, total_step 24500, total loss is 28.61 , inference loss is 16.11, weight deacy loss is 12.50, training accuracy is 0.125000, time 55.891 samples/sec
+epoch 1, total_step 24520, total loss is 29.65 , inference loss is 17.15, weight deacy loss is 12.50, training accuracy is 0.093750, time 55.736 samples/sec
+epoch 1, total_step 24540, total loss is 31.87 , inference loss is 19.38, weight deacy loss is 12.50, training accuracy is 0.062500, time 56.503 samples/sec
+epoch 1, total_step 24560, total loss is 27.92 , inference loss is 15.43, weight deacy loss is 12.49, training accuracy is 0.125000, time 55.632 samples/sec
+epoch 1, total_step 24580, total loss is 30.71 , inference loss is 18.22, weight deacy loss is 12.49, training accuracy is 0.125000, time 56.228 samples/sec
+epoch 1, total_step 24600, total loss is 31.70 , inference loss is 19.21, weight deacy loss is 12.49, training accuracy is 0.125000, time 57.181 samples/sec
+epoch 1, total_step 24620, total loss is 29.13 , inference loss is 16.64, weight deacy loss is 12.49, training accuracy is 0.093750, time 55.913 samples/sec
+epoch 1, total_step 24640, total loss is 31.74 , inference loss is 19.26, weight deacy loss is 12.49, training accuracy is 0.031250, time 52.717 samples/sec
+epoch 1, total_step 24660, total loss is 30.07 , inference loss is 17.58, weight deacy loss is 12.49, training accuracy is 0.062500, time 56.002 samples/sec
+epoch 1, total_step 24680, total loss is 28.18 , inference loss is 15.69, weight deacy loss is 12.48, training accuracy is 0.187500, time 56.138 samples/sec
+epoch 1, total_step 24700, total loss is 31.17 , inference loss is 18.69, weight deacy loss is 12.48, training accuracy is 0.125000, time 57.514 samples/sec
+epoch 1, total_step 24720, total loss is 29.64 , inference loss is 17.16, weight deacy loss is 12.48, training accuracy is 0.125000, time 56.649 samples/sec
+epoch 1, total_step 24740, total loss is 32.63 , inference loss is 20.15, weight deacy loss is 12.48, training accuracy is 0.031250, time 55.964 samples/sec
+epoch 1, total_step 24760, total loss is 27.73 , inference loss is 15.25, weight deacy loss is 12.48, training accuracy is 0.093750, time 56.352 samples/sec
+epoch 1, total_step 24780, total loss is 30.76 , inference loss is 18.28, weight deacy loss is 12.48, training accuracy is 0.062500, time 56.717 samples/sec
+epoch 1, total_step 24800, total loss is 30.04 , inference loss is 17.56, weight deacy loss is 12.47, training accuracy is 0.156250, time 56.499 samples/sec
+epoch 1, total_step 24820, total loss is 31.13 , inference loss is 18.65, weight deacy loss is 12.47, training accuracy is 0.062500, time 55.606 samples/sec
+epoch 1, total_step 24840, total loss is 26.07 , inference loss is 13.60, weight deacy loss is 12.47, training accuracy is 0.062500, time 55.922 samples/sec
+epoch 1, total_step 24860, total loss is 32.36 , inference loss is 19.89, weight deacy loss is 12.47, training accuracy is 0.093750, time 56.529 samples/sec
+epoch 1, total_step 24880, total loss is 28.98 , inference loss is 16.51, weight deacy loss is 12.47, training accuracy is 0.187500, time 56.333 samples/sec
+epoch 1, total_step 24900, total loss is 27.57 , inference loss is 15.10, weight deacy loss is 12.47, training accuracy is 0.187500, time 56.421 samples/sec
+epoch 1, total_step 24920, total loss is 30.80 , inference loss is 18.34, weight deacy loss is 12.46, training accuracy is 0.062500, time 56.520 samples/sec
+epoch 1, total_step 24940, total loss is 34.77 , inference loss is 22.30, weight deacy loss is 12.46, training accuracy is 0.031250, time 56.244 samples/sec
+epoch 1, total_step 24960, total loss is 34.38 , inference loss is 21.92, weight deacy loss is 12.46, training accuracy is 0.125000, time 56.025 samples/sec
+epoch 1, total_step 24980, total loss is 30.18 , inference loss is 17.72, weight deacy loss is 12.46, training accuracy is 0.062500, time 56.214 samples/sec
+epoch 1, total_step 25000, total loss is 31.76 , inference loss is 19.30, weight deacy loss is 12.46, training accuracy is 0.031250, time 56.198 samples/sec
+epoch 1, total_step 25020, total loss is 27.16 , inference loss is 14.70, weight deacy loss is 12.46, training accuracy is 0.156250, time 55.886 samples/sec
+epoch 1, total_step 25040, total loss is 30.38 , inference loss is 17.92, weight deacy loss is 12.45, training accuracy is 0.156250, time 55.890 samples/sec
+epoch 1, total_step 25060, total loss is 27.52 , inference loss is 15.07, weight deacy loss is 12.45, training accuracy is 0.093750, time 56.797 samples/sec
+epoch 1, total_step 25080, total loss is 31.50 , inference loss is 19.05, weight deacy loss is 12.45, training accuracy is 0.156250, time 55.575 samples/sec
+epoch 1, total_step 25100, total loss is 27.22 , inference loss is 14.77, weight deacy loss is 12.45, training accuracy is 0.312500, time 55.971 samples/sec
+epoch 1, total_step 25120, total loss is 31.57 , inference loss is 19.13, weight deacy loss is 12.45, training accuracy is 0.125000, time 55.539 samples/sec
+epoch 1, total_step 25140, total loss is 28.16 , inference loss is 15.71, weight deacy loss is 12.45, training accuracy is 0.250000, time 56.120 samples/sec
+epoch 1, total_step 25160, total loss is 31.00 , inference loss is 18.55, weight deacy loss is 12.44, training accuracy is 0.125000, time 56.886 samples/sec
+epoch 1, total_step 25180, total loss is 28.67 , inference loss is 16.23, weight deacy loss is 12.44, training accuracy is 0.031250, time 56.021 samples/sec
+epoch 1, total_step 25200, total loss is 27.83 , inference loss is 15.39, weight deacy loss is 12.44, training accuracy is 0.125000, time 55.777 samples/sec
+epoch 1, total_step 25220, total loss is 29.27 , inference loss is 16.83, weight deacy loss is 12.44, training accuracy is 0.093750, time 56.500 samples/sec
+epoch 1, total_step 25240, total loss is 28.34 , inference loss is 15.90, weight deacy loss is 12.44, training accuracy is 0.156250, time 55.852 samples/sec
+epoch 1, total_step 25260, total loss is 27.97 , inference loss is 15.54, weight deacy loss is 12.44, training accuracy is 0.125000, time 56.640 samples/sec
+epoch 1, total_step 25280, total loss is 28.58 , inference loss is 16.15, weight deacy loss is 12.43, training accuracy is 0.156250, time 55.854 samples/sec
+epoch 1, total_step 25300, total loss is 24.97 , inference loss is 12.53, weight deacy loss is 12.43, training accuracy is 0.187500, time 55.695 samples/sec
+epoch 1, total_step 25320, total loss is 28.30 , inference loss is 15.87, weight deacy loss is 12.43, training accuracy is 0.093750, time 53.283 samples/sec
+epoch 1, total_step 25340, total loss is 32.54 , inference loss is 20.11, weight deacy loss is 12.43, training accuracy is 0.125000, time 55.681 samples/sec
+epoch 1, total_step 25360, total loss is 31.60 , inference loss is 19.18, weight deacy loss is 12.43, training accuracy is 0.093750, time 56.045 samples/sec
+epoch 1, total_step 25380, total loss is 25.38 , inference loss is 12.96, weight deacy loss is 12.43, training accuracy is 0.125000, time 55.652 samples/sec
+epoch 1, total_step 25400, total loss is 28.60 , inference loss is 16.17, weight deacy loss is 12.42, training accuracy is 0.093750, time 56.004 samples/sec
+epoch 1, total_step 25420, total loss is 25.83 , inference loss is 13.41, weight deacy loss is 12.42, training accuracy is 0.062500, time 53.769 samples/sec
+epoch 1, total_step 25440, total loss is 29.10 , inference loss is 16.68, weight deacy loss is 12.42, training accuracy is 0.125000, time 56.559 samples/sec
+epoch 1, total_step 25460, total loss is 28.15 , inference loss is 15.73, weight deacy loss is 12.42, training accuracy is 0.125000, time 54.372 samples/sec
+epoch 1, total_step 25480, total loss is 31.50 , inference loss is 19.08, weight deacy loss is 12.42, training accuracy is 0.093750, time 56.372 samples/sec
+epoch 1, total_step 25500, total loss is 29.36 , inference loss is 16.95, weight deacy loss is 12.42, training accuracy is 0.125000, time 56.013 samples/sec
+epoch 1, total_step 25520, total loss is 38.39 , inference loss is 25.98, weight deacy loss is 12.41, training accuracy is 0.031250, time 54.540 samples/sec
+epoch 1, total_step 25540, total loss is 32.17 , inference loss is 19.76, weight deacy loss is 12.41, training accuracy is 0.156250, time 56.204 samples/sec
+epoch 1, total_step 25560, total loss is 31.90 , inference loss is 19.49, weight deacy loss is 12.41, training accuracy is 0.125000, time 56.016 samples/sec
+epoch 1, total_step 25580, total loss is 29.20 , inference loss is 16.79, weight deacy loss is 12.41, training accuracy is 0.093750, time 56.030 samples/sec
+epoch 1, total_step 25600, total loss is 32.23 , inference loss is 19.82, weight deacy loss is 12.41, training accuracy is 0.156250, time 55.560 samples/sec
+epoch 1, total_step 25620, total loss is 27.67 , inference loss is 15.26, weight deacy loss is 12.41, training accuracy is 0.156250, time 56.233 samples/sec
+epoch 1, total_step 25640, total loss is 30.64 , inference loss is 18.24, weight deacy loss is 12.40, training accuracy is 0.031250, time 56.694 samples/sec
+epoch 1, total_step 25660, total loss is 32.24 , inference loss is 19.84, weight deacy loss is 12.40, training accuracy is 0.093750, time 56.488 samples/sec
+epoch 1, total_step 25680, total loss is 26.40 , inference loss is 14.00, weight deacy loss is 12.40, training accuracy is 0.218750, time 56.549 samples/sec
+epoch 1, total_step 25700, total loss is 31.25 , inference loss is 18.85, weight deacy loss is 12.40, training accuracy is 0.093750, time 54.421 samples/sec
+epoch 1, total_step 25720, total loss is 30.05 , inference loss is 17.65, weight deacy loss is 12.40, training accuracy is 0.031250, time 56.197 samples/sec
+epoch 1, total_step 25740, total loss is 30.22 , inference loss is 17.82, weight deacy loss is 12.39, training accuracy is 0.125000, time 56.485 samples/sec
+epoch 1, total_step 25760, total loss is 32.39 , inference loss is 20.00, weight deacy loss is 12.39, training accuracy is 0.062500, time 53.376 samples/sec
+epoch 1, total_step 25780, total loss is 25.14 , inference loss is 12.75, weight deacy loss is 12.39, training accuracy is 0.218750, time 56.330 samples/sec
+epoch 1, total_step 25800, total loss is 31.49 , inference loss is 19.10, weight deacy loss is 12.39, training accuracy is 0.062500, time 55.366 samples/sec
+epoch 1, total_step 25820, total loss is 26.89 , inference loss is 14.50, weight deacy loss is 12.39, training accuracy is 0.187500, time 55.941 samples/sec
+epoch 1, total_step 25840, total loss is 30.90 , inference loss is 18.51, weight deacy loss is 12.39, training accuracy is 0.093750, time 56.464 samples/sec
+epoch 1, total_step 25860, total loss is 27.78 , inference loss is 15.39, weight deacy loss is 12.38, training accuracy is 0.187500, time 55.831 samples/sec
+epoch 1, total_step 25880, total loss is 31.08 , inference loss is 18.69, weight deacy loss is 12.38, training accuracy is 0.062500, time 56.749 samples/sec
+epoch 1, total_step 25900, total loss is 28.64 , inference loss is 16.25, weight deacy loss is 12.38, training accuracy is 0.062500, time 54.149 samples/sec
+epoch 1, total_step 25920, total loss is 30.34 , inference loss is 17.96, weight deacy loss is 12.38, training accuracy is 0.031250, time 56.370 samples/sec
+epoch 1, total_step 25940, total loss is 32.62 , inference loss is 20.24, weight deacy loss is 12.38, training accuracy is 0.125000, time 55.963 samples/sec
+epoch 1, total_step 25960, total loss is 28.06 , inference loss is 15.68, weight deacy loss is 12.38, training accuracy is 0.125000, time 55.710 samples/sec
+epoch 1, total_step 25980, total loss is 29.51 , inference loss is 17.14, weight deacy loss is 12.37, training accuracy is 0.156250, time 56.003 samples/sec
+testing verification..
+(12000, 512)
+infer time 75.27866699999998
+best_threshold_index 104 0.9696296296296296
+best_threshold_index 109 0.9688888888888889
+best_threshold_index 104 0.9712962962962963
+best_threshold_index 102 0.9703703703703703
+best_threshold_index 102 0.9696296296296296
+best_threshold_index 102 0.9679629629629629
+best_threshold_index 102 0.9692592592592593
+best_threshold_index 102 0.9687037037037037
+best_threshold_index 102 0.9668518518518519
+best_threshold_index 102 0.9685185185185186
+[lfw][26000]XNorm: 23.360109
+[lfw][26000]Accuracy-Flip: 0.96667+-0.01145
+testing verification..
+(14000, 512)
+infer time 87.93837100000012
+best_threshold_index 111 0.9687301587301588
+best_threshold_index 111 0.9674603174603175
+best_threshold_index 110 0.9674603174603175
+best_threshold_index 111 0.966984126984127
+best_threshold_index 111 0.967936507936508
+best_threshold_index 111 0.9674603174603175
+best_threshold_index 111 0.967936507936508
+best_threshold_index 111 0.9673015873015873
+best_threshold_index 111 0.9688888888888889
+best_threshold_index 111 0.9684126984126984
+[cfp_ff][26000]XNorm: 23.019888
+[cfp_ff][26000]Accuracy-Flip: 0.96743+-0.00534
+testing verification..
+(14000, 512)
+infer time 87.46562500000007
+best_threshold_index 128 0.8511111111111112
+best_threshold_index 131 0.8493650793650793
+best_threshold_index 128 0.8501587301587301
+best_threshold_index 128 0.8520634920634921
+best_threshold_index 127 0.8539682539682539
+best_threshold_index 127 0.8482539682539683
+best_threshold_index 128 0.8484126984126984
+best_threshold_index 128 0.8517460317460317
+best_threshold_index 130 0.8515873015873016
+best_threshold_index 130 0.8517460317460317
+[cfp_fp][26000]XNorm: 22.257646
+[cfp_fp][26000]Accuracy-Flip: 0.84686+-0.01521
+testing verification..
+(12000, 512)
+infer time 74.56175999999996
+best_threshold_index 132 0.8212962962962963
+best_threshold_index 131 0.8198148148148148
+best_threshold_index 131 0.8237037037037037
+best_threshold_index 128 0.8183333333333334
+best_threshold_index 131 0.8275925925925925
+best_threshold_index 131 0.8192592592592592
+best_threshold_index 131 0.8237037037037037
+best_threshold_index 131 0.8244444444444444
+best_threshold_index 131 0.8253703703703704
+best_threshold_index 131 0.8216666666666667
+[agedb_30][26000]XNorm: 22.465840
+[agedb_30][26000]Accuracy-Flip: 0.82033+-0.02290
+test accuracy is: 0.9666666666666666
+epoch 1, total_step 26000, total loss is 33.30 , inference loss is 20.93, weight deacy loss is 12.37, training accuracy is 0.093750, time 53.486 samples/sec
+epoch 1, total_step 26020, total loss is 31.52 , inference loss is 19.14, weight deacy loss is 12.37, training accuracy is 0.093750, time 56.742 samples/sec
+epoch 1, total_step 26040, total loss is 28.96 , inference loss is 16.59, weight deacy loss is 12.37, training accuracy is 0.125000, time 56.339 samples/sec
+epoch 1, total_step 26060, total loss is 31.32 , inference loss is 18.96, weight deacy loss is 12.37, training accuracy is 0.156250, time 56.435 samples/sec
+epoch 1, total_step 26080, total loss is 27.74 , inference loss is 15.37, weight deacy loss is 12.37, training accuracy is 0.093750, time 55.972 samples/sec
+epoch 1, total_step 26100, total loss is 27.41 , inference loss is 15.05, weight deacy loss is 12.36, training accuracy is 0.187500, time 55.678 samples/sec
+epoch 1, total_step 26120, total loss is 29.37 , inference loss is 17.00, weight deacy loss is 12.36, training accuracy is 0.093750, time 56.997 samples/sec
+epoch 1, total_step 26140, total loss is 28.15 , inference loss is 15.79, weight deacy loss is 12.36, training accuracy is 0.125000, time 56.094 samples/sec
+epoch 1, total_step 26160, total loss is 32.04 , inference loss is 19.69, weight deacy loss is 12.36, training accuracy is 0.156250, time 55.930 samples/sec
+epoch 1, total_step 26180, total loss is 33.05 , inference loss is 20.69, weight deacy loss is 12.36, training accuracy is 0.062500, time 56.030 samples/sec
+epoch 1, total_step 26200, total loss is 27.51 , inference loss is 15.16, weight deacy loss is 12.36, training accuracy is 0.187500, time 55.931 samples/sec
+epoch 1, total_step 26220, total loss is 32.01 , inference loss is 19.66, weight deacy loss is 12.35, training accuracy is 0.062500, time 56.448 samples/sec
+epoch 1, total_step 26240, total loss is 29.91 , inference loss is 17.55, weight deacy loss is 12.35, training accuracy is 0.093750, time 56.097 samples/sec
+epoch 1, total_step 26260, total loss is 29.30 , inference loss is 16.95, weight deacy loss is 12.35, training accuracy is 0.156250, time 56.185 samples/sec
+epoch 1, total_step 26280, total loss is 26.93 , inference loss is 14.58, weight deacy loss is 12.35, training accuracy is 0.093750, time 56.248 samples/sec
+epoch 1, total_step 26300, total loss is 33.35 , inference loss is 21.01, weight deacy loss is 12.35, training accuracy is 0.031250, time 56.222 samples/sec
+epoch 1, total_step 26320, total loss is 30.04 , inference loss is 17.70, weight deacy loss is 12.35, training accuracy is 0.062500, time 54.753 samples/sec
+epoch 1, total_step 26340, total loss is 28.38 , inference loss is 16.03, weight deacy loss is 12.35, training accuracy is 0.125000, time 56.152 samples/sec
+epoch 1, total_step 26360, total loss is 29.62 , inference loss is 17.28, weight deacy loss is 12.34, training accuracy is 0.218750, time 56.246 samples/sec
+epoch 1, total_step 26380, total loss is 28.78 , inference loss is 16.43, weight deacy loss is 12.34, training accuracy is 0.156250, time 56.275 samples/sec
+epoch 1, total_step 26400, total loss is 33.46 , inference loss is 21.12, weight deacy loss is 12.34, training accuracy is 0.031250, time 56.027 samples/sec
+epoch 1, total_step 26420, total loss is 27.56 , inference loss is 15.22, weight deacy loss is 12.34, training accuracy is 0.156250, time 56.487 samples/sec
+epoch 1, total_step 26440, total loss is 27.39 , inference loss is 15.05, weight deacy loss is 12.34, training accuracy is 0.312500, time 55.798 samples/sec
+epoch 1, total_step 26460, total loss is 28.46 , inference loss is 16.12, weight deacy loss is 12.34, training accuracy is 0.187500, time 56.149 samples/sec
+epoch 1, total_step 26480, total loss is 28.59 , inference loss is 16.26, weight deacy loss is 12.33, training accuracy is 0.093750, time 56.817 samples/sec
+epoch 1, total_step 26500, total loss is 30.53 , inference loss is 18.19, weight deacy loss is 12.33, training accuracy is 0.250000, time 55.490 samples/sec
+epoch 1, total_step 26520, total loss is 33.94 , inference loss is 21.61, weight deacy loss is 12.33, training accuracy is 0.093750, time 56.495 samples/sec
+epoch 1, total_step 26540, total loss is 33.50 , inference loss is 21.17, weight deacy loss is 12.33, training accuracy is 0.062500, time 55.830 samples/sec
+epoch 1, total_step 26560, total loss is 31.31 , inference loss is 18.99, weight deacy loss is 12.33, training accuracy is 0.125000, time 55.767 samples/sec
+epoch 1, total_step 26580, total loss is 33.24 , inference loss is 20.92, weight deacy loss is 12.33, training accuracy is 0.000000, time 56.459 samples/sec
+epoch 1, total_step 26600, total loss is 27.48 , inference loss is 15.16, weight deacy loss is 12.32, training accuracy is 0.125000, time 56.493 samples/sec
+epoch 1, total_step 26620, total loss is 29.29 , inference loss is 16.97, weight deacy loss is 12.32, training accuracy is 0.000000, time 56.356 samples/sec
+epoch 1, total_step 26640, total loss is 28.57 , inference loss is 16.25, weight deacy loss is 12.32, training accuracy is 0.125000, time 56.346 samples/sec
+epoch 1, total_step 26660, total loss is 27.64 , inference loss is 15.32, weight deacy loss is 12.32, training accuracy is 0.093750, time 56.962 samples/sec
+epoch 1, total_step 26680, total loss is 28.14 , inference loss is 15.82, weight deacy loss is 12.32, training accuracy is 0.062500, time 56.408 samples/sec
+epoch 1, total_step 26700, total loss is 31.88 , inference loss is 19.56, weight deacy loss is 12.32, training accuracy is 0.062500, time 55.336 samples/sec
+epoch 1, total_step 26720, total loss is 32.08 , inference loss is 19.77, weight deacy loss is 12.31, training accuracy is 0.031250, time 56.144 samples/sec
+epoch 1, total_step 26740, total loss is 30.55 , inference loss is 18.24, weight deacy loss is 12.31, training accuracy is 0.156250, time 56.397 samples/sec
+epoch 1, total_step 26760, total loss is 32.91 , inference loss is 20.60, weight deacy loss is 12.31, training accuracy is 0.062500, time 55.482 samples/sec
+epoch 1, total_step 26780, total loss is 34.21 , inference loss is 21.90, weight deacy loss is 12.31, training accuracy is 0.062500, time 55.185 samples/sec
+epoch 1, total_step 26800, total loss is 29.02 , inference loss is 16.71, weight deacy loss is 12.31, training accuracy is 0.187500, time 56.338 samples/sec
+epoch 1, total_step 26820, total loss is 33.11 , inference loss is 20.80, weight deacy loss is 12.31, training accuracy is 0.062500, time 56.286 samples/sec
+epoch 1, total_step 26840, total loss is 31.43 , inference loss is 19.12, weight deacy loss is 12.30, training accuracy is 0.031250, time 56.362 samples/sec
+epoch 1, total_step 26860, total loss is 33.01 , inference loss is 20.71, weight deacy loss is 12.30, training accuracy is 0.000000, time 55.526 samples/sec
+epoch 1, total_step 26880, total loss is 30.59 , inference loss is 18.29, weight deacy loss is 12.30, training accuracy is 0.125000, time 56.309 samples/sec
+epoch 1, total_step 26900, total loss is 29.29 , inference loss is 17.00, weight deacy loss is 12.30, training accuracy is 0.156250, time 57.314 samples/sec
+epoch 1, total_step 26920, total loss is 30.51 , inference loss is 18.22, weight deacy loss is 12.30, training accuracy is 0.062500, time 56.087 samples/sec
+epoch 1, total_step 26940, total loss is 34.04 , inference loss is 21.74, weight deacy loss is 12.30, training accuracy is 0.062500, time 56.300 samples/sec
+epoch 1, total_step 26960, total loss is 29.98 , inference loss is 17.69, weight deacy loss is 12.29, training accuracy is 0.125000, time 55.799 samples/sec
+epoch 1, total_step 26980, total loss is 31.40 , inference loss is 19.11, weight deacy loss is 12.29, training accuracy is 0.031250, time 57.601 samples/sec
+epoch 1, total_step 27000, total loss is 25.45 , inference loss is 13.16, weight deacy loss is 12.29, training accuracy is 0.093750, time 55.503 samples/sec
+epoch 1, total_step 27020, total loss is 28.80 , inference loss is 16.51, weight deacy loss is 12.29, training accuracy is 0.093750, time 55.739 samples/sec
+epoch 1, total_step 27040, total loss is 35.52 , inference loss is 23.23, weight deacy loss is 12.29, training accuracy is 0.125000, time 57.247 samples/sec
+epoch 1, total_step 27060, total loss is 28.29 , inference loss is 16.01, weight deacy loss is 12.29, training accuracy is 0.062500, time 55.981 samples/sec
+epoch 1, total_step 27080, total loss is 32.93 , inference loss is 20.65, weight deacy loss is 12.28, training accuracy is 0.062500, time 56.580 samples/sec
+epoch 1, total_step 27100, total loss is 28.57 , inference loss is 16.29, weight deacy loss is 12.28, training accuracy is 0.125000, time 57.519 samples/sec
+epoch 1, total_step 27120, total loss is 33.36 , inference loss is 21.08, weight deacy loss is 12.28, training accuracy is 0.062500, time 56.511 samples/sec
+epoch 1, total_step 27140, total loss is 27.22 , inference loss is 14.94, weight deacy loss is 12.28, training accuracy is 0.031250, time 56.414 samples/sec
+epoch 1, total_step 27160, total loss is 30.82 , inference loss is 18.54, weight deacy loss is 12.28, training accuracy is 0.062500, time 56.927 samples/sec
+epoch 1, total_step 27180, total loss is 31.88 , inference loss is 19.60, weight deacy loss is 12.28, training accuracy is 0.031250, time 53.914 samples/sec
+epoch 1, total_step 27200, total loss is 30.50 , inference loss is 18.22, weight deacy loss is 12.27, training accuracy is 0.093750, time 54.912 samples/sec
+epoch 1, total_step 27220, total loss is 29.96 , inference loss is 17.68, weight deacy loss is 12.27, training accuracy is 0.125000, time 56.195 samples/sec
+epoch 1, total_step 27240, total loss is 28.73 , inference loss is 16.46, weight deacy loss is 12.27, training accuracy is 0.093750, time 56.337 samples/sec
+epoch 1, total_step 27260, total loss is 33.63 , inference loss is 21.36, weight deacy loss is 12.27, training accuracy is 0.000000, time 56.470 samples/sec
+epoch 1, total_step 27280, total loss is 29.31 , inference loss is 17.05, weight deacy loss is 12.27, training accuracy is 0.156250, time 55.901 samples/sec
+epoch 1, total_step 27300, total loss is 29.27 , inference loss is 17.00, weight deacy loss is 12.27, training accuracy is 0.093750, time 55.076 samples/sec
+epoch 1, total_step 27320, total loss is 31.93 , inference loss is 19.67, weight deacy loss is 12.26, training accuracy is 0.031250, time 56.478 samples/sec
+epoch 1, total_step 27340, total loss is 30.40 , inference loss is 18.13, weight deacy loss is 12.26, training accuracy is 0.187500, time 55.048 samples/sec
+epoch 1, total_step 27360, total loss is 32.93 , inference loss is 20.67, weight deacy loss is 12.26, training accuracy is 0.062500, time 56.302 samples/sec
+epoch 1, total_step 27380, total loss is 30.63 , inference loss is 18.37, weight deacy loss is 12.26, training accuracy is 0.125000, time 55.647 samples/sec
+epoch 1, total_step 27400, total loss is 33.17 , inference loss is 20.91, weight deacy loss is 12.26, training accuracy is 0.062500, time 56.106 samples/sec
+epoch 1, total_step 27420, total loss is 34.65 , inference loss is 22.39, weight deacy loss is 12.26, training accuracy is 0.031250, time 56.940 samples/sec
+epoch 1, total_step 27440, total loss is 28.80 , inference loss is 16.55, weight deacy loss is 12.25, training accuracy is 0.093750, time 55.898 samples/sec
+epoch 1, total_step 27460, total loss is 30.99 , inference loss is 18.73, weight deacy loss is 12.25, training accuracy is 0.093750, time 56.580 samples/sec
+epoch 1, total_step 27480, total loss is 30.60 , inference loss is 18.35, weight deacy loss is 12.25, training accuracy is 0.031250, time 55.825 samples/sec
+epoch 1, total_step 27500, total loss is 30.70 , inference loss is 18.45, weight deacy loss is 12.25, training accuracy is 0.187500, time 56.722 samples/sec
+epoch 1, total_step 27520, total loss is 29.08 , inference loss is 16.83, weight deacy loss is 12.25, training accuracy is 0.031250, time 56.301 samples/sec
+epoch 1, total_step 27540, total loss is 30.96 , inference loss is 18.71, weight deacy loss is 12.25, training accuracy is 0.093750, time 56.440 samples/sec
+epoch 1, total_step 27560, total loss is 25.14 , inference loss is 12.90, weight deacy loss is 12.25, training accuracy is 0.125000, time 55.592 samples/sec
+epoch 1, total_step 27580, total loss is 31.83 , inference loss is 19.58, weight deacy loss is 12.24, training accuracy is 0.093750, time 55.978 samples/sec
+epoch 1, total_step 27600, total loss is 31.96 , inference loss is 19.72, weight deacy loss is 12.24, training accuracy is 0.187500, time 55.957 samples/sec
+epoch 1, total_step 27620, total loss is 29.81 , inference loss is 17.57, weight deacy loss is 12.24, training accuracy is 0.031250, time 54.227 samples/sec
+epoch 1, total_step 27640, total loss is 31.05 , inference loss is 18.81, weight deacy loss is 12.24, training accuracy is 0.031250, time 56.506 samples/sec
+epoch 1, total_step 27660, total loss is 29.86 , inference loss is 17.63, weight deacy loss is 12.24, training accuracy is 0.093750, time 56.013 samples/sec
+epoch 1, total_step 27680, total loss is 31.74 , inference loss is 19.51, weight deacy loss is 12.24, training accuracy is 0.062500, time 55.313 samples/sec
+epoch 1, total_step 27700, total loss is 30.66 , inference loss is 18.43, weight deacy loss is 12.23, training accuracy is 0.031250, time 55.679 samples/sec
+epoch 1, total_step 27720, total loss is 30.04 , inference loss is 17.80, weight deacy loss is 12.23, training accuracy is 0.062500, time 56.331 samples/sec
+epoch 1, total_step 27740, total loss is 30.00 , inference loss is 17.77, weight deacy loss is 12.23, training accuracy is 0.062500, time 55.683 samples/sec
+epoch 1, total_step 27760, total loss is 32.47 , inference loss is 20.25, weight deacy loss is 12.23, training accuracy is 0.031250, time 56.269 samples/sec
+epoch 1, total_step 27780, total loss is 30.43 , inference loss is 18.20, weight deacy loss is 12.23, training accuracy is 0.000000, time 56.693 samples/sec
+epoch 1, total_step 27800, total loss is 32.70 , inference loss is 20.47, weight deacy loss is 12.23, training accuracy is 0.031250, time 55.802 samples/sec
+epoch 1, total_step 27820, total loss is 31.75 , inference loss is 19.52, weight deacy loss is 12.22, training accuracy is 0.156250, time 56.471 samples/sec
+epoch 1, total_step 27840, total loss is 29.22 , inference loss is 17.00, weight deacy loss is 12.22, training accuracy is 0.125000, time 55.976 samples/sec
+epoch 1, total_step 27860, total loss is 31.31 , inference loss is 19.08, weight deacy loss is 12.22, training accuracy is 0.031250, time 56.800 samples/sec
+epoch 1, total_step 27880, total loss is 28.74 , inference loss is 16.52, weight deacy loss is 12.22, training accuracy is 0.093750, time 56.450 samples/sec
+epoch 1, total_step 27900, total loss is 32.60 , inference loss is 20.39, weight deacy loss is 12.22, training accuracy is 0.062500, time 56.516 samples/sec
+epoch 1, total_step 27920, total loss is 30.67 , inference loss is 18.45, weight deacy loss is 12.22, training accuracy is 0.031250, time 55.544 samples/sec
+epoch 1, total_step 27940, total loss is 28.47 , inference loss is 16.25, weight deacy loss is 12.21, training accuracy is 0.125000, time 56.489 samples/sec
+epoch 1, total_step 27960, total loss is 30.86 , inference loss is 18.64, weight deacy loss is 12.21, training accuracy is 0.062500, time 55.867 samples/sec
+epoch 1, total_step 27980, total loss is 28.14 , inference loss is 15.93, weight deacy loss is 12.21, training accuracy is 0.125000, time 56.154 samples/sec
+testing verification..
+(12000, 512)
+infer time 75.04763499999999
+best_threshold_index 99 0.9651851851851851
+best_threshold_index 104 0.9653703703703703
+best_threshold_index 99 0.9666666666666667
+best_threshold_index 100 0.9668518518518519
+best_threshold_index 99 0.9655555555555555
+best_threshold_index 99 0.9642592592592593
+best_threshold_index 100 0.965925925925926
+best_threshold_index 98 0.965
+best_threshold_index 99 0.9640740740740741
+best_threshold_index 99 0.9633333333333334
+[lfw][28000]XNorm: 25.117473
+[lfw][28000]Accuracy-Flip: 0.96383+-0.00963
+testing verification..
+(14000, 512)
+infer time 87.41785199999997
+best_threshold_index 108 0.9693650793650793
+best_threshold_index 108 0.9692063492063492
+best_threshold_index 108 0.97
+best_threshold_index 108 0.9687301587301588
+best_threshold_index 108 0.9687301587301588
+best_threshold_index 107 0.9680952380952381
+best_threshold_index 108 0.9682539682539683
+best_threshold_index 108 0.9671428571428572
+best_threshold_index 107 0.9684126984126984
+best_threshold_index 108 0.9684126984126984
+[cfp_ff][28000]XNorm: 24.156003
+[cfp_ff][28000]Accuracy-Flip: 0.96743+-0.00648
+testing verification..
+(14000, 512)
+infer time 87.703439
+best_threshold_index 123 0.8465079365079365
+best_threshold_index 123 0.8476190476190476
+best_threshold_index 123 0.844920634920635
+best_threshold_index 123 0.8476190476190476
+best_threshold_index 123 0.8482539682539683
+best_threshold_index 123 0.8433333333333334
+best_threshold_index 123 0.8417460317460318
+best_threshold_index 123 0.8447619047619047
+best_threshold_index 123 0.8450793650793651
+best_threshold_index 123 0.8458730158730159
+[cfp_fp][28000]XNorm: 23.550098
+[cfp_fp][28000]Accuracy-Flip: 0.84557+-0.01743
+testing verification..
+(12000, 512)
+infer time 74.94469899999994
+best_threshold_index 128 0.8331481481481482
+best_threshold_index 128 0.8357407407407408
+best_threshold_index 128 0.8351851851851851
+best_threshold_index 125 0.8361111111111111
+best_threshold_index 125 0.8407407407407408
+best_threshold_index 125 0.8344444444444444
+best_threshold_index 128 0.8372222222222222
+best_threshold_index 128 0.8401851851851851
+best_threshold_index 128 0.8362962962962963
+best_threshold_index 128 0.8361111111111111
+[agedb_30][28000]XNorm: 23.781962
+[agedb_30][28000]Accuracy-Flip: 0.83433+-0.02090
+test accuracy is: 0.9638333333333333
+epoch 1, total_step 28000, total loss is 30.52 , inference loss is 18.31, weight deacy loss is 12.21, training accuracy is 0.062500, time 53.511 samples/sec
+epoch 1, total_step 28020, total loss is 25.95 , inference loss is 13.74, weight deacy loss is 12.21, training accuracy is 0.062500, time 55.979 samples/sec
+epoch 1, total_step 28040, total loss is 26.82 , inference loss is 14.61, weight deacy loss is 12.21, training accuracy is 0.093750, time 56.352 samples/sec
+epoch 1, total_step 28060, total loss is 31.15 , inference loss is 18.95, weight deacy loss is 12.20, training accuracy is 0.093750, time 55.930 samples/sec
+epoch 1, total_step 28080, total loss is 32.18 , inference loss is 19.98, weight deacy loss is 12.20, training accuracy is 0.062500, time 56.436 samples/sec
+epoch 1, total_step 28100, total loss is 33.41 , inference loss is 21.21, weight deacy loss is 12.20, training accuracy is 0.093750, time 55.992 samples/sec
+epoch 1, total_step 28120, total loss is 34.43 , inference loss is 22.23, weight deacy loss is 12.20, training accuracy is 0.031250, time 56.823 samples/sec
+epoch 1, total_step 28140, total loss is 29.32 , inference loss is 17.12, weight deacy loss is 12.20, training accuracy is 0.062500, time 56.298 samples/sec
+epoch 1, total_step 28160, total loss is 30.98 , inference loss is 18.78, weight deacy loss is 12.20, training accuracy is 0.093750, time 56.655 samples/sec
+epoch 1, total_step 28180, total loss is 34.56 , inference loss is 22.36, weight deacy loss is 12.20, training accuracy is 0.031250, time 56.309 samples/sec
+epoch 1, total_step 28200, total loss is 30.72 , inference loss is 18.52, weight deacy loss is 12.19, training accuracy is 0.062500, time 56.103 samples/sec
+epoch 1, total_step 28220, total loss is 28.26 , inference loss is 16.07, weight deacy loss is 12.19, training accuracy is 0.093750, time 55.458 samples/sec
+epoch 1, total_step 28240, total loss is 31.92 , inference loss is 19.73, weight deacy loss is 12.19, training accuracy is 0.062500, time 55.410 samples/sec
+epoch 1, total_step 28260, total loss is 34.16 , inference loss is 21.98, weight deacy loss is 12.19, training accuracy is 0.062500, time 55.780 samples/sec
+epoch 1, total_step 28280, total loss is 31.50 , inference loss is 19.31, weight deacy loss is 12.19, training accuracy is 0.062500, time 56.822 samples/sec
+epoch 1, total_step 28300, total loss is 30.66 , inference loss is 18.48, weight deacy loss is 12.19, training accuracy is 0.031250, time 55.460 samples/sec
+epoch 1, total_step 28320, total loss is 32.74 , inference loss is 20.55, weight deacy loss is 12.18, training accuracy is 0.062500, time 56.479 samples/sec
+epoch 1, total_step 28340, total loss is 31.57 , inference loss is 19.39, weight deacy loss is 12.18, training accuracy is 0.031250, time 56.085 samples/sec
+epoch 1, total_step 28360, total loss is 29.53 , inference loss is 17.35, weight deacy loss is 12.18, training accuracy is 0.031250, time 57.110 samples/sec
+epoch 1, total_step 28380, total loss is 32.18 , inference loss is 20.00, weight deacy loss is 12.18, training accuracy is 0.093750, time 56.413 samples/sec
+epoch 1, total_step 28400, total loss is 29.91 , inference loss is 17.73, weight deacy loss is 12.18, training accuracy is 0.093750, time 56.066 samples/sec
+epoch 1, total_step 28420, total loss is 32.36 , inference loss is 20.19, weight deacy loss is 12.18, training accuracy is 0.062500, time 56.588 samples/sec
+epoch 1, total_step 28440, total loss is 29.34 , inference loss is 17.16, weight deacy loss is 12.17, training accuracy is 0.125000, time 56.160 samples/sec
+epoch 1, total_step 28460, total loss is 30.72 , inference loss is 18.55, weight deacy loss is 12.17, training accuracy is 0.125000, time 56.540 samples/sec
+epoch 1, total_step 28480, total loss is 32.35 , inference loss is 20.18, weight deacy loss is 12.17, training accuracy is 0.093750, time 55.619 samples/sec
+epoch 1, total_step 28500, total loss is 29.50 , inference loss is 17.33, weight deacy loss is 12.17, training accuracy is 0.000000, time 56.913 samples/sec
+epoch 1, total_step 28520, total loss is 30.11 , inference loss is 17.94, weight deacy loss is 12.17, training accuracy is 0.062500, time 55.586 samples/sec
+epoch 1, total_step 28540, total loss is 30.14 , inference loss is 17.98, weight deacy loss is 12.17, training accuracy is 0.031250, time 56.676 samples/sec
+epoch 1, total_step 28560, total loss is 28.22 , inference loss is 16.05, weight deacy loss is 12.17, training accuracy is 0.031250, time 56.514 samples/sec
+epoch 1, total_step 28580, total loss is 29.46 , inference loss is 17.30, weight deacy loss is 12.16, training accuracy is 0.093750, time 57.248 samples/sec
+epoch 1, total_step 28600, total loss is 31.43 , inference loss is 19.27, weight deacy loss is 12.16, training accuracy is 0.031250, time 55.221 samples/sec
+epoch 1, total_step 28620, total loss is 30.40 , inference loss is 18.24, weight deacy loss is 12.16, training accuracy is 0.062500, time 55.117 samples/sec
+epoch 1, total_step 28640, total loss is 34.89 , inference loss is 22.73, weight deacy loss is 12.16, training accuracy is 0.062500, time 55.985 samples/sec
+epoch 1, total_step 28660, total loss is 29.40 , inference loss is 17.24, weight deacy loss is 12.16, training accuracy is 0.093750, time 55.851 samples/sec
+epoch 1, total_step 28680, total loss is 35.19 , inference loss is 23.04, weight deacy loss is 12.16, training accuracy is 0.031250, time 57.003 samples/sec
+epoch 1, total_step 28700, total loss is 28.89 , inference loss is 16.73, weight deacy loss is 12.15, training accuracy is 0.031250, time 55.996 samples/sec
+epoch 1, total_step 28720, total loss is 29.89 , inference loss is 17.74, weight deacy loss is 12.15, training accuracy is 0.031250, time 56.248 samples/sec
+epoch 1, total_step 28740, total loss is 33.04 , inference loss is 20.89, weight deacy loss is 12.15, training accuracy is 0.093750, time 56.574 samples/sec
+epoch 1, total_step 28760, total loss is 31.29 , inference loss is 19.14, weight deacy loss is 12.15, training accuracy is 0.093750, time 56.813 samples/sec
+epoch 1, total_step 28780, total loss is 31.24 , inference loss is 19.09, weight deacy loss is 12.15, training accuracy is 0.125000, time 57.826 samples/sec
+epoch 1, total_step 28800, total loss is 30.48 , inference loss is 18.33, weight deacy loss is 12.15, training accuracy is 0.062500, time 56.382 samples/sec
+epoch 1, total_step 28820, total loss is 29.05 , inference loss is 16.91, weight deacy loss is 12.14, training accuracy is 0.125000, time 55.827 samples/sec
+epoch 1, total_step 28840, total loss is 31.85 , inference loss is 19.71, weight deacy loss is 12.14, training accuracy is 0.125000, time 55.698 samples/sec
+epoch 1, total_step 28860, total loss is 32.32 , inference loss is 20.18, weight deacy loss is 12.14, training accuracy is 0.031250, time 56.101 samples/sec
+epoch 1, total_step 28880, total loss is 32.92 , inference loss is 20.78, weight deacy loss is 12.14, training accuracy is 0.093750, time 57.262 samples/sec
+epoch 1, total_step 28900, total loss is 31.28 , inference loss is 19.15, weight deacy loss is 12.14, training accuracy is 0.093750, time 56.645 samples/sec
+epoch 1, total_step 28920, total loss is 31.03 , inference loss is 18.89, weight deacy loss is 12.14, training accuracy is 0.062500, time 55.745 samples/sec
+epoch 1, total_step 28940, total loss is 29.47 , inference loss is 17.33, weight deacy loss is 12.13, training accuracy is 0.093750, time 56.330 samples/sec
+epoch 1, total_step 28960, total loss is 33.37 , inference loss is 21.24, weight deacy loss is 12.13, training accuracy is 0.062500, time 56.032 samples/sec
+epoch 1, total_step 28980, total loss is 32.93 , inference loss is 20.80, weight deacy loss is 12.13, training accuracy is 0.093750, time 56.753 samples/sec
+epoch 1, total_step 29000, total loss is 33.38 , inference loss is 21.25, weight deacy loss is 12.13, training accuracy is 0.062500, time 55.754 samples/sec
+epoch 1, total_step 29020, total loss is 25.83 , inference loss is 13.70, weight deacy loss is 12.13, training accuracy is 0.125000, time 56.716 samples/sec
+epoch 1, total_step 29040, total loss is 30.15 , inference loss is 18.02, weight deacy loss is 12.13, training accuracy is 0.093750, time 56.035 samples/sec
+epoch 1, total_step 29060, total loss is 31.43 , inference loss is 19.30, weight deacy loss is 12.13, training accuracy is 0.031250, time 56.144 samples/sec
+epoch 1, total_step 29080, total loss is 29.33 , inference loss is 17.21, weight deacy loss is 12.12, training accuracy is 0.187500, time 57.044 samples/sec
+epoch 1, total_step 29100, total loss is 31.01 , inference loss is 18.89, weight deacy loss is 12.12, training accuracy is 0.031250, time 55.465 samples/sec
+epoch 1, total_step 29120, total loss is 31.29 , inference loss is 19.17, weight deacy loss is 12.12, training accuracy is 0.156250, time 56.055 samples/sec
+epoch 1, total_step 29140, total loss is 31.82 , inference loss is 19.70, weight deacy loss is 12.12, training accuracy is 0.062500, time 55.105 samples/sec
+epoch 1, total_step 29160, total loss is 33.44 , inference loss is 21.32, weight deacy loss is 12.12, training accuracy is 0.031250, time 56.802 samples/sec
+epoch 1, total_step 29180, total loss is 28.03 , inference loss is 15.92, weight deacy loss is 12.12, training accuracy is 0.093750, time 56.530 samples/sec
+epoch 1, total_step 29200, total loss is 30.01 , inference loss is 17.90, weight deacy loss is 12.11, training accuracy is 0.125000, time 56.451 samples/sec
+epoch 1, total_step 29220, total loss is 31.93 , inference loss is 19.81, weight deacy loss is 12.11, training accuracy is 0.156250, time 55.692 samples/sec
+epoch 1, total_step 29240, total loss is 31.31 , inference loss is 19.20, weight deacy loss is 12.11, training accuracy is 0.062500, time 56.332 samples/sec
+epoch 1, total_step 29260, total loss is 33.27 , inference loss is 21.16, weight deacy loss is 12.11, training accuracy is 0.062500, time 55.787 samples/sec
+epoch 1, total_step 29280, total loss is 29.12 , inference loss is 17.01, weight deacy loss is 12.11, training accuracy is 0.093750, time 56.124 samples/sec
+epoch 1, total_step 29300, total loss is 28.37 , inference loss is 16.26, weight deacy loss is 12.11, training accuracy is 0.062500, time 56.496 samples/sec
+epoch 1, total_step 29320, total loss is 31.84 , inference loss is 19.73, weight deacy loss is 12.10, training accuracy is 0.031250, time 55.242 samples/sec
+epoch 1, total_step 29340, total loss is 29.40 , inference loss is 17.30, weight deacy loss is 12.10, training accuracy is 0.000000, time 56.308 samples/sec
+epoch 1, total_step 29360, total loss is 29.99 , inference loss is 17.88, weight deacy loss is 12.10, training accuracy is 0.093750, time 57.125 samples/sec
+epoch 1, total_step 29380, total loss is 32.04 , inference loss is 19.94, weight deacy loss is 12.10, training accuracy is 0.093750, time 56.490 samples/sec
+epoch 1, total_step 29400, total loss is 34.95 , inference loss is 22.85, weight deacy loss is 12.10, training accuracy is 0.000000, time 55.815 samples/sec
+epoch 1, total_step 29420, total loss is 28.94 , inference loss is 16.84, weight deacy loss is 12.10, training accuracy is 0.156250, time 56.802 samples/sec
+epoch 1, total_step 29440, total loss is 28.03 , inference loss is 15.94, weight deacy loss is 12.10, training accuracy is 0.125000, time 55.657 samples/sec
+epoch 1, total_step 29460, total loss is 29.35 , inference loss is 17.25, weight deacy loss is 12.09, training accuracy is 0.062500, time 55.884 samples/sec
+epoch 1, total_step 29480, total loss is 26.19 , inference loss is 14.10, weight deacy loss is 12.09, training accuracy is 0.000000, time 54.987 samples/sec
+epoch 1, total_step 29500, total loss is 25.30 , inference loss is 13.21, weight deacy loss is 12.09, training accuracy is 0.125000, time 55.818 samples/sec
+epoch 1, total_step 29520, total loss is 28.41 , inference loss is 16.32, weight deacy loss is 12.09, training accuracy is 0.062500, time 56.800 samples/sec
+epoch 1, total_step 29540, total loss is 28.91 , inference loss is 16.83, weight deacy loss is 12.09, training accuracy is 0.062500, time 56.719 samples/sec
+epoch 1, total_step 29560, total loss is 34.96 , inference loss is 22.87, weight deacy loss is 12.09, training accuracy is 0.000000, time 55.061 samples/sec
+epoch 1, total_step 29580, total loss is 29.90 , inference loss is 17.82, weight deacy loss is 12.08, training accuracy is 0.062500, time 55.015 samples/sec
+epoch 1, total_step 29600, total loss is 28.59 , inference loss is 16.51, weight deacy loss is 12.08, training accuracy is 0.093750, time 55.668 samples/sec
+epoch 1, total_step 29620, total loss is 34.01 , inference loss is 21.93, weight deacy loss is 12.08, training accuracy is 0.062500, time 55.857 samples/sec
+epoch 1, total_step 29640, total loss is 31.55 , inference loss is 19.47, weight deacy loss is 12.08, training accuracy is 0.125000, time 56.576 samples/sec
+epoch 1, total_step 29660, total loss is 28.49 , inference loss is 16.41, weight deacy loss is 12.08, training accuracy is 0.031250, time 56.275 samples/sec
+epoch 1, total_step 29680, total loss is 33.01 , inference loss is 20.94, weight deacy loss is 12.08, training accuracy is 0.000000, time 56.025 samples/sec
+epoch 1, total_step 29700, total loss is 30.37 , inference loss is 18.29, weight deacy loss is 12.07, training accuracy is 0.000000, time 55.015 samples/sec
+epoch 1, total_step 29720, total loss is 28.96 , inference loss is 16.88, weight deacy loss is 12.07, training accuracy is 0.125000, time 56.193 samples/sec
+epoch 1, total_step 29740, total loss is 31.48 , inference loss is 19.41, weight deacy loss is 12.07, training accuracy is 0.062500, time 55.756 samples/sec
+epoch 1, total_step 29760, total loss is 31.21 , inference loss is 19.14, weight deacy loss is 12.07, training accuracy is 0.062500, time 56.286 samples/sec
+epoch 1, total_step 29780, total loss is 30.36 , inference loss is 18.29, weight deacy loss is 12.07, training accuracy is 0.062500, time 55.007 samples/sec
+epoch 1, total_step 29800, total loss is 32.51 , inference loss is 20.44, weight deacy loss is 12.07, training accuracy is 0.062500, time 55.619 samples/sec
+epoch 1, total_step 29820, total loss is 29.99 , inference loss is 17.92, weight deacy loss is 12.07, training accuracy is 0.062500, time 56.194 samples/sec
+epoch 1, total_step 29840, total loss is 31.09 , inference loss is 19.03, weight deacy loss is 12.06, training accuracy is 0.000000, time 56.446 samples/sec
+epoch 1, total_step 29860, total loss is 32.20 , inference loss is 20.14, weight deacy loss is 12.06, training accuracy is 0.000000, time 55.881 samples/sec
+epoch 1, total_step 29880, total loss is 31.99 , inference loss is 19.93, weight deacy loss is 12.06, training accuracy is 0.031250, time 55.961 samples/sec
+epoch 1, total_step 29900, total loss is 33.49 , inference loss is 21.43, weight deacy loss is 12.06, training accuracy is 0.031250, time 56.060 samples/sec
+epoch 1, total_step 29920, total loss is 28.36 , inference loss is 16.30, weight deacy loss is 12.06, training accuracy is 0.125000, time 55.775 samples/sec
+epoch 1, total_step 29940, total loss is 32.69 , inference loss is 20.63, weight deacy loss is 12.06, training accuracy is 0.062500, time 56.250 samples/sec
+epoch 1, total_step 29960, total loss is 31.92 , inference loss is 19.86, weight deacy loss is 12.05, training accuracy is 0.000000, time 55.453 samples/sec
+epoch 1, total_step 29980, total loss is 29.33 , inference loss is 17.28, weight deacy loss is 12.05, training accuracy is 0.031250, time 56.010 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.38361300000004
+best_threshold_index 98 0.9683333333333334
+best_threshold_index 98 0.9685185185185186
+best_threshold_index 98 0.9685185185185186
+best_threshold_index 98 0.9701851851851852
+best_threshold_index 98 0.9685185185185186
+best_threshold_index 98 0.967037037037037
+best_threshold_index 98 0.9681481481481482
+best_threshold_index 98 0.9675925925925926
+best_threshold_index 98 0.9668518518518519
+best_threshold_index 98 0.9679629629629629
+[lfw][30000]XNorm: 24.519175
+[lfw][30000]Accuracy-Flip: 0.96817+-0.00797
+testing verification..
+(14000, 512)
+infer time 86.81265400000011
+best_threshold_index 109 0.9712698412698413
+best_threshold_index 108 0.9685714285714285
+best_threshold_index 109 0.9693650793650793
+best_threshold_index 109 0.969047619047619
+best_threshold_index 109 0.9692063492063492
+best_threshold_index 109 0.969047619047619
+best_threshold_index 109 0.9703174603174604
+best_threshold_index 109 0.9680952380952381
+best_threshold_index 109 0.9693650793650793
+best_threshold_index 109 0.97
+[cfp_ff][30000]XNorm: 24.017048
+[cfp_ff][30000]Accuracy-Flip: 0.96914+-0.00748
+testing verification..
+(14000, 512)
+infer time 86.74146
+best_threshold_index 125 0.846031746031746
+best_threshold_index 125 0.8442857142857143
+best_threshold_index 125 0.8436507936507937
+best_threshold_index 125 0.8480952380952381
+best_threshold_index 125 0.849047619047619
+best_threshold_index 125 0.8398412698412698
+best_threshold_index 125 0.8411111111111111
+best_threshold_index 125 0.8458730158730159
+best_threshold_index 125 0.8447619047619047
+best_threshold_index 125 0.8458730158730159
+[cfp_fp][30000]XNorm: 23.253498
+[cfp_fp][30000]Accuracy-Flip: 0.84486+-0.02423
+testing verification..
+(12000, 512)
+infer time 73.636611
+best_threshold_index 127 0.8490740740740741
+best_threshold_index 127 0.8468518518518519
+best_threshold_index 127 0.8494444444444444
+best_threshold_index 124 0.8490740740740741
+best_threshold_index 126 0.8538888888888889
+best_threshold_index 127 0.8479629629629629
+best_threshold_index 123 0.8483333333333334
+best_threshold_index 127 0.855
+best_threshold_index 131 0.8522222222222222
+best_threshold_index 131 0.8505555555555555
+[agedb_30][30000]XNorm: 23.465841
+[agedb_30][30000]Accuracy-Flip: 0.84433+-0.02244
+test accuracy is: 0.9681666666666668
+epoch 1, total_step 30000, total loss is 33.06 , inference loss is 21.01, weight deacy loss is 12.05, training accuracy is 0.000000, time 53.578 samples/sec
+epoch 1, total_step 30020, total loss is 32.14 , inference loss is 20.09, weight deacy loss is 12.05, training accuracy is 0.062500, time 56.414 samples/sec
+epoch 1, total_step 30040, total loss is 32.27 , inference loss is 20.22, weight deacy loss is 12.05, training accuracy is 0.187500, time 56.520 samples/sec
+epoch 1, total_step 30060, total loss is 32.71 , inference loss is 20.66, weight deacy loss is 12.05, training accuracy is 0.031250, time 58.023 samples/sec
+epoch 1, total_step 30080, total loss is 28.75 , inference loss is 16.70, weight deacy loss is 12.05, training accuracy is 0.031250, time 57.134 samples/sec
+epoch 1, total_step 30100, total loss is 30.05 , inference loss is 18.00, weight deacy loss is 12.04, training accuracy is 0.125000, time 56.633 samples/sec
+epoch 1, total_step 30120, total loss is 27.10 , inference loss is 15.06, weight deacy loss is 12.04, training accuracy is 0.031250, time 54.953 samples/sec
+epoch 1, total_step 30140, total loss is 33.69 , inference loss is 21.65, weight deacy loss is 12.04, training accuracy is 0.062500, time 55.469 samples/sec2022-12-01 04:06:31.143472: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:111] Filling up shuffle buffer (this may take a while): 7926 of 10000
+2022-12-01 04:06:33.480690: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:162] Shuffle buffer filled.
+
+epoch 1, total_step 30160, total loss is 28.07 , inference loss is 16.03, weight deacy loss is 12.04, training accuracy is 0.031250, time 55.579 samples/sec
+epoch 1, total_step 30180, total loss is 29.57 , inference loss is 17.53, weight deacy loss is 12.04, training accuracy is 0.062500, time 56.173 samples/sec
+epoch 1, total_step 30200, total loss is 30.15 , inference loss is 18.11, weight deacy loss is 12.04, training accuracy is 0.062500, time 55.891 samples/sec
+epoch 1, total_step 30220, total loss is 33.29 , inference loss is 21.25, weight deacy loss is 12.03, training accuracy is 0.031250, time 55.913 samples/sec
+epoch 1, total_step 30240, total loss is 31.33 , inference loss is 19.29, weight deacy loss is 12.03, training accuracy is 0.156250, time 56.332 samples/sec
+epoch 1, total_step 30260, total loss is 28.83 , inference loss is 16.80, weight deacy loss is 12.03, training accuracy is 0.031250, time 53.714 samples/sec
+epoch 1, total_step 30280, total loss is 31.11 , inference loss is 19.08, weight deacy loss is 12.03, training accuracy is 0.000000, time 56.606 samples/sec
+epoch 1, total_step 30300, total loss is 31.01 , inference loss is 18.98, weight deacy loss is 12.03, training accuracy is 0.062500, time 56.895 samples/sec
+epoch 1, total_step 30320, total loss is 30.03 , inference loss is 18.01, weight deacy loss is 12.03, training accuracy is 0.031250, time 55.231 samples/sec
+epoch 1, total_step 30340, total loss is 28.14 , inference loss is 16.12, weight deacy loss is 12.03, training accuracy is 0.125000, time 56.402 samples/sec
+epoch 1, total_step 30360, total loss is 29.44 , inference loss is 17.41, weight deacy loss is 12.02, training accuracy is 0.031250, time 56.168 samples/sec
+epoch 1, total_step 30380, total loss is 36.41 , inference loss is 24.39, weight deacy loss is 12.02, training accuracy is 0.000000, time 54.315 samples/sec
+epoch 1, total_step 30400, total loss is 34.45 , inference loss is 22.43, weight deacy loss is 12.02, training accuracy is 0.093750, time 55.235 samples/sec
+epoch 1, total_step 30420, total loss is 32.10 , inference loss is 20.08, weight deacy loss is 12.02, training accuracy is 0.093750, time 53.493 samples/sec
+epoch 1, total_step 30440, total loss is 29.28 , inference loss is 17.26, weight deacy loss is 12.02, training accuracy is 0.062500, time 54.828 samples/sec
+epoch 1, total_step 30460, total loss is 29.19 , inference loss is 17.17, weight deacy loss is 12.02, training accuracy is 0.062500, time 55.940 samples/sec
+epoch 1, total_step 30480, total loss is 29.46 , inference loss is 17.45, weight deacy loss is 12.01, training accuracy is 0.062500, time 53.272 samples/sec
+epoch 1, total_step 30500, total loss is 30.00 , inference loss is 17.99, weight deacy loss is 12.01, training accuracy is 0.125000, time 56.144 samples/sec
+epoch 1, total_step 30520, total loss is 33.56 , inference loss is 21.55, weight deacy loss is 12.01, training accuracy is 0.125000, time 54.292 samples/sec
+epoch 1, total_step 30540, total loss is 28.52 , inference loss is 16.51, weight deacy loss is 12.01, training accuracy is 0.156250, time 56.042 samples/sec
+epoch 1, total_step 30560, total loss is 27.23 , inference loss is 15.22, weight deacy loss is 12.01, training accuracy is 0.125000, time 55.878 samples/sec
+epoch 1, total_step 30580, total loss is 26.40 , inference loss is 14.40, weight deacy loss is 12.01, training accuracy is 0.156250, time 56.764 samples/sec
+epoch 1, total_step 30600, total loss is 31.91 , inference loss is 19.91, weight deacy loss is 12.00, training accuracy is 0.031250, time 55.631 samples/sec
+epoch 1, total_step 30620, total loss is 29.60 , inference loss is 17.60, weight deacy loss is 12.00, training accuracy is 0.093750, time 53.537 samples/sec
+epoch 1, total_step 30640, total loss is 29.37 , inference loss is 17.37, weight deacy loss is 12.00, training accuracy is 0.125000, time 53.865 samples/sec
+epoch 1, total_step 30660, total loss is 31.58 , inference loss is 19.58, weight deacy loss is 12.00, training accuracy is 0.062500, time 56.455 samples/sec
+
+
+
+
+End of epoch 1
+epoch 2, total_step 30680, total loss is 37.25 , inference loss is 25.25, weight deacy loss is 12.00, training accuracy is 0.000000, time 56.477 samples/sec
+epoch 2, total_step 30700, total loss is 27.37 , inference loss is 15.37, weight deacy loss is 12.00, training accuracy is 0.218750, time 56.132 samples/sec
+epoch 2, total_step 30720, total loss is 29.80 , inference loss is 17.80, weight deacy loss is 12.00, training accuracy is 0.125000, time 56.550 samples/sec
+epoch 2, total_step 30740, total loss is 27.24 , inference loss is 15.25, weight deacy loss is 11.99, training accuracy is 0.218750, time 55.957 samples/sec
+epoch 2, total_step 30760, total loss is 26.60 , inference loss is 14.60, weight deacy loss is 11.99, training accuracy is 0.250000, time 56.204 samples/sec
+epoch 2, total_step 30780, total loss is 23.02 , inference loss is 11.03, weight deacy loss is 11.99, training accuracy is 0.312500, time 55.961 samples/sec
+epoch 2, total_step 30800, total loss is 29.46 , inference loss is 17.47, weight deacy loss is 11.99, training accuracy is 0.218750, time 56.391 samples/sec
+epoch 2, total_step 30820, total loss is 25.85 , inference loss is 13.86, weight deacy loss is 11.99, training accuracy is 0.250000, time 55.852 samples/sec
+epoch 2, total_step 30840, total loss is 21.58 , inference loss is 9.59, weight deacy loss is 11.99, training accuracy is 0.218750, time 55.944 samples/sec
+epoch 2, total_step 30860, total loss is 21.95 , inference loss is 9.96, weight deacy loss is 11.98, training accuracy is 0.250000, time 56.361 samples/sec
+epoch 2, total_step 30880, total loss is 25.25 , inference loss is 13.26, weight deacy loss is 11.98, training accuracy is 0.312500, time 55.946 samples/sec
+epoch 2, total_step 30900, total loss is 24.40 , inference loss is 12.42, weight deacy loss is 11.98, training accuracy is 0.218750, time 56.063 samples/sec
+epoch 2, total_step 30920, total loss is 27.77 , inference loss is 15.79, weight deacy loss is 11.98, training accuracy is 0.156250, time 55.402 samples/sec
+epoch 2, total_step 30940, total loss is 24.35 , inference loss is 12.37, weight deacy loss is 11.98, training accuracy is 0.375000, time 53.660 samples/sec
+epoch 2, total_step 30960, total loss is 26.69 , inference loss is 14.72, weight deacy loss is 11.98, training accuracy is 0.156250, time 52.822 samples/sec
+epoch 2, total_step 30980, total loss is 25.95 , inference loss is 13.98, weight deacy loss is 11.97, training accuracy is 0.281250, time 53.530 samples/sec
+epoch 2, total_step 31000, total loss is 28.86 , inference loss is 16.88, weight deacy loss is 11.97, training accuracy is 0.218750, time 56.121 samples/sec
+epoch 2, total_step 31020, total loss is 29.65 , inference loss is 17.68, weight deacy loss is 11.97, training accuracy is 0.312500, time 56.122 samples/sec
+epoch 2, total_step 31040, total loss is 22.94 , inference loss is 10.97, weight deacy loss is 11.97, training accuracy is 0.218750, time 56.059 samples/sec
+epoch 2, total_step 31060, total loss is 25.67 , inference loss is 13.70, weight deacy loss is 11.97, training accuracy is 0.093750, time 55.861 samples/sec
+epoch 2, total_step 31080, total loss is 26.13 , inference loss is 14.17, weight deacy loss is 11.97, training accuracy is 0.250000, time 55.700 samples/sec
+epoch 2, total_step 31100, total loss is 23.65 , inference loss is 11.68, weight deacy loss is 11.96, training accuracy is 0.250000, time 56.282 samples/sec
+epoch 2, total_step 31120, total loss is 29.39 , inference loss is 17.42, weight deacy loss is 11.96, training accuracy is 0.125000, time 56.199 samples/sec
+epoch 2, total_step 31140, total loss is 28.59 , inference loss is 16.63, weight deacy loss is 11.96, training accuracy is 0.125000, time 55.597 samples/sec
+epoch 2, total_step 31160, total loss is 27.95 , inference loss is 15.99, weight deacy loss is 11.96, training accuracy is 0.343750, time 56.960 samples/sec
+epoch 2, total_step 31180, total loss is 25.30 , inference loss is 13.34, weight deacy loss is 11.96, training accuracy is 0.187500, time 55.909 samples/sec
+epoch 2, total_step 31200, total loss is 25.33 , inference loss is 13.38, weight deacy loss is 11.96, training accuracy is 0.343750, time 56.219 samples/sec
+epoch 2, total_step 31220, total loss is 25.58 , inference loss is 13.62, weight deacy loss is 11.95, training accuracy is 0.156250, time 55.746 samples/sec
+epoch 2, total_step 31240, total loss is 23.85 , inference loss is 11.90, weight deacy loss is 11.95, training accuracy is 0.312500, time 54.721 samples/sec
+epoch 2, total_step 31260, total loss is 31.38 , inference loss is 19.43, weight deacy loss is 11.95, training accuracy is 0.187500, time 56.112 samples/sec
+epoch 2, total_step 31280, total loss is 28.58 , inference loss is 16.63, weight deacy loss is 11.95, training accuracy is 0.218750, time 56.002 samples/sec
+epoch 2, total_step 31300, total loss is 25.54 , inference loss is 13.59, weight deacy loss is 11.95, training accuracy is 0.218750, time 57.174 samples/sec
+epoch 2, total_step 31320, total loss is 22.77 , inference loss is 10.83, weight deacy loss is 11.95, training accuracy is 0.250000, time 55.987 samples/sec
+epoch 2, total_step 31340, total loss is 28.29 , inference loss is 16.35, weight deacy loss is 11.94, training accuracy is 0.187500, time 56.337 samples/sec
+epoch 2, total_step 31360, total loss is 26.02 , inference loss is 14.07, weight deacy loss is 11.94, training accuracy is 0.281250, time 55.826 samples/sec
+epoch 2, total_step 31380, total loss is 30.07 , inference loss is 18.13, weight deacy loss is 11.94, training accuracy is 0.250000, time 56.561 samples/sec
+epoch 2, total_step 31400, total loss is 27.58 , inference loss is 15.65, weight deacy loss is 11.94, training accuracy is 0.281250, time 57.505 samples/sec
+epoch 2, total_step 31420, total loss is 26.96 , inference loss is 15.02, weight deacy loss is 11.94, training accuracy is 0.156250, time 56.030 samples/sec
+epoch 2, total_step 31440, total loss is 25.17 , inference loss is 13.24, weight deacy loss is 11.94, training accuracy is 0.312500, time 56.594 samples/sec
+epoch 2, total_step 31460, total loss is 28.42 , inference loss is 16.48, weight deacy loss is 11.93, training accuracy is 0.125000, time 56.748 samples/sec
+epoch 2, total_step 31480, total loss is 26.37 , inference loss is 14.43, weight deacy loss is 11.93, training accuracy is 0.093750, time 56.318 samples/sec
+epoch 2, total_step 31500, total loss is 27.60 , inference loss is 15.67, weight deacy loss is 11.93, training accuracy is 0.187500, time 56.055 samples/sec
+epoch 2, total_step 31520, total loss is 28.07 , inference loss is 16.14, weight deacy loss is 11.93, training accuracy is 0.093750, time 56.327 samples/sec
+epoch 2, total_step 31540, total loss is 29.64 , inference loss is 17.71, weight deacy loss is 11.93, training accuracy is 0.093750, time 56.346 samples/sec
+epoch 2, total_step 31560, total loss is 24.37 , inference loss is 12.44, weight deacy loss is 11.93, training accuracy is 0.218750, time 56.569 samples/sec
+epoch 2, total_step 31580, total loss is 27.58 , inference loss is 15.66, weight deacy loss is 11.93, training accuracy is 0.093750, time 56.261 samples/sec
+epoch 2, total_step 31600, total loss is 24.68 , inference loss is 12.76, weight deacy loss is 11.92, training accuracy is 0.187500, time 54.212 samples/sec
+epoch 2, total_step 31620, total loss is 25.28 , inference loss is 13.36, weight deacy loss is 11.92, training accuracy is 0.250000, time 55.743 samples/sec
+epoch 2, total_step 31640, total loss is 25.89 , inference loss is 13.97, weight deacy loss is 11.92, training accuracy is 0.281250, time 56.076 samples/sec
+epoch 2, total_step 31660, total loss is 28.33 , inference loss is 16.41, weight deacy loss is 11.92, training accuracy is 0.187500, time 56.195 samples/sec
+epoch 2, total_step 31680, total loss is 24.62 , inference loss is 12.70, weight deacy loss is 11.92, training accuracy is 0.093750, time 55.939 samples/sec
+epoch 2, total_step 31700, total loss is 26.43 , inference loss is 14.52, weight deacy loss is 11.92, training accuracy is 0.187500, time 56.081 samples/sec
+epoch 2, total_step 31720, total loss is 24.84 , inference loss is 12.92, weight deacy loss is 11.91, training accuracy is 0.156250, time 56.516 samples/sec
+epoch 2, total_step 31740, total loss is 30.32 , inference loss is 18.41, weight deacy loss is 11.91, training accuracy is 0.218750, time 55.914 samples/sec
+epoch 2, total_step 31760, total loss is 26.53 , inference loss is 14.62, weight deacy loss is 11.91, training accuracy is 0.218750, time 53.451 samples/sec
+epoch 2, total_step 31780, total loss is 27.89 , inference loss is 15.98, weight deacy loss is 11.91, training accuracy is 0.187500, time 54.755 samples/sec
+epoch 2, total_step 31800, total loss is 24.77 , inference loss is 12.86, weight deacy loss is 11.91, training accuracy is 0.156250, time 55.692 samples/sec
+epoch 2, total_step 31820, total loss is 25.89 , inference loss is 13.99, weight deacy loss is 11.91, training accuracy is 0.187500, time 55.454 samples/sec
+epoch 2, total_step 31840, total loss is 27.36 , inference loss is 15.45, weight deacy loss is 11.90, training accuracy is 0.156250, time 54.256 samples/sec
+epoch 2, total_step 31860, total loss is 27.75 , inference loss is 15.85, weight deacy loss is 11.90, training accuracy is 0.156250, time 55.705 samples/sec
+epoch 2, total_step 31880, total loss is 27.48 , inference loss is 15.58, weight deacy loss is 11.90, training accuracy is 0.156250, time 55.607 samples/sec
+epoch 2, total_step 31900, total loss is 30.15 , inference loss is 18.25, weight deacy loss is 11.90, training accuracy is 0.093750, time 56.274 samples/sec
+epoch 2, total_step 31920, total loss is 26.98 , inference loss is 15.08, weight deacy loss is 11.90, training accuracy is 0.062500, time 56.474 samples/sec
+epoch 2, total_step 31940, total loss is 30.36 , inference loss is 18.47, weight deacy loss is 11.90, training accuracy is 0.125000, time 56.020 samples/sec
+epoch 2, total_step 31960, total loss is 28.46 , inference loss is 16.56, weight deacy loss is 11.89, training accuracy is 0.187500, time 56.851 samples/sec
+epoch 2, total_step 31980, total loss is 28.80 , inference loss is 16.91, weight deacy loss is 11.89, training accuracy is 0.125000, time 56.948 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.54464900000005
+best_threshold_index 112 0.9783333333333334
+best_threshold_index 112 0.9790740740740741
+best_threshold_index 112 0.9787037037037037
+best_threshold_index 112 0.9792592592592593
+best_threshold_index 112 0.9792592592592593
+best_threshold_index 112 0.9781481481481481
+best_threshold_index 112 0.9783333333333334
+best_threshold_index 112 0.9777777777777777
+best_threshold_index 112 0.9766666666666667
+best_threshold_index 112 0.9777777777777777
+[lfw][32000]XNorm: 24.832845
+[lfw][32000]Accuracy-Flip: 0.97833+-0.00687
+testing verification..
+(14000, 512)
+infer time 86.34647600000008
+best_threshold_index 111 0.9714285714285714
+best_threshold_index 111 0.9703174603174604
+best_threshold_index 111 0.9714285714285714
+best_threshold_index 111 0.9704761904761905
+best_threshold_index 111 0.9693650793650793
+best_threshold_index 111 0.9687301587301588
+best_threshold_index 111 0.9709523809523809
+best_threshold_index 111 0.9696825396825397
+best_threshold_index 111 0.9688888888888889
+best_threshold_index 111 0.9715873015873016
+[cfp_ff][32000]XNorm: 24.546969
+[cfp_ff][32000]Accuracy-Flip: 0.97029+-0.00917
+testing verification..
+(14000, 512)
+infer time 86.80295300000009
+best_threshold_index 131 0.8587301587301587
+best_threshold_index 131 0.8563492063492063
+best_threshold_index 129 0.8541269841269842
+best_threshold_index 129 0.8588888888888889
+best_threshold_index 131 0.86
+best_threshold_index 131 0.8522222222222222
+best_threshold_index 131 0.8536507936507937
+best_threshold_index 131 0.8580952380952381
+best_threshold_index 131 0.8563492063492063
+best_threshold_index 131 0.8574603174603175
+[cfp_fp][32000]XNorm: 23.947544
+[cfp_fp][32000]Accuracy-Flip: 0.85557+-0.02179
+testing verification..
+(12000, 512)
+infer time 74.54066499999995
+best_threshold_index 131 0.8448148148148148
+best_threshold_index 130 0.8444444444444444
+best_threshold_index 136 0.8475925925925926
+best_threshold_index 131 0.8464814814814815
+best_threshold_index 130 0.8516666666666667
+best_threshold_index 131 0.8457407407407408
+best_threshold_index 130 0.8511111111111112
+best_threshold_index 131 0.8537037037037037
+best_threshold_index 130 0.85
+best_threshold_index 130 0.8507407407407407
+[agedb_30][32000]XNorm: 23.959389
+[agedb_30][32000]Accuracy-Flip: 0.84467+-0.02787
+test accuracy is: 0.9783333333333333
+epoch 2, total_step 32000, total loss is 31.28 , inference loss is 19.39, weight deacy loss is 11.89, training accuracy is 0.125000, time 52.947 samples/sec
+epoch 2, total_step 32020, total loss is 29.69 , inference loss is 17.80, weight deacy loss is 11.89, training accuracy is 0.062500, time 56.255 samples/sec
+epoch 2, total_step 32040, total loss is 32.58 , inference loss is 20.69, weight deacy loss is 11.89, training accuracy is 0.125000, time 56.454 samples/sec
+epoch 2, total_step 32060, total loss is 28.44 , inference loss is 16.55, weight deacy loss is 11.89, training accuracy is 0.187500, time 56.308 samples/sec
+epoch 2, total_step 32080, total loss is 32.25 , inference loss is 20.36, weight deacy loss is 11.88, training accuracy is 0.187500, time 56.429 samples/sec
+epoch 2, total_step 32100, total loss is 30.70 , inference loss is 18.82, weight deacy loss is 11.88, training accuracy is 0.187500, time 56.653 samples/sec
+epoch 2, total_step 32120, total loss is 28.15 , inference loss is 16.27, weight deacy loss is 11.88, training accuracy is 0.125000, time 56.205 samples/sec
+epoch 2, total_step 32140, total loss is 29.47 , inference loss is 17.59, weight deacy loss is 11.88, training accuracy is 0.187500, time 55.282 samples/sec
+epoch 2, total_step 32160, total loss is 27.69 , inference loss is 15.81, weight deacy loss is 11.88, training accuracy is 0.156250, time 55.600 samples/sec
+epoch 2, total_step 32180, total loss is 29.43 , inference loss is 17.55, weight deacy loss is 11.88, training accuracy is 0.062500, time 55.783 samples/sec
+epoch 2, total_step 32200, total loss is 27.87 , inference loss is 15.99, weight deacy loss is 11.88, training accuracy is 0.187500, time 56.040 samples/sec
+epoch 2, total_step 32220, total loss is 28.78 , inference loss is 16.90, weight deacy loss is 11.87, training accuracy is 0.125000, time 55.504 samples/sec
+epoch 2, total_step 32240, total loss is 31.57 , inference loss is 19.70, weight deacy loss is 11.87, training accuracy is 0.250000, time 55.059 samples/sec
+epoch 2, total_step 32260, total loss is 27.18 , inference loss is 15.31, weight deacy loss is 11.87, training accuracy is 0.125000, time 55.900 samples/sec
+epoch 2, total_step 32280, total loss is 29.84 , inference loss is 17.97, weight deacy loss is 11.87, training accuracy is 0.156250, time 55.601 samples/sec
+epoch 2, total_step 32300, total loss is 25.53 , inference loss is 13.66, weight deacy loss is 11.87, training accuracy is 0.187500, time 55.565 samples/sec
+epoch 2, total_step 32320, total loss is 33.53 , inference loss is 21.67, weight deacy loss is 11.87, training accuracy is 0.218750, time 56.439 samples/sec
+epoch 2, total_step 32340, total loss is 30.90 , inference loss is 19.04, weight deacy loss is 11.86, training accuracy is 0.156250, time 55.853 samples/sec
+epoch 2, total_step 32360, total loss is 28.30 , inference loss is 16.44, weight deacy loss is 11.86, training accuracy is 0.218750, time 55.963 samples/sec
+epoch 2, total_step 32380, total loss is 26.12 , inference loss is 14.26, weight deacy loss is 11.86, training accuracy is 0.218750, time 56.283 samples/sec
+epoch 2, total_step 32400, total loss is 25.52 , inference loss is 13.66, weight deacy loss is 11.86, training accuracy is 0.281250, time 56.063 samples/sec
+epoch 2, total_step 32420, total loss is 29.30 , inference loss is 17.45, weight deacy loss is 11.86, training accuracy is 0.156250, time 56.480 samples/sec
+epoch 2, total_step 32440, total loss is 25.09 , inference loss is 13.23, weight deacy loss is 11.86, training accuracy is 0.218750, time 56.302 samples/sec
+epoch 2, total_step 32460, total loss is 24.91 , inference loss is 13.06, weight deacy loss is 11.85, training accuracy is 0.250000, time 55.591 samples/sec
+epoch 2, total_step 32480, total loss is 28.70 , inference loss is 16.84, weight deacy loss is 11.85, training accuracy is 0.187500, time 55.891 samples/sec
+epoch 2, total_step 32500, total loss is 23.84 , inference loss is 11.99, weight deacy loss is 11.85, training accuracy is 0.312500, time 55.538 samples/sec
+epoch 2, total_step 32520, total loss is 26.97 , inference loss is 15.12, weight deacy loss is 11.85, training accuracy is 0.156250, time 56.015 samples/sec
+epoch 2, total_step 32540, total loss is 26.02 , inference loss is 14.18, weight deacy loss is 11.85, training accuracy is 0.218750, time 56.459 samples/sec
+epoch 2, total_step 32560, total loss is 27.79 , inference loss is 15.94, weight deacy loss is 11.85, training accuracy is 0.156250, time 55.830 samples/sec
+epoch 2, total_step 32580, total loss is 24.35 , inference loss is 12.50, weight deacy loss is 11.84, training accuracy is 0.187500, time 55.769 samples/sec
+epoch 2, total_step 32600, total loss is 25.31 , inference loss is 13.47, weight deacy loss is 11.84, training accuracy is 0.250000, time 52.827 samples/sec
+epoch 2, total_step 32620, total loss is 28.56 , inference loss is 16.72, weight deacy loss is 11.84, training accuracy is 0.125000, time 56.855 samples/sec
+epoch 2, total_step 32640, total loss is 24.73 , inference loss is 12.89, weight deacy loss is 11.84, training accuracy is 0.156250, time 56.378 samples/sec
+epoch 2, total_step 32660, total loss is 27.50 , inference loss is 15.67, weight deacy loss is 11.84, training accuracy is 0.187500, time 56.534 samples/sec
+epoch 2, total_step 32680, total loss is 18.46 , inference loss is 6.63, weight deacy loss is 11.84, training accuracy is 0.343750, time 56.006 samples/sec
+epoch 2, total_step 32700, total loss is 27.13 , inference loss is 15.30, weight deacy loss is 11.83, training accuracy is 0.312500, time 53.649 samples/sec
+epoch 2, total_step 32720, total loss is 29.83 , inference loss is 18.00, weight deacy loss is 11.83, training accuracy is 0.187500, time 56.511 samples/sec
+epoch 2, total_step 32740, total loss is 24.07 , inference loss is 12.24, weight deacy loss is 11.83, training accuracy is 0.312500, time 55.717 samples/sec
+epoch 2, total_step 32760, total loss is 25.22 , inference loss is 13.39, weight deacy loss is 11.83, training accuracy is 0.187500, time 55.649 samples/sec
+epoch 2, total_step 32780, total loss is 26.02 , inference loss is 14.19, weight deacy loss is 11.83, training accuracy is 0.187500, time 56.161 samples/sec
+epoch 2, total_step 32800, total loss is 25.61 , inference loss is 13.78, weight deacy loss is 11.83, training accuracy is 0.187500, time 56.356 samples/sec
+epoch 2, total_step 32820, total loss is 27.66 , inference loss is 15.84, weight deacy loss is 11.82, training accuracy is 0.125000, time 56.054 samples/sec
+epoch 2, total_step 32840, total loss is 25.38 , inference loss is 13.56, weight deacy loss is 11.82, training accuracy is 0.250000, time 56.204 samples/sec
+epoch 2, total_step 32860, total loss is 31.09 , inference loss is 19.27, weight deacy loss is 11.82, training accuracy is 0.156250, time 56.452 samples/sec
+epoch 2, total_step 32880, total loss is 22.76 , inference loss is 10.94, weight deacy loss is 11.82, training accuracy is 0.375000, time 55.994 samples/sec
+epoch 2, total_step 32900, total loss is 22.83 , inference loss is 11.01, weight deacy loss is 11.82, training accuracy is 0.218750, time 56.161 samples/sec
+epoch 2, total_step 32920, total loss is 26.52 , inference loss is 14.70, weight deacy loss is 11.82, training accuracy is 0.281250, time 56.531 samples/sec
+epoch 2, total_step 32940, total loss is 23.25 , inference loss is 11.44, weight deacy loss is 11.81, training accuracy is 0.218750, time 52.690 samples/sec
+epoch 2, total_step 32960, total loss is 26.35 , inference loss is 14.53, weight deacy loss is 11.81, training accuracy is 0.218750, time 57.074 samples/sec
+epoch 2, total_step 32980, total loss is 27.01 , inference loss is 15.19, weight deacy loss is 11.81, training accuracy is 0.187500, time 54.849 samples/sec
+epoch 2, total_step 33000, total loss is 24.48 , inference loss is 12.67, weight deacy loss is 11.81, training accuracy is 0.218750, time 55.888 samples/sec
+epoch 2, total_step 33020, total loss is 29.94 , inference loss is 18.13, weight deacy loss is 11.81, training accuracy is 0.187500, time 55.641 samples/sec
+epoch 2, total_step 33040, total loss is 21.75 , inference loss is 9.94, weight deacy loss is 11.81, training accuracy is 0.312500, time 55.785 samples/sec
+epoch 2, total_step 33060, total loss is 25.87 , inference loss is 14.06, weight deacy loss is 11.80, training accuracy is 0.125000, time 56.397 samples/sec
+epoch 2, total_step 33080, total loss is 26.25 , inference loss is 14.45, weight deacy loss is 11.80, training accuracy is 0.187500, time 56.348 samples/sec
+epoch 2, total_step 33100, total loss is 25.07 , inference loss is 13.27, weight deacy loss is 11.80, training accuracy is 0.156250, time 56.421 samples/sec
+epoch 2, total_step 33120, total loss is 25.57 , inference loss is 13.77, weight deacy loss is 11.80, training accuracy is 0.218750, time 56.728 samples/sec
+epoch 2, total_step 33140, total loss is 27.13 , inference loss is 15.33, weight deacy loss is 11.80, training accuracy is 0.125000, time 55.566 samples/sec
+epoch 2, total_step 33160, total loss is 20.76 , inference loss is 8.96, weight deacy loss is 11.80, training accuracy is 0.281250, time 56.067 samples/sec
+epoch 2, total_step 33180, total loss is 23.60 , inference loss is 11.80, weight deacy loss is 11.79, training accuracy is 0.281250, time 56.221 samples/sec
+epoch 2, total_step 33200, total loss is 22.81 , inference loss is 11.02, weight deacy loss is 11.79, training accuracy is 0.093750, time 55.518 samples/sec
+epoch 2, total_step 33220, total loss is 27.02 , inference loss is 15.22, weight deacy loss is 11.79, training accuracy is 0.187500, time 54.284 samples/sec
+epoch 2, total_step 33240, total loss is 19.63 , inference loss is 7.84, weight deacy loss is 11.79, training accuracy is 0.343750, time 55.542 samples/sec
+epoch 2, total_step 33260, total loss is 21.22 , inference loss is 9.43, weight deacy loss is 11.79, training accuracy is 0.281250, time 53.470 samples/sec
+epoch 2, total_step 33280, total loss is 22.58 , inference loss is 10.79, weight deacy loss is 11.79, training accuracy is 0.375000, time 53.977 samples/sec
+epoch 2, total_step 33300, total loss is 24.67 , inference loss is 12.89, weight deacy loss is 11.78, training accuracy is 0.187500, time 56.441 samples/sec
+epoch 2, total_step 33320, total loss is 25.96 , inference loss is 14.18, weight deacy loss is 11.78, training accuracy is 0.187500, time 56.214 samples/sec
+epoch 2, total_step 33340, total loss is 25.59 , inference loss is 13.81, weight deacy loss is 11.78, training accuracy is 0.281250, time 55.353 samples/sec
+epoch 2, total_step 33360, total loss is 27.28 , inference loss is 15.50, weight deacy loss is 11.78, training accuracy is 0.187500, time 56.154 samples/sec
+epoch 2, total_step 33380, total loss is 25.38 , inference loss is 13.60, weight deacy loss is 11.78, training accuracy is 0.250000, time 56.012 samples/sec
+epoch 2, total_step 33400, total loss is 23.78 , inference loss is 12.00, weight deacy loss is 11.78, training accuracy is 0.187500, time 56.437 samples/sec
+epoch 2, total_step 33420, total loss is 23.48 , inference loss is 11.71, weight deacy loss is 11.78, training accuracy is 0.187500, time 54.890 samples/sec
+epoch 2, total_step 33440, total loss is 24.56 , inference loss is 12.79, weight deacy loss is 11.77, training accuracy is 0.156250, time 55.130 samples/sec
+epoch 2, total_step 33460, total loss is 24.76 , inference loss is 12.99, weight deacy loss is 11.77, training accuracy is 0.125000, time 56.451 samples/sec
+epoch 2, total_step 33480, total loss is 31.78 , inference loss is 20.01, weight deacy loss is 11.77, training accuracy is 0.093750, time 55.779 samples/sec
+epoch 2, total_step 33500, total loss is 25.04 , inference loss is 13.27, weight deacy loss is 11.77, training accuracy is 0.187500, time 56.393 samples/sec
+epoch 2, total_step 33520, total loss is 19.91 , inference loss is 8.15, weight deacy loss is 11.77, training accuracy is 0.281250, time 56.268 samples/sec
+epoch 2, total_step 33540, total loss is 22.73 , inference loss is 10.97, weight deacy loss is 11.77, training accuracy is 0.125000, time 55.927 samples/sec
+epoch 2, total_step 33560, total loss is 22.53 , inference loss is 10.77, weight deacy loss is 11.76, training accuracy is 0.281250, time 55.711 samples/sec
+epoch 2, total_step 33580, total loss is 24.60 , inference loss is 12.84, weight deacy loss is 11.76, training accuracy is 0.250000, time 55.557 samples/sec
+epoch 2, total_step 33600, total loss is 25.01 , inference loss is 13.25, weight deacy loss is 11.76, training accuracy is 0.250000, time 56.537 samples/sec
+epoch 2, total_step 33620, total loss is 22.45 , inference loss is 10.69, weight deacy loss is 11.76, training accuracy is 0.281250, time 56.577 samples/sec
+epoch 2, total_step 33640, total loss is 26.79 , inference loss is 15.04, weight deacy loss is 11.76, training accuracy is 0.250000, time 55.880 samples/sec
+epoch 2, total_step 33660, total loss is 26.06 , inference loss is 14.30, weight deacy loss is 11.76, training accuracy is 0.281250, time 56.655 samples/sec
+epoch 2, total_step 33680, total loss is 26.03 , inference loss is 14.28, weight deacy loss is 11.75, training accuracy is 0.125000, time 56.856 samples/sec
+epoch 2, total_step 33700, total loss is 23.00 , inference loss is 11.25, weight deacy loss is 11.75, training accuracy is 0.281250, time 56.387 samples/sec
+epoch 2, total_step 33720, total loss is 24.82 , inference loss is 13.06, weight deacy loss is 11.75, training accuracy is 0.218750, time 55.659 samples/sec
+epoch 2, total_step 33740, total loss is 22.51 , inference loss is 10.76, weight deacy loss is 11.75, training accuracy is 0.281250, time 57.763 samples/sec
+epoch 2, total_step 33760, total loss is 30.34 , inference loss is 18.59, weight deacy loss is 11.75, training accuracy is 0.093750, time 55.554 samples/sec
+epoch 2, total_step 33780, total loss is 29.56 , inference loss is 17.81, weight deacy loss is 11.75, training accuracy is 0.250000, time 57.034 samples/sec
+epoch 2, total_step 33800, total loss is 24.56 , inference loss is 12.82, weight deacy loss is 11.74, training accuracy is 0.281250, time 56.386 samples/sec
+epoch 2, total_step 33820, total loss is 25.17 , inference loss is 13.43, weight deacy loss is 11.74, training accuracy is 0.187500, time 55.565 samples/sec
+epoch 2, total_step 33840, total loss is 26.11 , inference loss is 14.37, weight deacy loss is 11.74, training accuracy is 0.187500, time 56.090 samples/sec
+epoch 2, total_step 33860, total loss is 26.73 , inference loss is 14.99, weight deacy loss is 11.74, training accuracy is 0.218750, time 56.108 samples/sec
+epoch 2, total_step 33880, total loss is 26.22 , inference loss is 14.48, weight deacy loss is 11.74, training accuracy is 0.281250, time 55.922 samples/sec
+epoch 2, total_step 33900, total loss is 21.95 , inference loss is 10.21, weight deacy loss is 11.74, training accuracy is 0.156250, time 56.509 samples/sec
+epoch 2, total_step 33920, total loss is 29.49 , inference loss is 17.76, weight deacy loss is 11.73, training accuracy is 0.218750, time 56.077 samples/sec
+epoch 2, total_step 33940, total loss is 26.18 , inference loss is 14.44, weight deacy loss is 11.73, training accuracy is 0.312500, time 57.184 samples/sec
+epoch 2, total_step 33960, total loss is 27.79 , inference loss is 16.06, weight deacy loss is 11.73, training accuracy is 0.250000, time 57.066 samples/sec
+epoch 2, total_step 33980, total loss is 28.05 , inference loss is 16.32, weight deacy loss is 11.73, training accuracy is 0.125000, time 56.343 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.90797800000004
+best_threshold_index 112 0.9788888888888889
+best_threshold_index 112 0.9792592592592593
+best_threshold_index 117 0.9796296296296296
+best_threshold_index 112 0.9788888888888889
+best_threshold_index 117 0.9798148148148148
+best_threshold_index 117 0.9783333333333334
+best_threshold_index 117 0.9794444444444445
+best_threshold_index 112 0.9792592592592593
+best_threshold_index 112 0.9781481481481481
+best_threshold_index 112 0.9794444444444445
+[lfw][34000]XNorm: 23.220113
+[lfw][34000]Accuracy-Flip: 0.97733+-0.00496
+testing verification..
+(14000, 512)
+infer time 87.05466000000001
+best_threshold_index 112 0.9726984126984127
+best_threshold_index 112 0.9714285714285714
+best_threshold_index 112 0.9722222222222222
+best_threshold_index 112 0.9715873015873016
+best_threshold_index 115 0.9722222222222222
+best_threshold_index 115 0.9722222222222222
+best_threshold_index 115 0.9728571428571429
+best_threshold_index 116 0.9720634920634921
+best_threshold_index 115 0.9726984126984127
+best_threshold_index 112 0.9731746031746031
+[cfp_ff][34000]XNorm: 22.451691
+[cfp_ff][34000]Accuracy-Flip: 0.97114+-0.00455
+testing verification..
+(14000, 512)
+infer time 86.97091999999996
+best_threshold_index 135 0.8758730158730159
+best_threshold_index 134 0.8757142857142857
+best_threshold_index 135 0.8752380952380953
+best_threshold_index 134 0.8771428571428571
+best_threshold_index 135 0.8777777777777778
+best_threshold_index 130 0.8720634920634921
+best_threshold_index 134 0.8744444444444445
+best_threshold_index 134 0.8752380952380953
+best_threshold_index 135 0.8750793650793651
+best_threshold_index 137 0.876031746031746
+[cfp_fp][34000]XNorm: 21.628063
+[cfp_fp][34000]Accuracy-Flip: 0.87200+-0.01077
+testing verification..
+(12000, 512)
+infer time 74.34086100000002
+best_threshold_index 139 0.8290740740740741
+best_threshold_index 139 0.8296296296296296
+best_threshold_index 139 0.8327777777777777
+best_threshold_index 139 0.832962962962963
+best_threshold_index 139 0.835
+best_threshold_index 139 0.8303703703703704
+best_threshold_index 139 0.8338888888888889
+best_threshold_index 139 0.8366666666666667
+best_threshold_index 139 0.8351851851851851
+best_threshold_index 139 0.8344444444444444
+[agedb_30][34000]XNorm: 22.507396
+[agedb_30][34000]Accuracy-Flip: 0.83300+-0.02183
+test accuracy is: 0.9773333333333334
+epoch 2, total_step 34000, total loss is 30.95 , inference loss is 19.22, weight deacy loss is 11.73, training accuracy is 0.187500, time 53.623 samples/sec
+epoch 2, total_step 34020, total loss is 26.23 , inference loss is 14.50, weight deacy loss is 11.73, training accuracy is 0.250000, time 55.650 samples/sec
+epoch 2, total_step 34040, total loss is 26.96 , inference loss is 15.23, weight deacy loss is 11.73, training accuracy is 0.125000, time 56.471 samples/sec
+epoch 2, total_step 34060, total loss is 29.84 , inference loss is 18.11, weight deacy loss is 11.72, training accuracy is 0.218750, time 56.124 samples/sec
+epoch 2, total_step 34080, total loss is 30.16 , inference loss is 18.44, weight deacy loss is 11.72, training accuracy is 0.062500, time 56.373 samples/sec
+epoch 2, total_step 34100, total loss is 25.98 , inference loss is 14.25, weight deacy loss is 11.72, training accuracy is 0.281250, time 54.864 samples/sec
+epoch 2, total_step 34120, total loss is 22.70 , inference loss is 10.99, weight deacy loss is 11.72, training accuracy is 0.281250, time 56.015 samples/sec
+epoch 2, total_step 34140, total loss is 27.27 , inference loss is 15.55, weight deacy loss is 11.72, training accuracy is 0.156250, time 52.430 samples/sec
+epoch 2, total_step 34160, total loss is 30.35 , inference loss is 18.63, weight deacy loss is 11.72, training accuracy is 0.156250, time 56.951 samples/sec
+epoch 2, total_step 34180, total loss is 28.89 , inference loss is 17.17, weight deacy loss is 11.71, training accuracy is 0.250000, time 56.444 samples/sec
+epoch 2, total_step 34200, total loss is 28.58 , inference loss is 16.86, weight deacy loss is 11.71, training accuracy is 0.062500, time 56.525 samples/sec
+epoch 2, total_step 34220, total loss is 26.03 , inference loss is 14.32, weight deacy loss is 11.71, training accuracy is 0.156250, time 56.793 samples/sec
+epoch 2, total_step 34240, total loss is 25.32 , inference loss is 13.61, weight deacy loss is 11.71, training accuracy is 0.156250, time 53.179 samples/sec
+epoch 2, total_step 34260, total loss is 25.35 , inference loss is 13.64, weight deacy loss is 11.71, training accuracy is 0.312500, time 55.378 samples/sec
+epoch 2, total_step 34280, total loss is 25.85 , inference loss is 14.14, weight deacy loss is 11.71, training accuracy is 0.250000, time 53.974 samples/sec
+epoch 2, total_step 34300, total loss is 24.45 , inference loss is 12.75, weight deacy loss is 11.71, training accuracy is 0.250000, time 53.261 samples/sec
+epoch 2, total_step 34320, total loss is 27.97 , inference loss is 16.27, weight deacy loss is 11.70, training accuracy is 0.093750, time 56.359 samples/sec
+epoch 2, total_step 34340, total loss is 29.03 , inference loss is 17.33, weight deacy loss is 11.70, training accuracy is 0.125000, time 56.412 samples/sec
+epoch 2, total_step 34360, total loss is 29.80 , inference loss is 18.10, weight deacy loss is 11.70, training accuracy is 0.093750, time 56.437 samples/sec
+epoch 2, total_step 34380, total loss is 25.48 , inference loss is 13.78, weight deacy loss is 11.70, training accuracy is 0.156250, time 54.444 samples/sec
+epoch 2, total_step 34400, total loss is 26.51 , inference loss is 14.81, weight deacy loss is 11.70, training accuracy is 0.125000, time 55.881 samples/sec
+epoch 2, total_step 34420, total loss is 30.14 , inference loss is 18.45, weight deacy loss is 11.70, training accuracy is 0.031250, time 56.538 samples/sec
+epoch 2, total_step 34440, total loss is 29.94 , inference loss is 18.25, weight deacy loss is 11.69, training accuracy is 0.125000, time 56.507 samples/sec
+epoch 2, total_step 34460, total loss is 26.35 , inference loss is 14.66, weight deacy loss is 11.69, training accuracy is 0.156250, time 55.773 samples/sec
+epoch 2, total_step 34480, total loss is 25.52 , inference loss is 13.82, weight deacy loss is 11.69, training accuracy is 0.125000, time 56.297 samples/sec
+epoch 2, total_step 34500, total loss is 27.55 , inference loss is 15.86, weight deacy loss is 11.69, training accuracy is 0.187500, time 56.430 samples/sec
+epoch 2, total_step 34520, total loss is 27.75 , inference loss is 16.06, weight deacy loss is 11.69, training accuracy is 0.093750, time 55.906 samples/sec
+epoch 2, total_step 34540, total loss is 28.95 , inference loss is 17.26, weight deacy loss is 11.69, training accuracy is 0.062500, time 54.602 samples/sec
+epoch 2, total_step 34560, total loss is 27.71 , inference loss is 16.03, weight deacy loss is 11.68, training accuracy is 0.156250, time 55.605 samples/sec
+epoch 2, total_step 34580, total loss is 30.69 , inference loss is 19.01, weight deacy loss is 11.68, training accuracy is 0.031250, time 56.739 samples/sec
+epoch 2, total_step 34600, total loss is 30.07 , inference loss is 18.38, weight deacy loss is 11.68, training accuracy is 0.125000, time 56.307 samples/sec
+epoch 2, total_step 34620, total loss is 25.10 , inference loss is 13.42, weight deacy loss is 11.68, training accuracy is 0.093750, time 55.820 samples/sec
+epoch 2, total_step 34640, total loss is 27.77 , inference loss is 16.09, weight deacy loss is 11.68, training accuracy is 0.250000, time 56.407 samples/sec
+epoch 2, total_step 34660, total loss is 29.39 , inference loss is 17.72, weight deacy loss is 11.68, training accuracy is 0.031250, time 55.436 samples/sec
+epoch 2, total_step 34680, total loss is 29.10 , inference loss is 17.42, weight deacy loss is 11.68, training accuracy is 0.156250, time 56.408 samples/sec
+epoch 2, total_step 34700, total loss is 28.22 , inference loss is 16.55, weight deacy loss is 11.67, training accuracy is 0.125000, time 56.209 samples/sec
+epoch 2, total_step 34720, total loss is 28.21 , inference loss is 16.54, weight deacy loss is 11.67, training accuracy is 0.156250, time 57.701 samples/sec
+epoch 2, total_step 34740, total loss is 28.88 , inference loss is 17.21, weight deacy loss is 11.67, training accuracy is 0.156250, time 55.749 samples/sec
+epoch 2, total_step 34760, total loss is 26.94 , inference loss is 15.27, weight deacy loss is 11.67, training accuracy is 0.125000, time 55.606 samples/sec
+epoch 2, total_step 34780, total loss is 28.57 , inference loss is 16.90, weight deacy loss is 11.67, training accuracy is 0.125000, time 56.555 samples/sec
+epoch 2, total_step 34800, total loss is 29.09 , inference loss is 17.43, weight deacy loss is 11.67, training accuracy is 0.218750, time 55.689 samples/sec
+epoch 2, total_step 34820, total loss is 29.01 , inference loss is 17.35, weight deacy loss is 11.66, training accuracy is 0.125000, time 56.196 samples/sec
+epoch 2, total_step 34840, total loss is 29.07 , inference loss is 17.41, weight deacy loss is 11.66, training accuracy is 0.187500, time 56.293 samples/sec
+epoch 2, total_step 34860, total loss is 24.77 , inference loss is 13.11, weight deacy loss is 11.66, training accuracy is 0.343750, time 55.947 samples/sec
+epoch 2, total_step 34880, total loss is 27.22 , inference loss is 15.56, weight deacy loss is 11.66, training accuracy is 0.093750, time 55.433 samples/sec
+epoch 2, total_step 34900, total loss is 25.29 , inference loss is 13.64, weight deacy loss is 11.66, training accuracy is 0.093750, time 55.561 samples/sec
+epoch 2, total_step 34920, total loss is 27.62 , inference loss is 15.96, weight deacy loss is 11.66, training accuracy is 0.218750, time 56.033 samples/sec
+epoch 2, total_step 34940, total loss is 25.06 , inference loss is 13.40, weight deacy loss is 11.66, training accuracy is 0.187500, time 56.316 samples/sec
+epoch 2, total_step 34960, total loss is 25.50 , inference loss is 13.85, weight deacy loss is 11.65, training accuracy is 0.187500, time 56.894 samples/sec
+epoch 2, total_step 34980, total loss is 24.07 , inference loss is 12.41, weight deacy loss is 11.65, training accuracy is 0.218750, time 55.922 samples/sec
+epoch 2, total_step 35000, total loss is 27.01 , inference loss is 15.36, weight deacy loss is 11.65, training accuracy is 0.156250, time 56.497 samples/sec
+epoch 2, total_step 35020, total loss is 26.10 , inference loss is 14.45, weight deacy loss is 11.65, training accuracy is 0.156250, time 56.297 samples/sec
+epoch 2, total_step 35040, total loss is 30.08 , inference loss is 18.43, weight deacy loss is 11.65, training accuracy is 0.093750, time 56.290 samples/sec
+epoch 2, total_step 35060, total loss is 29.91 , inference loss is 18.26, weight deacy loss is 11.65, training accuracy is 0.187500, time 56.289 samples/sec
+epoch 2, total_step 35080, total loss is 25.54 , inference loss is 13.90, weight deacy loss is 11.64, training accuracy is 0.093750, time 56.377 samples/sec
+epoch 2, total_step 35100, total loss is 26.40 , inference loss is 14.75, weight deacy loss is 11.64, training accuracy is 0.156250, time 56.186 samples/sec
+epoch 2, total_step 35120, total loss is 29.88 , inference loss is 18.24, weight deacy loss is 11.64, training accuracy is 0.156250, time 55.561 samples/sec
+epoch 2, total_step 35140, total loss is 27.93 , inference loss is 16.29, weight deacy loss is 11.64, training accuracy is 0.156250, time 56.222 samples/sec
+epoch 2, total_step 35160, total loss is 28.11 , inference loss is 16.47, weight deacy loss is 11.64, training accuracy is 0.093750, time 54.876 samples/sec
+epoch 2, total_step 35180, total loss is 25.86 , inference loss is 14.22, weight deacy loss is 11.64, training accuracy is 0.156250, time 56.091 samples/sec
+epoch 2, total_step 35200, total loss is 25.56 , inference loss is 13.93, weight deacy loss is 11.63, training accuracy is 0.218750, time 54.807 samples/sec
+epoch 2, total_step 35220, total loss is 27.38 , inference loss is 15.75, weight deacy loss is 11.63, training accuracy is 0.218750, time 55.372 samples/sec
+epoch 2, total_step 35240, total loss is 25.99 , inference loss is 14.36, weight deacy loss is 11.63, training accuracy is 0.156250, time 55.056 samples/sec
+epoch 2, total_step 35260, total loss is 32.46 , inference loss is 20.83, weight deacy loss is 11.63, training accuracy is 0.031250, time 56.430 samples/sec
+epoch 2, total_step 35280, total loss is 26.09 , inference loss is 14.46, weight deacy loss is 11.63, training accuracy is 0.125000, time 57.046 samples/sec
+epoch 2, total_step 35300, total loss is 28.62 , inference loss is 16.99, weight deacy loss is 11.63, training accuracy is 0.250000, time 57.077 samples/sec
+epoch 2, total_step 35320, total loss is 25.21 , inference loss is 13.59, weight deacy loss is 11.63, training accuracy is 0.187500, time 56.399 samples/sec
+epoch 2, total_step 35340, total loss is 29.31 , inference loss is 17.68, weight deacy loss is 11.62, training accuracy is 0.125000, time 56.575 samples/sec
+epoch 2, total_step 35360, total loss is 27.52 , inference loss is 15.90, weight deacy loss is 11.62, training accuracy is 0.250000, time 53.637 samples/sec
+epoch 2, total_step 35380, total loss is 26.63 , inference loss is 15.01, weight deacy loss is 11.62, training accuracy is 0.218750, time 56.451 samples/sec
+epoch 2, total_step 35400, total loss is 29.04 , inference loss is 17.42, weight deacy loss is 11.62, training accuracy is 0.156250, time 56.326 samples/sec
+epoch 2, total_step 35420, total loss is 24.24 , inference loss is 12.63, weight deacy loss is 11.62, training accuracy is 0.218750, time 55.352 samples/sec
+epoch 2, total_step 35440, total loss is 27.55 , inference loss is 15.94, weight deacy loss is 11.62, training accuracy is 0.093750, time 53.701 samples/sec
+epoch 2, total_step 35460, total loss is 26.60 , inference loss is 14.98, weight deacy loss is 11.61, training accuracy is 0.218750, time 55.820 samples/sec
+epoch 2, total_step 35480, total loss is 25.45 , inference loss is 13.83, weight deacy loss is 11.61, training accuracy is 0.093750, time 56.239 samples/sec
+epoch 2, total_step 35500, total loss is 25.49 , inference loss is 13.88, weight deacy loss is 11.61, training accuracy is 0.125000, time 56.052 samples/sec
+epoch 2, total_step 35520, total loss is 25.40 , inference loss is 13.79, weight deacy loss is 11.61, training accuracy is 0.125000, time 57.078 samples/sec
+epoch 2, total_step 35540, total loss is 23.63 , inference loss is 12.02, weight deacy loss is 11.61, training accuracy is 0.250000, time 56.350 samples/sec
+epoch 2, total_step 35560, total loss is 28.00 , inference loss is 16.40, weight deacy loss is 11.61, training accuracy is 0.093750, time 55.353 samples/sec
+epoch 2, total_step 35580, total loss is 29.53 , inference loss is 17.92, weight deacy loss is 11.61, training accuracy is 0.062500, time 55.682 samples/sec
+epoch 2, total_step 35600, total loss is 26.62 , inference loss is 15.01, weight deacy loss is 11.60, training accuracy is 0.187500, time 56.328 samples/sec
+epoch 2, total_step 35620, total loss is 25.59 , inference loss is 13.99, weight deacy loss is 11.60, training accuracy is 0.187500, time 55.818 samples/sec
+epoch 2, total_step 35640, total loss is 26.70 , inference loss is 15.10, weight deacy loss is 11.60, training accuracy is 0.187500, time 56.280 samples/sec
+epoch 2, total_step 35660, total loss is 27.07 , inference loss is 15.47, weight deacy loss is 11.60, training accuracy is 0.187500, time 56.372 samples/sec
+epoch 2, total_step 35680, total loss is 31.26 , inference loss is 19.67, weight deacy loss is 11.60, training accuracy is 0.093750, time 56.400 samples/sec
+epoch 2, total_step 35700, total loss is 27.09 , inference loss is 15.50, weight deacy loss is 11.60, training accuracy is 0.156250, time 56.200 samples/sec
+epoch 2, total_step 35720, total loss is 26.19 , inference loss is 14.59, weight deacy loss is 11.59, training accuracy is 0.187500, time 56.252 samples/sec
+epoch 2, total_step 35740, total loss is 23.74 , inference loss is 12.15, weight deacy loss is 11.59, training accuracy is 0.218750, time 56.470 samples/sec
+epoch 2, total_step 35760, total loss is 24.80 , inference loss is 13.21, weight deacy loss is 11.59, training accuracy is 0.125000, time 56.425 samples/sec
+epoch 2, total_step 35780, total loss is 29.00 , inference loss is 17.41, weight deacy loss is 11.59, training accuracy is 0.156250, time 55.584 samples/sec
+epoch 2, total_step 35800, total loss is 23.48 , inference loss is 11.89, weight deacy loss is 11.59, training accuracy is 0.281250, time 55.732 samples/sec
+epoch 2, total_step 35820, total loss is 29.31 , inference loss is 17.72, weight deacy loss is 11.59, training accuracy is 0.062500, time 55.394 samples/sec
+epoch 2, total_step 35840, total loss is 28.94 , inference loss is 17.35, weight deacy loss is 11.59, training accuracy is 0.125000, time 56.453 samples/sec
+epoch 2, total_step 35860, total loss is 27.39 , inference loss is 15.81, weight deacy loss is 11.58, training accuracy is 0.156250, time 56.192 samples/sec
+epoch 2, total_step 35880, total loss is 28.78 , inference loss is 17.20, weight deacy loss is 11.58, training accuracy is 0.125000, time 56.310 samples/sec
+epoch 2, total_step 35900, total loss is 27.38 , inference loss is 15.80, weight deacy loss is 11.58, training accuracy is 0.093750, time 56.577 samples/sec
+epoch 2, total_step 35920, total loss is 30.05 , inference loss is 18.47, weight deacy loss is 11.58, training accuracy is 0.281250, time 56.354 samples/sec
+epoch 2, total_step 35940, total loss is 23.27 , inference loss is 11.69, weight deacy loss is 11.58, training accuracy is 0.343750, time 55.937 samples/sec
+epoch 2, total_step 35960, total loss is 24.63 , inference loss is 13.05, weight deacy loss is 11.58, training accuracy is 0.218750, time 56.322 samples/sec
+epoch 2, total_step 35980, total loss is 29.32 , inference loss is 17.75, weight deacy loss is 11.57, training accuracy is 0.156250, time 55.870 samples/sec
+testing verification..
+(12000, 512)
+infer time 75.15278099999993
+best_threshold_index 107 0.980925925925926
+best_threshold_index 111 0.9803703703703703
+best_threshold_index 111 0.9822222222222222
+best_threshold_index 110 0.9812962962962963
+best_threshold_index 111 0.9812962962962963
+best_threshold_index 111 0.9811111111111112
+best_threshold_index 111 0.9814814814814815
+best_threshold_index 111 0.9811111111111112
+best_threshold_index 111 0.9796296296296296
+best_threshold_index 111 0.9807407407407407
+[lfw][36000]XNorm: 23.330442
+[lfw][36000]Accuracy-Flip: 0.98000+-0.00619
+testing verification..
+(14000, 512)
+infer time 86.98192600000012
+best_threshold_index 113 0.9758730158730159
+best_threshold_index 112 0.973015873015873
+best_threshold_index 109 0.9744444444444444
+best_threshold_index 113 0.9736507936507937
+best_threshold_index 113 0.9734920634920635
+best_threshold_index 113 0.9723809523809523
+best_threshold_index 109 0.9734920634920635
+best_threshold_index 114 0.973015873015873
+best_threshold_index 114 0.9741269841269842
+best_threshold_index 113 0.9746031746031746
+[cfp_ff][36000]XNorm: 22.717406
+[cfp_ff][36000]Accuracy-Flip: 0.97157+-0.00861
+testing verification..
+(14000, 512)
+infer time 87.44157499999997
+best_threshold_index 128 0.8704761904761905
+best_threshold_index 128 0.8665079365079366
+best_threshold_index 134 0.8666666666666667
+best_threshold_index 128 0.8684126984126984
+best_threshold_index 128 0.8677777777777778
+best_threshold_index 128 0.8642857142857143
+best_threshold_index 128 0.8650793650793651
+best_threshold_index 128 0.8665079365079366
+best_threshold_index 128 0.8671428571428571
+best_threshold_index 128 0.866984126984127
+[cfp_fp][36000]XNorm: 22.383073
+[cfp_fp][36000]Accuracy-Flip: 0.86471+-0.01472
+testing verification..
+(12000, 512)
+infer time 75.20203699999998
+best_threshold_index 129 0.840925925925926
+best_threshold_index 135 0.8431481481481482
+best_threshold_index 135 0.8455555555555555
+best_threshold_index 135 0.840925925925926
+best_threshold_index 135 0.8483333333333334
+best_threshold_index 135 0.8433333333333334
+best_threshold_index 135 0.8429629629629629
+best_threshold_index 135 0.8488888888888889
+best_threshold_index 136 0.8462962962962963
+best_threshold_index 135 0.845
+[agedb_30][36000]XNorm: 23.795915
+[agedb_30][36000]Accuracy-Flip: 0.84300+-0.02296
+test accuracy is: 0.9799999999999999
+epoch 2, total_step 36000, total loss is 23.93 , inference loss is 12.36, weight deacy loss is 11.57, training accuracy is 0.218750, time 53.114 samples/sec
+epoch 2, total_step 36020, total loss is 26.55 , inference loss is 14.98, weight deacy loss is 11.57, training accuracy is 0.187500, time 56.498 samples/sec
+epoch 2, total_step 36040, total loss is 28.09 , inference loss is 16.52, weight deacy loss is 11.57, training accuracy is 0.281250, time 55.579 samples/sec
+epoch 2, total_step 36060, total loss is 25.74 , inference loss is 14.17, weight deacy loss is 11.57, training accuracy is 0.218750, time 56.059 samples/sec
+epoch 2, total_step 36080, total loss is 25.00 , inference loss is 13.44, weight deacy loss is 11.57, training accuracy is 0.187500, time 56.698 samples/sec
+epoch 2, total_step 36100, total loss is 31.37 , inference loss is 19.81, weight deacy loss is 11.57, training accuracy is 0.125000, time 56.694 samples/sec
+epoch 2, total_step 36120, total loss is 22.23 , inference loss is 10.67, weight deacy loss is 11.56, training accuracy is 0.218750, time 56.813 samples/sec
+epoch 2, total_step 36140, total loss is 30.06 , inference loss is 18.50, weight deacy loss is 11.56, training accuracy is 0.062500, time 56.698 samples/sec
+epoch 2, total_step 36160, total loss is 24.36 , inference loss is 12.80, weight deacy loss is 11.56, training accuracy is 0.125000, time 56.778 samples/sec
+epoch 2, total_step 36180, total loss is 25.66 , inference loss is 14.10, weight deacy loss is 11.56, training accuracy is 0.156250, time 56.602 samples/sec
+epoch 2, total_step 36200, total loss is 26.48 , inference loss is 14.92, weight deacy loss is 11.56, training accuracy is 0.250000, time 56.319 samples/sec
+epoch 2, total_step 36220, total loss is 28.94 , inference loss is 17.38, weight deacy loss is 11.56, training accuracy is 0.031250, time 56.742 samples/sec
+epoch 2, total_step 36240, total loss is 22.99 , inference loss is 11.44, weight deacy loss is 11.55, training accuracy is 0.156250, time 57.171 samples/sec
+epoch 2, total_step 36260, total loss is 29.33 , inference loss is 17.77, weight deacy loss is 11.55, training accuracy is 0.125000, time 56.855 samples/sec
+epoch 2, total_step 36280, total loss is 25.21 , inference loss is 13.66, weight deacy loss is 11.55, training accuracy is 0.281250, time 56.270 samples/sec
+epoch 2, total_step 36300, total loss is 26.81 , inference loss is 15.27, weight deacy loss is 11.55, training accuracy is 0.250000, time 56.704 samples/sec
+epoch 2, total_step 36320, total loss is 24.10 , inference loss is 12.56, weight deacy loss is 11.55, training accuracy is 0.312500, time 55.584 samples/sec
+epoch 2, total_step 36340, total loss is 26.27 , inference loss is 14.72, weight deacy loss is 11.55, training accuracy is 0.156250, time 56.469 samples/sec
+epoch 2, total_step 36360, total loss is 21.27 , inference loss is 9.73, weight deacy loss is 11.55, training accuracy is 0.156250, time 56.401 samples/sec
+epoch 2, total_step 36380, total loss is 23.00 , inference loss is 11.45, weight deacy loss is 11.54, training accuracy is 0.187500, time 55.656 samples/sec
+epoch 2, total_step 36400, total loss is 27.79 , inference loss is 16.25, weight deacy loss is 11.54, training accuracy is 0.187500, time 56.198 samples/sec
+epoch 2, total_step 36420, total loss is 30.50 , inference loss is 18.96, weight deacy loss is 11.54, training accuracy is 0.156250, time 55.906 samples/sec
+epoch 2, total_step 36440, total loss is 27.48 , inference loss is 15.94, weight deacy loss is 11.54, training accuracy is 0.187500, time 56.023 samples/sec
+epoch 2, total_step 36460, total loss is 26.93 , inference loss is 15.39, weight deacy loss is 11.54, training accuracy is 0.250000, time 55.253 samples/sec
+epoch 2, total_step 36480, total loss is 27.01 , inference loss is 15.47, weight deacy loss is 11.54, training accuracy is 0.187500, time 56.055 samples/sec
+epoch 2, total_step 36500, total loss is 31.26 , inference loss is 19.73, weight deacy loss is 11.53, training accuracy is 0.062500, time 55.763 samples/sec
+epoch 2, total_step 36520, total loss is 28.30 , inference loss is 16.77, weight deacy loss is 11.53, training accuracy is 0.250000, time 56.162 samples/sec
+epoch 2, total_step 36540, total loss is 23.80 , inference loss is 12.26, weight deacy loss is 11.53, training accuracy is 0.218750, time 56.186 samples/sec
+epoch 2, total_step 36560, total loss is 24.71 , inference loss is 13.18, weight deacy loss is 11.53, training accuracy is 0.218750, time 55.130 samples/sec
+epoch 2, total_step 36580, total loss is 31.91 , inference loss is 20.38, weight deacy loss is 11.53, training accuracy is 0.125000, time 55.933 samples/sec
+epoch 2, total_step 36600, total loss is 28.18 , inference loss is 16.66, weight deacy loss is 11.53, training accuracy is 0.156250, time 55.961 samples/sec
+epoch 2, total_step 36620, total loss is 27.37 , inference loss is 15.85, weight deacy loss is 11.53, training accuracy is 0.187500, time 55.925 samples/sec
+epoch 2, total_step 36640, total loss is 25.83 , inference loss is 14.31, weight deacy loss is 11.52, training accuracy is 0.125000, time 55.747 samples/sec
+epoch 2, total_step 36660, total loss is 24.96 , inference loss is 13.44, weight deacy loss is 11.52, training accuracy is 0.187500, time 56.802 samples/sec
+epoch 2, total_step 36680, total loss is 28.65 , inference loss is 17.13, weight deacy loss is 11.52, training accuracy is 0.187500, time 56.597 samples/sec
+epoch 2, total_step 36700, total loss is 28.26 , inference loss is 16.74, weight deacy loss is 11.52, training accuracy is 0.125000, time 56.269 samples/sec
+epoch 2, total_step 36720, total loss is 27.40 , inference loss is 15.88, weight deacy loss is 11.52, training accuracy is 0.187500, time 56.757 samples/sec
+epoch 2, total_step 36740, total loss is 27.67 , inference loss is 16.15, weight deacy loss is 11.52, training accuracy is 0.156250, time 55.815 samples/sec
+epoch 2, total_step 36760, total loss is 24.95 , inference loss is 13.44, weight deacy loss is 11.51, training accuracy is 0.156250, time 56.165 samples/sec
+epoch 2, total_step 36780, total loss is 26.67 , inference loss is 15.16, weight deacy loss is 11.51, training accuracy is 0.062500, time 55.646 samples/sec
+epoch 2, total_step 36800, total loss is 24.61 , inference loss is 13.10, weight deacy loss is 11.51, training accuracy is 0.187500, time 55.365 samples/sec
+epoch 2, total_step 36820, total loss is 28.53 , inference loss is 17.02, weight deacy loss is 11.51, training accuracy is 0.125000, time 57.063 samples/sec
+epoch 2, total_step 36840, total loss is 27.23 , inference loss is 15.72, weight deacy loss is 11.51, training accuracy is 0.187500, time 55.978 samples/sec
+epoch 2, total_step 36860, total loss is 24.57 , inference loss is 13.07, weight deacy loss is 11.51, training accuracy is 0.250000, time 56.332 samples/sec
+epoch 2, total_step 36880, total loss is 23.57 , inference loss is 12.07, weight deacy loss is 11.51, training accuracy is 0.343750, time 56.266 samples/sec
+epoch 2, total_step 36900, total loss is 28.21 , inference loss is 16.70, weight deacy loss is 11.50, training accuracy is 0.187500, time 54.850 samples/sec
+epoch 2, total_step 36920, total loss is 22.57 , inference loss is 11.07, weight deacy loss is 11.50, training accuracy is 0.218750, time 56.595 samples/sec
+epoch 2, total_step 36940, total loss is 29.78 , inference loss is 18.28, weight deacy loss is 11.50, training accuracy is 0.187500, time 54.582 samples/sec
+epoch 2, total_step 36960, total loss is 26.48 , inference loss is 14.98, weight deacy loss is 11.50, training accuracy is 0.218750, time 55.981 samples/sec
+epoch 2, total_step 36980, total loss is 26.52 , inference loss is 15.02, weight deacy loss is 11.50, training accuracy is 0.187500, time 55.781 samples/sec
+epoch 2, total_step 37000, total loss is 33.87 , inference loss is 22.38, weight deacy loss is 11.50, training accuracy is 0.125000, time 55.941 samples/sec
+epoch 2, total_step 37020, total loss is 25.97 , inference loss is 14.48, weight deacy loss is 11.49, training accuracy is 0.218750, time 56.255 samples/sec
+epoch 2, total_step 37040, total loss is 28.12 , inference loss is 16.63, weight deacy loss is 11.49, training accuracy is 0.281250, time 56.254 samples/sec
+epoch 2, total_step 37060, total loss is 23.30 , inference loss is 11.81, weight deacy loss is 11.49, training accuracy is 0.156250, time 55.696 samples/sec
+epoch 2, total_step 37080, total loss is 31.81 , inference loss is 20.32, weight deacy loss is 11.49, training accuracy is 0.125000, time 56.858 samples/sec
+epoch 2, total_step 37100, total loss is 31.72 , inference loss is 20.23, weight deacy loss is 11.49, training accuracy is 0.031250, time 56.152 samples/sec
+epoch 2, total_step 37120, total loss is 26.82 , inference loss is 15.33, weight deacy loss is 11.49, training accuracy is 0.093750, time 55.165 samples/sec
+epoch 2, total_step 37140, total loss is 27.62 , inference loss is 16.14, weight deacy loss is 11.49, training accuracy is 0.093750, time 54.821 samples/sec
+epoch 2, total_step 37160, total loss is 27.48 , inference loss is 16.00, weight deacy loss is 11.48, training accuracy is 0.156250, time 56.352 samples/sec
+epoch 2, total_step 37180, total loss is 29.87 , inference loss is 18.39, weight deacy loss is 11.48, training accuracy is 0.187500, time 56.567 samples/sec
+epoch 2, total_step 37200, total loss is 25.83 , inference loss is 14.35, weight deacy loss is 11.48, training accuracy is 0.218750, time 56.028 samples/sec
+epoch 2, total_step 37220, total loss is 33.65 , inference loss is 22.17, weight deacy loss is 11.48, training accuracy is 0.000000, time 56.032 samples/sec
+epoch 2, total_step 37240, total loss is 27.31 , inference loss is 15.83, weight deacy loss is 11.48, training accuracy is 0.125000, time 56.621 samples/sec
+epoch 2, total_step 37260, total loss is 26.92 , inference loss is 15.45, weight deacy loss is 11.48, training accuracy is 0.218750, time 56.885 samples/sec
+epoch 2, total_step 37280, total loss is 26.17 , inference loss is 14.70, weight deacy loss is 11.47, training accuracy is 0.125000, time 56.039 samples/sec
+epoch 2, total_step 37300, total loss is 29.75 , inference loss is 18.28, weight deacy loss is 11.47, training accuracy is 0.156250, time 56.083 samples/sec
+epoch 2, total_step 37320, total loss is 25.26 , inference loss is 13.79, weight deacy loss is 11.47, training accuracy is 0.218750, time 56.441 samples/sec
+epoch 2, total_step 37340, total loss is 24.85 , inference loss is 13.38, weight deacy loss is 11.47, training accuracy is 0.218750, time 56.146 samples/sec
+epoch 2, total_step 37360, total loss is 31.80 , inference loss is 20.33, weight deacy loss is 11.47, training accuracy is 0.093750, time 56.740 samples/sec
+epoch 2, total_step 37380, total loss is 29.64 , inference loss is 18.17, weight deacy loss is 11.47, training accuracy is 0.031250, time 56.191 samples/sec
+epoch 2, total_step 37400, total loss is 23.63 , inference loss is 12.17, weight deacy loss is 11.47, training accuracy is 0.343750, time 56.826 samples/sec
+epoch 2, total_step 37420, total loss is 27.29 , inference loss is 15.83, weight deacy loss is 11.46, training accuracy is 0.250000, time 56.568 samples/sec
+epoch 2, total_step 37440, total loss is 28.37 , inference loss is 16.91, weight deacy loss is 11.46, training accuracy is 0.125000, time 56.258 samples/sec
+epoch 2, total_step 37460, total loss is 27.18 , inference loss is 15.72, weight deacy loss is 11.46, training accuracy is 0.250000, time 55.802 samples/sec
+epoch 2, total_step 37480, total loss is 25.40 , inference loss is 13.94, weight deacy loss is 11.46, training accuracy is 0.218750, time 55.766 samples/sec
+epoch 2, total_step 37500, total loss is 28.36 , inference loss is 16.90, weight deacy loss is 11.46, training accuracy is 0.062500, time 55.945 samples/sec
+epoch 2, total_step 37520, total loss is 27.80 , inference loss is 16.35, weight deacy loss is 11.46, training accuracy is 0.156250, time 55.844 samples/sec
+epoch 2, total_step 37540, total loss is 27.70 , inference loss is 16.25, weight deacy loss is 11.45, training accuracy is 0.093750, time 55.834 samples/sec
+epoch 2, total_step 37560, total loss is 26.10 , inference loss is 14.65, weight deacy loss is 11.45, training accuracy is 0.187500, time 56.664 samples/sec
+epoch 2, total_step 37580, total loss is 26.83 , inference loss is 15.38, weight deacy loss is 11.45, training accuracy is 0.093750, time 56.705 samples/sec
+epoch 2, total_step 37600, total loss is 30.50 , inference loss is 19.05, weight deacy loss is 11.45, training accuracy is 0.093750, time 56.166 samples/sec
+epoch 2, total_step 37620, total loss is 25.58 , inference loss is 14.13, weight deacy loss is 11.45, training accuracy is 0.125000, time 56.117 samples/sec
+epoch 2, total_step 37640, total loss is 24.01 , inference loss is 12.56, weight deacy loss is 11.45, training accuracy is 0.187500, time 55.065 samples/sec
+epoch 2, total_step 37660, total loss is 31.17 , inference loss is 19.73, weight deacy loss is 11.45, training accuracy is 0.031250, time 56.337 samples/sec
+epoch 2, total_step 37680, total loss is 27.51 , inference loss is 16.06, weight deacy loss is 11.44, training accuracy is 0.125000, time 56.038 samples/sec
+epoch 2, total_step 37700, total loss is 28.55 , inference loss is 17.11, weight deacy loss is 11.44, training accuracy is 0.000000, time 56.955 samples/sec
+epoch 2, total_step 37720, total loss is 29.32 , inference loss is 17.88, weight deacy loss is 11.44, training accuracy is 0.125000, time 55.480 samples/sec
+epoch 2, total_step 37740, total loss is 26.75 , inference loss is 15.31, weight deacy loss is 11.44, training accuracy is 0.125000, time 56.649 samples/sec
+epoch 2, total_step 37760, total loss is 28.85 , inference loss is 17.42, weight deacy loss is 11.44, training accuracy is 0.187500, time 55.872 samples/sec
+epoch 2, total_step 37780, total loss is 25.68 , inference loss is 14.24, weight deacy loss is 11.44, training accuracy is 0.125000, time 56.276 samples/sec
+epoch 2, total_step 37800, total loss is 30.61 , inference loss is 19.17, weight deacy loss is 11.44, training accuracy is 0.125000, time 55.776 samples/sec
+epoch 2, total_step 37820, total loss is 30.46 , inference loss is 19.03, weight deacy loss is 11.43, training accuracy is 0.062500, time 56.004 samples/sec
+epoch 2, total_step 37840, total loss is 26.90 , inference loss is 15.46, weight deacy loss is 11.43, training accuracy is 0.125000, time 57.074 samples/sec
+epoch 2, total_step 37860, total loss is 25.29 , inference loss is 13.85, weight deacy loss is 11.43, training accuracy is 0.218750, time 56.667 samples/sec
+epoch 2, total_step 37880, total loss is 26.57 , inference loss is 15.14, weight deacy loss is 11.43, training accuracy is 0.281250, time 57.349 samples/sec
+epoch 2, total_step 37900, total loss is 30.05 , inference loss is 18.62, weight deacy loss is 11.43, training accuracy is 0.125000, time 55.777 samples/sec
+epoch 2, total_step 37920, total loss is 23.46 , inference loss is 12.03, weight deacy loss is 11.43, training accuracy is 0.250000, time 55.738 samples/sec
+epoch 2, total_step 37940, total loss is 25.06 , inference loss is 13.64, weight deacy loss is 11.43, training accuracy is 0.156250, time 56.346 samples/sec
+epoch 2, total_step 37960, total loss is 24.94 , inference loss is 13.52, weight deacy loss is 11.42, training accuracy is 0.250000, time 56.461 samples/sec
+epoch 2, total_step 37980, total loss is 31.43 , inference loss is 20.01, weight deacy loss is 11.42, training accuracy is 0.156250, time 53.596 samples/sec
+testing verification..
+(12000, 512)
+infer time 75.37912700000001
+best_threshold_index 114 0.982037037037037
+best_threshold_index 114 0.9811111111111112
+best_threshold_index 111 0.9827777777777778
+best_threshold_index 114 0.9837037037037037
+best_threshold_index 114 0.9824074074074074
+best_threshold_index 114 0.9812962962962963
+best_threshold_index 114 0.982037037037037
+best_threshold_index 114 0.9807407407407407
+best_threshold_index 114 0.98
+best_threshold_index 111 0.9805555555555555
+[lfw][38000]XNorm: 22.842519
+[lfw][38000]Accuracy-Flip: 0.98100+-0.00975
+testing verification..
+(14000, 512)
+infer time 87.05877499999995
+best_threshold_index 114 0.9753968253968254
+best_threshold_index 114 0.9741269841269842
+best_threshold_index 114 0.9742857142857143
+best_threshold_index 114 0.9733333333333334
+best_threshold_index 114 0.9738095238095238
+best_threshold_index 114 0.9726984126984127
+best_threshold_index 114 0.974920634920635
+best_threshold_index 114 0.9734920634920635
+best_threshold_index 114 0.9736507936507937
+best_threshold_index 114 0.9757142857142858
+[cfp_ff][38000]XNorm: 22.263371
+[cfp_ff][38000]Accuracy-Flip: 0.97414+-0.00814
+testing verification..
+(14000, 512)
+infer time 86.68211800000007
+best_threshold_index 136 0.8720634920634921
+best_threshold_index 134 0.8703174603174603
+best_threshold_index 136 0.8687301587301587
+best_threshold_index 135 0.87
+best_threshold_index 136 0.8722222222222222
+best_threshold_index 134 0.866984126984127
+best_threshold_index 134 0.8688888888888889
+best_threshold_index 134 0.8688888888888889
+best_threshold_index 134 0.8704761904761905
+best_threshold_index 134 0.87
+[cfp_fp][38000]XNorm: 21.814284
+[cfp_fp][38000]Accuracy-Flip: 0.86700+-0.01459
+testing verification..
+(12000, 512)
+infer time 73.91691600000001
+best_threshold_index 131 0.8494444444444444
+best_threshold_index 136 0.8483333333333334
+best_threshold_index 131 0.8531481481481481
+best_threshold_index 131 0.8505555555555555
+best_threshold_index 131 0.8555555555555555
+best_threshold_index 131 0.8501851851851852
+best_threshold_index 131 0.8525925925925926
+best_threshold_index 131 0.8548148148148148
+best_threshold_index 131 0.8518518518518519
+best_threshold_index 131 0.8525925925925926
+[agedb_30][38000]XNorm: 22.259738
+[agedb_30][38000]Accuracy-Flip: 0.85050+-0.01868
+test accuracy is: 0.9809999999999999
+epoch 2, total_step 38000, total loss is 23.96 , inference loss is 12.54, weight deacy loss is 11.42, training accuracy is 0.156250, time 52.753 samples/sec
+epoch 2, total_step 38020, total loss is 29.31 , inference loss is 17.89, weight deacy loss is 11.42, training accuracy is 0.093750, time 56.549 samples/sec
+epoch 2, total_step 38040, total loss is 25.18 , inference loss is 13.76, weight deacy loss is 11.42, training accuracy is 0.250000, time 55.436 samples/sec
+epoch 2, total_step 38060, total loss is 24.09 , inference loss is 12.67, weight deacy loss is 11.42, training accuracy is 0.281250, time 56.792 samples/sec
+epoch 2, total_step 38080, total loss is 26.03 , inference loss is 14.61, weight deacy loss is 11.41, training accuracy is 0.250000, time 57.559 samples/sec
+epoch 2, total_step 38100, total loss is 27.47 , inference loss is 16.06, weight deacy loss is 11.41, training accuracy is 0.156250, time 55.711 samples/sec
+epoch 2, total_step 38120, total loss is 24.45 , inference loss is 13.03, weight deacy loss is 11.41, training accuracy is 0.218750, time 55.789 samples/sec
+epoch 2, total_step 38140, total loss is 25.09 , inference loss is 13.68, weight deacy loss is 11.41, training accuracy is 0.093750, time 56.514 samples/sec
+epoch 2, total_step 38160, total loss is 26.13 , inference loss is 14.72, weight deacy loss is 11.41, training accuracy is 0.218750, time 56.233 samples/sec
+epoch 2, total_step 38180, total loss is 26.96 , inference loss is 15.56, weight deacy loss is 11.41, training accuracy is 0.250000, time 56.232 samples/sec
+epoch 2, total_step 38200, total loss is 24.52 , inference loss is 13.11, weight deacy loss is 11.41, training accuracy is 0.281250, time 56.072 samples/sec
+epoch 2, total_step 38220, total loss is 23.08 , inference loss is 11.67, weight deacy loss is 11.40, training accuracy is 0.250000, time 56.201 samples/sec
+epoch 2, total_step 38240, total loss is 26.60 , inference loss is 15.20, weight deacy loss is 11.40, training accuracy is 0.156250, time 56.516 samples/sec
+epoch 2, total_step 38260, total loss is 23.62 , inference loss is 12.22, weight deacy loss is 11.40, training accuracy is 0.218750, time 56.104 samples/sec
+epoch 2, total_step 38280, total loss is 25.30 , inference loss is 13.90, weight deacy loss is 11.40, training accuracy is 0.218750, time 52.757 samples/sec
+epoch 2, total_step 38300, total loss is 25.38 , inference loss is 13.98, weight deacy loss is 11.40, training accuracy is 0.218750, time 57.018 samples/sec
+epoch 2, total_step 38320, total loss is 24.24 , inference loss is 12.84, weight deacy loss is 11.40, training accuracy is 0.218750, time 56.164 samples/sec
+epoch 2, total_step 38340, total loss is 27.17 , inference loss is 15.77, weight deacy loss is 11.40, training accuracy is 0.218750, time 56.698 samples/sec
+epoch 2, total_step 38360, total loss is 27.77 , inference loss is 16.38, weight deacy loss is 11.39, training accuracy is 0.250000, time 57.046 samples/sec
+epoch 2, total_step 38380, total loss is 20.07 , inference loss is 8.67, weight deacy loss is 11.39, training accuracy is 0.250000, time 55.683 samples/sec
+epoch 2, total_step 38400, total loss is 22.69 , inference loss is 11.30, weight deacy loss is 11.39, training accuracy is 0.250000, time 56.586 samples/sec
+epoch 2, total_step 38420, total loss is 24.55 , inference loss is 13.16, weight deacy loss is 11.39, training accuracy is 0.218750, time 55.858 samples/sec
+epoch 2, total_step 38440, total loss is 26.58 , inference loss is 15.19, weight deacy loss is 11.39, training accuracy is 0.156250, time 55.696 samples/sec
+epoch 2, total_step 38460, total loss is 24.55 , inference loss is 13.16, weight deacy loss is 11.39, training accuracy is 0.250000, time 55.811 samples/sec
+epoch 2, total_step 38480, total loss is 27.11 , inference loss is 15.72, weight deacy loss is 11.38, training accuracy is 0.156250, time 56.933 samples/sec
+epoch 2, total_step 38500, total loss is 23.42 , inference loss is 12.03, weight deacy loss is 11.38, training accuracy is 0.250000, time 57.056 samples/sec
+epoch 2, total_step 38520, total loss is 26.59 , inference loss is 15.21, weight deacy loss is 11.38, training accuracy is 0.281250, time 56.458 samples/sec
+epoch 2, total_step 38540, total loss is 28.94 , inference loss is 17.56, weight deacy loss is 11.38, training accuracy is 0.218750, time 56.084 samples/sec
+epoch 2, total_step 38560, total loss is 28.56 , inference loss is 17.18, weight deacy loss is 11.38, training accuracy is 0.093750, time 55.315 samples/sec
+epoch 2, total_step 38580, total loss is 28.77 , inference loss is 17.39, weight deacy loss is 11.38, training accuracy is 0.218750, time 56.019 samples/sec
+epoch 2, total_step 38600, total loss is 23.68 , inference loss is 12.30, weight deacy loss is 11.38, training accuracy is 0.218750, time 54.265 samples/sec
+epoch 2, total_step 38620, total loss is 27.53 , inference loss is 16.16, weight deacy loss is 11.37, training accuracy is 0.093750, time 55.795 samples/sec
+epoch 2, total_step 38640, total loss is 27.75 , inference loss is 16.38, weight deacy loss is 11.37, training accuracy is 0.156250, time 57.632 samples/sec
+epoch 2, total_step 38660, total loss is 27.84 , inference loss is 16.47, weight deacy loss is 11.37, training accuracy is 0.062500, time 56.356 samples/sec
+epoch 2, total_step 38680, total loss is 23.96 , inference loss is 12.59, weight deacy loss is 11.37, training accuracy is 0.187500, time 55.409 samples/sec
+epoch 2, total_step 38700, total loss is 24.35 , inference loss is 12.98, weight deacy loss is 11.37, training accuracy is 0.156250, time 55.185 samples/sec
+epoch 2, total_step 38720, total loss is 27.96 , inference loss is 16.59, weight deacy loss is 11.37, training accuracy is 0.218750, time 57.415 samples/sec
+epoch 2, total_step 38740, total loss is 23.63 , inference loss is 12.27, weight deacy loss is 11.36, training accuracy is 0.218750, time 56.562 samples/sec
+epoch 2, total_step 38760, total loss is 29.61 , inference loss is 18.25, weight deacy loss is 11.36, training accuracy is 0.156250, time 56.266 samples/sec
+epoch 2, total_step 38780, total loss is 26.90 , inference loss is 15.54, weight deacy loss is 11.36, training accuracy is 0.187500, time 55.872 samples/sec
+epoch 2, total_step 38800, total loss is 27.25 , inference loss is 15.89, weight deacy loss is 11.36, training accuracy is 0.187500, time 55.884 samples/sec
+epoch 2, total_step 38820, total loss is 22.01 , inference loss is 10.65, weight deacy loss is 11.36, training accuracy is 0.218750, time 53.219 samples/sec
+epoch 2, total_step 38840, total loss is 31.98 , inference loss is 20.62, weight deacy loss is 11.36, training accuracy is 0.156250, time 55.991 samples/sec
+epoch 2, total_step 38860, total loss is 23.53 , inference loss is 12.18, weight deacy loss is 11.36, training accuracy is 0.125000, time 56.216 samples/sec
+epoch 2, total_step 38880, total loss is 31.41 , inference loss is 20.06, weight deacy loss is 11.35, training accuracy is 0.031250, time 55.872 samples/sec
+epoch 2, total_step 38900, total loss is 26.63 , inference loss is 15.28, weight deacy loss is 11.35, training accuracy is 0.187500, time 56.271 samples/sec
+epoch 2, total_step 38920, total loss is 30.07 , inference loss is 18.72, weight deacy loss is 11.35, training accuracy is 0.156250, time 56.243 samples/sec
+epoch 2, total_step 38940, total loss is 25.84 , inference loss is 14.49, weight deacy loss is 11.35, training accuracy is 0.125000, time 56.471 samples/sec
+epoch 2, total_step 38960, total loss is 30.40 , inference loss is 19.05, weight deacy loss is 11.35, training accuracy is 0.187500, time 56.208 samples/sec
+epoch 2, total_step 38980, total loss is 22.68 , inference loss is 11.33, weight deacy loss is 11.35, training accuracy is 0.187500, time 56.829 samples/sec
+epoch 2, total_step 39000, total loss is 26.22 , inference loss is 14.88, weight deacy loss is 11.35, training accuracy is 0.125000, time 55.802 samples/sec
+epoch 2, total_step 39020, total loss is 25.99 , inference loss is 14.65, weight deacy loss is 11.34, training accuracy is 0.187500, time 55.943 samples/sec
+epoch 2, total_step 39040, total loss is 24.21 , inference loss is 12.87, weight deacy loss is 11.34, training accuracy is 0.218750, time 52.574 samples/sec
+epoch 2, total_step 39060, total loss is 27.87 , inference loss is 16.53, weight deacy loss is 11.34, training accuracy is 0.125000, time 57.383 samples/sec
+epoch 2, total_step 39080, total loss is 24.97 , inference loss is 13.63, weight deacy loss is 11.34, training accuracy is 0.187500, time 56.420 samples/sec
+epoch 2, total_step 39100, total loss is 27.11 , inference loss is 15.77, weight deacy loss is 11.34, training accuracy is 0.156250, time 56.844 samples/sec
+epoch 2, total_step 39120, total loss is 31.71 , inference loss is 20.37, weight deacy loss is 11.34, training accuracy is 0.125000, time 53.959 samples/sec
+epoch 2, total_step 39140, total loss is 27.76 , inference loss is 16.43, weight deacy loss is 11.34, training accuracy is 0.125000, time 54.986 samples/sec
+epoch 2, total_step 39160, total loss is 24.37 , inference loss is 13.03, weight deacy loss is 11.33, training accuracy is 0.343750, time 56.573 samples/sec
+epoch 2, total_step 39180, total loss is 27.31 , inference loss is 15.98, weight deacy loss is 11.33, training accuracy is 0.218750, time 56.723 samples/sec
+epoch 2, total_step 39200, total loss is 29.78 , inference loss is 18.45, weight deacy loss is 11.33, training accuracy is 0.062500, time 55.869 samples/sec
+epoch 2, total_step 39220, total loss is 30.26 , inference loss is 18.93, weight deacy loss is 11.33, training accuracy is 0.218750, time 55.734 samples/sec
+epoch 2, total_step 39240, total loss is 23.86 , inference loss is 12.53, weight deacy loss is 11.33, training accuracy is 0.187500, time 56.134 samples/sec
+epoch 2, total_step 39260, total loss is 30.36 , inference loss is 19.03, weight deacy loss is 11.33, training accuracy is 0.156250, time 53.664 samples/sec
+epoch 2, total_step 39280, total loss is 30.28 , inference loss is 18.96, weight deacy loss is 11.32, training accuracy is 0.125000, time 57.106 samples/sec
+epoch 2, total_step 39300, total loss is 26.66 , inference loss is 15.34, weight deacy loss is 11.32, training accuracy is 0.156250, time 54.808 samples/sec
+epoch 2, total_step 39320, total loss is 27.68 , inference loss is 16.36, weight deacy loss is 11.32, training accuracy is 0.156250, time 56.445 samples/sec
+epoch 2, total_step 39340, total loss is 25.82 , inference loss is 14.50, weight deacy loss is 11.32, training accuracy is 0.156250, time 56.415 samples/sec
+epoch 2, total_step 39360, total loss is 23.29 , inference loss is 11.97, weight deacy loss is 11.32, training accuracy is 0.187500, time 56.352 samples/sec
+epoch 2, total_step 39380, total loss is 26.05 , inference loss is 14.74, weight deacy loss is 11.32, training accuracy is 0.156250, time 56.428 samples/sec
+epoch 2, total_step 39400, total loss is 30.88 , inference loss is 19.56, weight deacy loss is 11.32, training accuracy is 0.187500, time 56.734 samples/sec
+epoch 2, total_step 39420, total loss is 30.91 , inference loss is 19.60, weight deacy loss is 11.31, training accuracy is 0.125000, time 56.360 samples/sec
+epoch 2, total_step 39440, total loss is 28.51 , inference loss is 17.20, weight deacy loss is 11.31, training accuracy is 0.156250, time 55.785 samples/sec
+epoch 2, total_step 39460, total loss is 27.38 , inference loss is 16.07, weight deacy loss is 11.31, training accuracy is 0.156250, time 55.377 samples/sec
+epoch 2, total_step 39480, total loss is 26.29 , inference loss is 14.98, weight deacy loss is 11.31, training accuracy is 0.093750, time 56.463 samples/sec
+epoch 2, total_step 39500, total loss is 30.64 , inference loss is 19.33, weight deacy loss is 11.31, training accuracy is 0.218750, time 55.881 samples/sec
+epoch 2, total_step 39520, total loss is 24.40 , inference loss is 13.10, weight deacy loss is 11.31, training accuracy is 0.187500, time 55.442 samples/sec
+epoch 2, total_step 39540, total loss is 26.05 , inference loss is 14.74, weight deacy loss is 11.31, training accuracy is 0.093750, time 55.915 samples/sec
+epoch 2, total_step 39560, total loss is 27.63 , inference loss is 16.32, weight deacy loss is 11.30, training accuracy is 0.093750, time 56.495 samples/sec
+epoch 2, total_step 39580, total loss is 29.48 , inference loss is 18.18, weight deacy loss is 11.30, training accuracy is 0.250000, time 56.302 samples/sec
+epoch 2, total_step 39600, total loss is 29.05 , inference loss is 17.75, weight deacy loss is 11.30, training accuracy is 0.125000, time 55.971 samples/sec
+epoch 2, total_step 39620, total loss is 24.41 , inference loss is 13.11, weight deacy loss is 11.30, training accuracy is 0.250000, time 56.216 samples/sec
+epoch 2, total_step 39640, total loss is 28.46 , inference loss is 17.16, weight deacy loss is 11.30, training accuracy is 0.062500, time 55.943 samples/sec
+epoch 2, total_step 39660, total loss is 22.48 , inference loss is 11.19, weight deacy loss is 11.30, training accuracy is 0.218750, time 55.111 samples/sec
+epoch 2, total_step 39680, total loss is 25.49 , inference loss is 14.19, weight deacy loss is 11.30, training accuracy is 0.218750, time 56.488 samples/sec
+epoch 2, total_step 39700, total loss is 23.67 , inference loss is 12.37, weight deacy loss is 11.29, training accuracy is 0.093750, time 55.916 samples/sec
+epoch 2, total_step 39720, total loss is 24.72 , inference loss is 13.43, weight deacy loss is 11.29, training accuracy is 0.281250, time 56.738 samples/sec
+epoch 2, total_step 39740, total loss is 26.24 , inference loss is 14.95, weight deacy loss is 11.29, training accuracy is 0.218750, time 56.296 samples/sec
+epoch 2, total_step 39760, total loss is 28.05 , inference loss is 16.76, weight deacy loss is 11.29, training accuracy is 0.187500, time 56.578 samples/sec
+epoch 2, total_step 39780, total loss is 26.79 , inference loss is 15.50, weight deacy loss is 11.29, training accuracy is 0.156250, time 55.393 samples/sec
+epoch 2, total_step 39800, total loss is 23.27 , inference loss is 11.98, weight deacy loss is 11.29, training accuracy is 0.125000, time 56.389 samples/sec
+epoch 2, total_step 39820, total loss is 20.83 , inference loss is 9.54, weight deacy loss is 11.29, training accuracy is 0.281250, time 56.481 samples/sec
+epoch 2, total_step 39840, total loss is 29.71 , inference loss is 18.43, weight deacy loss is 11.28, training accuracy is 0.125000, time 56.036 samples/sec
+epoch 2, total_step 39860, total loss is 30.32 , inference loss is 19.03, weight deacy loss is 11.28, training accuracy is 0.093750, time 56.646 samples/sec
+epoch 2, total_step 39880, total loss is 28.93 , inference loss is 17.65, weight deacy loss is 11.28, training accuracy is 0.062500, time 56.490 samples/sec
+epoch 2, total_step 39900, total loss is 24.47 , inference loss is 13.19, weight deacy loss is 11.28, training accuracy is 0.218750, time 56.125 samples/sec
+epoch 2, total_step 39920, total loss is 27.75 , inference loss is 16.47, weight deacy loss is 11.28, training accuracy is 0.062500, time 56.088 samples/sec
+epoch 2, total_step 39940, total loss is 25.76 , inference loss is 14.49, weight deacy loss is 11.28, training accuracy is 0.093750, time 56.312 samples/sec
+epoch 2, total_step 39960, total loss is 23.94 , inference loss is 12.67, weight deacy loss is 11.28, training accuracy is 0.187500, time 56.081 samples/sec
+epoch 2, total_step 39980, total loss is 27.85 , inference loss is 16.57, weight deacy loss is 11.27, training accuracy is 0.250000, time 56.311 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.29367299999994
+best_threshold_index 107 0.9790740740740741
+best_threshold_index 107 0.9790740740740741
+best_threshold_index 108 0.9807407407407407
+best_threshold_index 107 0.9807407407407407
+best_threshold_index 107 0.9798148148148148
+best_threshold_index 111 0.9787037037037037
+best_threshold_index 107 0.9794444444444445
+best_threshold_index 108 0.9787037037037037
+best_threshold_index 107 0.9779629629629629
+best_threshold_index 107 0.9781481481481481
+[lfw][40000]XNorm: 23.928947
+[lfw][40000]Accuracy-Flip: 0.97817+-0.00808
+testing verification..
+(14000, 512)
+infer time 85.957247
+best_threshold_index 111 0.9742857142857143
+best_threshold_index 109 0.9722222222222222
+best_threshold_index 109 0.9736507936507937
+best_threshold_index 110 0.9726984126984127
+best_threshold_index 109 0.973015873015873
+best_threshold_index 109 0.9720634920634921
+best_threshold_index 111 0.973015873015873
+best_threshold_index 109 0.9722222222222222
+best_threshold_index 109 0.9736507936507937
+best_threshold_index 109 0.9741269841269842
+[cfp_ff][40000]XNorm: 23.124479
+[cfp_ff][40000]Accuracy-Flip: 0.97214+-0.00689
+testing verification..
+(14000, 512)
+infer time 86.12050599999999
+best_threshold_index 128 0.8679365079365079
+best_threshold_index 130 0.8638095238095238
+best_threshold_index 128 0.8649206349206349
+best_threshold_index 127 0.8657142857142858
+best_threshold_index 127 0.8685714285714285
+best_threshold_index 128 0.8642857142857143
+best_threshold_index 127 0.8634920634920635
+best_threshold_index 128 0.8657142857142858
+best_threshold_index 128 0.8653968253968254
+best_threshold_index 135 0.866031746031746
+[cfp_fp][40000]XNorm: 23.538579
+[cfp_fp][40000]Accuracy-Flip: 0.86229+-0.01440
+testing verification..
+(12000, 512)
+infer time 73.84221099999992
+best_threshold_index 131 0.8496296296296296
+best_threshold_index 131 0.8483333333333334
+best_threshold_index 131 0.85
+best_threshold_index 127 0.8490740740740741
+best_threshold_index 131 0.855
+best_threshold_index 127 0.8487037037037037
+best_threshold_index 126 0.8529629629629629
+best_threshold_index 131 0.8555555555555555
+best_threshold_index 131 0.8511111111111112
+best_threshold_index 131 0.8505555555555555
+[agedb_30][40000]XNorm: 23.883560
+[agedb_30][40000]Accuracy-Flip: 0.84700+-0.02284
+test accuracy is: 0.9781666666666666
+epoch 2, total_step 40000, total loss is 29.65 , inference loss is 18.38, weight deacy loss is 11.27, training accuracy is 0.125000, time 52.655 samples/sec
+epoch 2, total_step 40020, total loss is 24.26 , inference loss is 12.99, weight deacy loss is 11.27, training accuracy is 0.125000, time 56.330 samples/sec
+epoch 2, total_step 40040, total loss is 24.51 , inference loss is 13.24, weight deacy loss is 11.27, training accuracy is 0.125000, time 55.616 samples/sec
+epoch 2, total_step 40060, total loss is 22.11 , inference loss is 10.84, weight deacy loss is 11.27, training accuracy is 0.218750, time 56.084 samples/sec
+epoch 2, total_step 40080, total loss is 27.69 , inference loss is 16.43, weight deacy loss is 11.27, training accuracy is 0.125000, time 56.485 samples/sec
+epoch 2, total_step 40100, total loss is 25.67 , inference loss is 14.40, weight deacy loss is 11.27, training accuracy is 0.187500, time 56.791 samples/sec
+epoch 2, total_step 40120, total loss is 24.50 , inference loss is 13.23, weight deacy loss is 11.26, training accuracy is 0.250000, time 55.778 samples/sec
+epoch 2, total_step 40140, total loss is 30.55 , inference loss is 19.28, weight deacy loss is 11.26, training accuracy is 0.093750, time 56.270 samples/sec
+epoch 2, total_step 40160, total loss is 28.18 , inference loss is 16.92, weight deacy loss is 11.26, training accuracy is 0.156250, time 56.571 samples/sec
+epoch 2, total_step 40180, total loss is 24.55 , inference loss is 13.29, weight deacy loss is 11.26, training accuracy is 0.187500, time 54.998 samples/sec
+epoch 2, total_step 40200, total loss is 23.69 , inference loss is 12.43, weight deacy loss is 11.26, training accuracy is 0.187500, time 56.020 samples/sec
+epoch 2, total_step 40220, total loss is 31.34 , inference loss is 20.08, weight deacy loss is 11.26, training accuracy is 0.093750, time 55.071 samples/sec
+epoch 2, total_step 40240, total loss is 27.37 , inference loss is 16.11, weight deacy loss is 11.25, training accuracy is 0.093750, time 56.398 samples/sec
+epoch 2, total_step 40260, total loss is 29.05 , inference loss is 17.80, weight deacy loss is 11.25, training accuracy is 0.156250, time 53.776 samples/sec
+epoch 2, total_step 40280, total loss is 24.60 , inference loss is 13.35, weight deacy loss is 11.25, training accuracy is 0.250000, time 56.067 samples/sec
+epoch 2, total_step 40300, total loss is 24.63 , inference loss is 13.38, weight deacy loss is 11.25, training accuracy is 0.250000, time 55.732 samples/sec
+epoch 2, total_step 40320, total loss is 25.31 , inference loss is 14.06, weight deacy loss is 11.25, training accuracy is 0.250000, time 56.002 samples/sec
+epoch 2, total_step 40340, total loss is 29.65 , inference loss is 18.40, weight deacy loss is 11.25, training accuracy is 0.125000, time 55.618 samples/sec
+epoch 2, total_step 40360, total loss is 25.12 , inference loss is 13.87, weight deacy loss is 11.25, training accuracy is 0.156250, time 56.237 samples/sec
+epoch 2, total_step 40380, total loss is 21.94 , inference loss is 10.70, weight deacy loss is 11.24, training accuracy is 0.250000, time 57.223 samples/sec
+epoch 2, total_step 40400, total loss is 27.21 , inference loss is 15.97, weight deacy loss is 11.24, training accuracy is 0.125000, time 56.529 samples/sec
+epoch 2, total_step 40420, total loss is 26.19 , inference loss is 14.95, weight deacy loss is 11.24, training accuracy is 0.187500, time 55.841 samples/sec
+epoch 2, total_step 40440, total loss is 29.70 , inference loss is 18.46, weight deacy loss is 11.24, training accuracy is 0.093750, time 55.385 samples/sec
+epoch 2, total_step 40460, total loss is 24.00 , inference loss is 12.76, weight deacy loss is 11.24, training accuracy is 0.125000, time 56.292 samples/sec
+epoch 2, total_step 40480, total loss is 25.65 , inference loss is 14.41, weight deacy loss is 11.24, training accuracy is 0.187500, time 53.839 samples/sec
+epoch 2, total_step 40500, total loss is 24.61 , inference loss is 13.37, weight deacy loss is 11.24, training accuracy is 0.062500, time 55.715 samples/sec
+epoch 2, total_step 40520, total loss is 26.82 , inference loss is 15.59, weight deacy loss is 11.23, training accuracy is 0.125000, time 55.822 samples/sec
+epoch 2, total_step 40540, total loss is 28.96 , inference loss is 17.73, weight deacy loss is 11.23, training accuracy is 0.218750, time 56.590 samples/sec
+epoch 2, total_step 40560, total loss is 26.21 , inference loss is 14.98, weight deacy loss is 11.23, training accuracy is 0.156250, time 56.055 samples/sec
+epoch 2, total_step 40580, total loss is 26.46 , inference loss is 15.23, weight deacy loss is 11.23, training accuracy is 0.093750, time 56.394 samples/sec
+epoch 2, total_step 40600, total loss is 26.50 , inference loss is 15.28, weight deacy loss is 11.23, training accuracy is 0.093750, time 55.368 samples/sec
+epoch 2, total_step 40620, total loss is 23.55 , inference loss is 12.32, weight deacy loss is 11.23, training accuracy is 0.281250, time 55.371 samples/sec
+epoch 2, total_step 40640, total loss is 27.46 , inference loss is 16.23, weight deacy loss is 11.23, training accuracy is 0.125000, time 55.343 samples/sec
+epoch 2, total_step 40660, total loss is 27.93 , inference loss is 16.71, weight deacy loss is 11.22, training accuracy is 0.218750, time 55.925 samples/sec
+epoch 2, total_step 40680, total loss is 33.30 , inference loss is 22.08, weight deacy loss is 11.22, training accuracy is 0.062500, time 55.442 samples/sec
+epoch 2, total_step 40700, total loss is 27.67 , inference loss is 16.45, weight deacy loss is 11.22, training accuracy is 0.156250, time 56.449 samples/sec
+epoch 2, total_step 40720, total loss is 24.72 , inference loss is 13.50, weight deacy loss is 11.22, training accuracy is 0.187500, time 55.877 samples/sec
+epoch 2, total_step 40740, total loss is 24.40 , inference loss is 13.19, weight deacy loss is 11.22, training accuracy is 0.218750, time 56.340 samples/sec
+epoch 2, total_step 40760, total loss is 25.73 , inference loss is 14.51, weight deacy loss is 11.22, training accuracy is 0.218750, time 56.986 samples/sec
+epoch 2, total_step 40780, total loss is 25.00 , inference loss is 13.79, weight deacy loss is 11.22, training accuracy is 0.218750, time 56.162 samples/sec
+epoch 2, total_step 40800, total loss is 23.68 , inference loss is 12.47, weight deacy loss is 11.21, training accuracy is 0.250000, time 56.298 samples/sec
+epoch 2, total_step 40820, total loss is 28.87 , inference loss is 17.65, weight deacy loss is 11.21, training accuracy is 0.187500, time 55.847 samples/sec
+epoch 2, total_step 40840, total loss is 23.96 , inference loss is 12.74, weight deacy loss is 11.21, training accuracy is 0.312500, time 56.528 samples/sec
+epoch 2, total_step 40860, total loss is 24.68 , inference loss is 13.47, weight deacy loss is 11.21, training accuracy is 0.281250, time 56.323 samples/sec
+epoch 2, total_step 40880, total loss is 24.52 , inference loss is 13.31, weight deacy loss is 11.21, training accuracy is 0.187500, time 56.231 samples/sec
+epoch 2, total_step 40900, total loss is 23.72 , inference loss is 12.51, weight deacy loss is 11.21, training accuracy is 0.187500, time 54.867 samples/sec
+epoch 2, total_step 40920, total loss is 24.15 , inference loss is 12.95, weight deacy loss is 11.21, training accuracy is 0.218750, time 56.618 samples/sec
+epoch 2, total_step 40940, total loss is 25.33 , inference loss is 14.12, weight deacy loss is 11.20, training accuracy is 0.218750, time 56.934 samples/sec
+epoch 2, total_step 40960, total loss is 22.69 , inference loss is 11.49, weight deacy loss is 11.20, training accuracy is 0.187500, time 56.769 samples/sec
+epoch 2, total_step 40980, total loss is 22.17 , inference loss is 10.97, weight deacy loss is 11.20, training accuracy is 0.125000, time 55.821 samples/sec
+epoch 2, total_step 41000, total loss is 25.46 , inference loss is 14.26, weight deacy loss is 11.20, training accuracy is 0.218750, time 54.830 samples/sec
+epoch 2, total_step 41020, total loss is 27.48 , inference loss is 16.28, weight deacy loss is 11.20, training accuracy is 0.125000, time 55.621 samples/sec
+epoch 2, total_step 41040, total loss is 27.28 , inference loss is 16.08, weight deacy loss is 11.20, training accuracy is 0.156250, time 56.469 samples/sec
+epoch 2, total_step 41060, total loss is 23.41 , inference loss is 12.22, weight deacy loss is 11.19, training accuracy is 0.343750, time 56.608 samples/sec
+epoch 2, total_step 41080, total loss is 24.35 , inference loss is 13.15, weight deacy loss is 11.19, training accuracy is 0.062500, time 56.263 samples/sec
+epoch 2, total_step 41100, total loss is 28.72 , inference loss is 17.53, weight deacy loss is 11.19, training accuracy is 0.125000, time 56.079 samples/sec
+epoch 2, total_step 41120, total loss is 25.28 , inference loss is 14.09, weight deacy loss is 11.19, training accuracy is 0.250000, time 56.180 samples/sec
+epoch 2, total_step 41140, total loss is 22.62 , inference loss is 11.43, weight deacy loss is 11.19, training accuracy is 0.312500, time 56.516 samples/sec
+epoch 2, total_step 41160, total loss is 23.80 , inference loss is 12.61, weight deacy loss is 11.19, training accuracy is 0.281250, time 56.284 samples/sec
+epoch 2, total_step 41180, total loss is 28.56 , inference loss is 17.38, weight deacy loss is 11.19, training accuracy is 0.125000, time 56.072 samples/sec
+epoch 2, total_step 41200, total loss is 26.54 , inference loss is 15.36, weight deacy loss is 11.18, training accuracy is 0.187500, time 56.386 samples/sec
+epoch 2, total_step 41220, total loss is 27.20 , inference loss is 16.02, weight deacy loss is 11.18, training accuracy is 0.093750, time 56.042 samples/sec
+epoch 2, total_step 41240, total loss is 23.93 , inference loss is 12.75, weight deacy loss is 11.18, training accuracy is 0.281250, time 56.821 samples/sec
+epoch 2, total_step 41260, total loss is 26.17 , inference loss is 14.99, weight deacy loss is 11.18, training accuracy is 0.187500, time 56.354 samples/sec
+epoch 2, total_step 41280, total loss is 24.74 , inference loss is 13.56, weight deacy loss is 11.18, training accuracy is 0.125000, time 56.109 samples/sec
+epoch 2, total_step 41300, total loss is 27.17 , inference loss is 15.99, weight deacy loss is 11.18, training accuracy is 0.125000, time 56.077 samples/sec
+epoch 2, total_step 41320, total loss is 27.29 , inference loss is 16.11, weight deacy loss is 11.18, training accuracy is 0.218750, time 55.600 samples/sec
+epoch 2, total_step 41340, total loss is 26.64 , inference loss is 15.47, weight deacy loss is 11.17, training accuracy is 0.187500, time 56.323 samples/sec
+epoch 2, total_step 41360, total loss is 29.30 , inference loss is 18.12, weight deacy loss is 11.17, training accuracy is 0.062500, time 56.357 samples/sec
+epoch 2, total_step 41380, total loss is 28.95 , inference loss is 17.78, weight deacy loss is 11.17, training accuracy is 0.125000, time 55.480 samples/sec
+epoch 2, total_step 41400, total loss is 22.51 , inference loss is 11.34, weight deacy loss is 11.17, training accuracy is 0.187500, time 55.478 samples/sec
+epoch 2, total_step 41420, total loss is 24.20 , inference loss is 13.03, weight deacy loss is 11.17, training accuracy is 0.125000, time 55.198 samples/sec
+epoch 2, total_step 41440, total loss is 27.51 , inference loss is 16.34, weight deacy loss is 11.17, training accuracy is 0.156250, time 56.190 samples/sec
+epoch 2, total_step 41460, total loss is 26.75 , inference loss is 15.59, weight deacy loss is 11.17, training accuracy is 0.218750, time 56.553 samples/sec
+epoch 2, total_step 41480, total loss is 22.48 , inference loss is 11.32, weight deacy loss is 11.17, training accuracy is 0.250000, time 55.406 samples/sec
+epoch 2, total_step 41500, total loss is 26.95 , inference loss is 15.78, weight deacy loss is 11.16, training accuracy is 0.125000, time 55.557 samples/sec
+epoch 2, total_step 41520, total loss is 22.68 , inference loss is 11.51, weight deacy loss is 11.16, training accuracy is 0.125000, time 56.519 samples/sec
+epoch 2, total_step 41540, total loss is 27.13 , inference loss is 15.97, weight deacy loss is 11.16, training accuracy is 0.187500, time 55.597 samples/sec
+epoch 2, total_step 41560, total loss is 26.56 , inference loss is 15.40, weight deacy loss is 11.16, training accuracy is 0.062500, time 56.141 samples/sec
+epoch 2, total_step 41580, total loss is 29.28 , inference loss is 18.12, weight deacy loss is 11.16, training accuracy is 0.187500, time 55.856 samples/sec
+epoch 2, total_step 41600, total loss is 30.29 , inference loss is 19.13, weight deacy loss is 11.16, training accuracy is 0.125000, time 57.449 samples/sec
+epoch 2, total_step 41620, total loss is 24.98 , inference loss is 13.83, weight deacy loss is 11.16, training accuracy is 0.156250, time 55.649 samples/sec
+epoch 2, total_step 41640, total loss is 29.90 , inference loss is 18.75, weight deacy loss is 11.15, training accuracy is 0.125000, time 56.139 samples/sec
+epoch 2, total_step 41660, total loss is 29.37 , inference loss is 18.22, weight deacy loss is 11.15, training accuracy is 0.250000, time 56.066 samples/sec
+epoch 2, total_step 41680, total loss is 27.60 , inference loss is 16.45, weight deacy loss is 11.15, training accuracy is 0.218750, time 56.658 samples/sec
+epoch 2, total_step 41700, total loss is 23.85 , inference loss is 12.70, weight deacy loss is 11.15, training accuracy is 0.281250, time 55.643 samples/sec
+epoch 2, total_step 41720, total loss is 26.82 , inference loss is 15.68, weight deacy loss is 11.15, training accuracy is 0.062500, time 56.290 samples/sec
+epoch 2, total_step 41740, total loss is 25.07 , inference loss is 13.92, weight deacy loss is 11.15, training accuracy is 0.343750, time 56.560 samples/sec
+epoch 2, total_step 41760, total loss is 25.63 , inference loss is 14.48, weight deacy loss is 11.15, training accuracy is 0.156250, time 56.312 samples/sec
+epoch 2, total_step 41780, total loss is 27.33 , inference loss is 16.18, weight deacy loss is 11.14, training accuracy is 0.062500, time 56.693 samples/sec
+epoch 2, total_step 41800, total loss is 25.44 , inference loss is 14.30, weight deacy loss is 11.14, training accuracy is 0.125000, time 54.275 samples/sec
+epoch 2, total_step 41820, total loss is 22.74 , inference loss is 11.60, weight deacy loss is 11.14, training accuracy is 0.125000, time 56.634 samples/sec
+epoch 2, total_step 41840, total loss is 25.05 , inference loss is 13.91, weight deacy loss is 11.14, training accuracy is 0.156250, time 56.193 samples/sec
+epoch 2, total_step 41860, total loss is 25.97 , inference loss is 14.83, weight deacy loss is 11.14, training accuracy is 0.187500, time 56.775 samples/sec
+epoch 2, total_step 41880, total loss is 25.77 , inference loss is 14.63, weight deacy loss is 11.14, training accuracy is 0.125000, time 56.080 samples/sec
+epoch 2, total_step 41900, total loss is 26.91 , inference loss is 15.78, weight deacy loss is 11.14, training accuracy is 0.093750, time 53.008 samples/sec
+epoch 2, total_step 41920, total loss is 29.38 , inference loss is 18.24, weight deacy loss is 11.13, training accuracy is 0.062500, time 56.327 samples/sec
+epoch 2, total_step 41940, total loss is 24.91 , inference loss is 13.78, weight deacy loss is 11.13, training accuracy is 0.093750, time 56.420 samples/sec
+epoch 2, total_step 41960, total loss is 23.98 , inference loss is 12.85, weight deacy loss is 11.13, training accuracy is 0.125000, time 55.901 samples/sec
+epoch 2, total_step 41980, total loss is 29.98 , inference loss is 18.85, weight deacy loss is 11.13, training accuracy is 0.156250, time 56.308 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.64420099999988
+best_threshold_index 104 0.9742592592592593
+best_threshold_index 106 0.975
+best_threshold_index 104 0.975925925925926
+best_threshold_index 100 0.9762962962962963
+best_threshold_index 100 0.9748148148148148
+best_threshold_index 104 0.9753703703703703
+best_threshold_index 103 0.975
+best_threshold_index 104 0.9740740740740741
+best_threshold_index 111 0.9737037037037037
+best_threshold_index 104 0.9742592592592593
+[lfw][42000]XNorm: 23.241452
+[lfw][42000]Accuracy-Flip: 0.97300+-0.00741
+testing verification..
+(14000, 512)
+infer time 86.164023
+best_threshold_index 112 0.9752380952380952
+best_threshold_index 112 0.9746031746031746
+best_threshold_index 112 0.9752380952380952
+best_threshold_index 112 0.9741269841269842
+best_threshold_index 112 0.9742857142857143
+best_threshold_index 112 0.9731746031746031
+best_threshold_index 112 0.9738095238095238
+best_threshold_index 112 0.9736507936507937
+best_threshold_index 112 0.9747619047619047
+best_threshold_index 112 0.9753968253968254
+[cfp_ff][42000]XNorm: 22.697513
+[cfp_ff][42000]Accuracy-Flip: 0.97443+-0.00640
+testing verification..
+(14000, 512)
+infer time 85.79626300000008
+best_threshold_index 132 0.8742857142857143
+best_threshold_index 132 0.8720634920634921
+best_threshold_index 132 0.871904761904762
+best_threshold_index 131 0.8738095238095238
+best_threshold_index 132 0.8746031746031746
+best_threshold_index 132 0.8711111111111111
+best_threshold_index 132 0.8711111111111111
+best_threshold_index 131 0.8706349206349207
+best_threshold_index 132 0.8728571428571429
+best_threshold_index 132 0.8736507936507937
+[cfp_fp][42000]XNorm: 22.709831
+[cfp_fp][42000]Accuracy-Flip: 0.87171+-0.01215
+testing verification..
+(12000, 512)
+infer time 73.66946099999996
+best_threshold_index 136 0.8624074074074074
+best_threshold_index 136 0.8631481481481481
+best_threshold_index 134 0.8655555555555555
+best_threshold_index 136 0.8655555555555555
+best_threshold_index 136 0.87
+best_threshold_index 136 0.8635185185185185
+best_threshold_index 135 0.8662962962962963
+best_threshold_index 136 0.8711111111111111
+best_threshold_index 136 0.8681481481481481
+best_threshold_index 136 0.8677777777777778
+[agedb_30][42000]XNorm: 22.615868
+[agedb_30][42000]Accuracy-Flip: 0.86467+-0.02480
+test accuracy is: 0.9730000000000001
+epoch 2, total_step 42000, total loss is 24.64 , inference loss is 13.52, weight deacy loss is 11.13, training accuracy is 0.218750, time 53.246 samples/sec
+epoch 2, total_step 42020, total loss is 25.30 , inference loss is 14.17, weight deacy loss is 11.13, training accuracy is 0.093750, time 55.610 samples/sec
+epoch 2, total_step 42040, total loss is 27.29 , inference loss is 16.17, weight deacy loss is 11.13, training accuracy is 0.125000, time 56.040 samples/sec
+epoch 2, total_step 42060, total loss is 29.13 , inference loss is 18.00, weight deacy loss is 11.12, training accuracy is 0.031250, time 56.550 samples/sec
+epoch 2, total_step 42080, total loss is 22.95 , inference loss is 11.83, weight deacy loss is 11.12, training accuracy is 0.250000, time 56.756 samples/sec
+epoch 2, total_step 42100, total loss is 24.84 , inference loss is 13.72, weight deacy loss is 11.12, training accuracy is 0.187500, time 56.170 samples/sec
+epoch 2, total_step 42120, total loss is 24.31 , inference loss is 13.19, weight deacy loss is 11.12, training accuracy is 0.312500, time 56.049 samples/sec
+epoch 2, total_step 42140, total loss is 25.51 , inference loss is 14.39, weight deacy loss is 11.12, training accuracy is 0.218750, time 55.704 samples/sec
+epoch 2, total_step 42160, total loss is 30.23 , inference loss is 19.11, weight deacy loss is 11.12, training accuracy is 0.062500, time 56.223 samples/sec
+epoch 2, total_step 42180, total loss is 31.61 , inference loss is 20.50, weight deacy loss is 11.12, training accuracy is 0.062500, time 56.695 samples/sec
+epoch 2, total_step 42200, total loss is 26.23 , inference loss is 15.12, weight deacy loss is 11.11, training accuracy is 0.218750, time 54.387 samples/sec
+epoch 2, total_step 42220, total loss is 26.67 , inference loss is 15.55, weight deacy loss is 11.11, training accuracy is 0.062500, time 56.416 samples/sec
+epoch 2, total_step 42240, total loss is 31.60 , inference loss is 20.49, weight deacy loss is 11.11, training accuracy is 0.062500, time 56.249 samples/sec
+epoch 2, total_step 42260, total loss is 25.61 , inference loss is 14.50, weight deacy loss is 11.11, training accuracy is 0.093750, time 55.757 samples/sec
+epoch 2, total_step 42280, total loss is 27.62 , inference loss is 16.51, weight deacy loss is 11.11, training accuracy is 0.125000, time 56.369 samples/sec
+epoch 2, total_step 42300, total loss is 27.62 , inference loss is 16.51, weight deacy loss is 11.11, training accuracy is 0.031250, time 55.705 samples/sec
+epoch 2, total_step 42320, total loss is 29.26 , inference loss is 18.16, weight deacy loss is 11.11, training accuracy is 0.125000, time 56.532 samples/sec
+epoch 2, total_step 42340, total loss is 27.10 , inference loss is 15.99, weight deacy loss is 11.10, training accuracy is 0.156250, time 54.111 samples/sec
+epoch 2, total_step 42360, total loss is 24.48 , inference loss is 13.37, weight deacy loss is 11.10, training accuracy is 0.218750, time 56.693 samples/sec
+epoch 2, total_step 42380, total loss is 28.02 , inference loss is 16.92, weight deacy loss is 11.10, training accuracy is 0.187500, time 56.587 samples/sec
+epoch 2, total_step 42400, total loss is 21.40 , inference loss is 10.30, weight deacy loss is 11.10, training accuracy is 0.250000, time 55.923 samples/sec
+epoch 2, total_step 42420, total loss is 26.61 , inference loss is 15.51, weight deacy loss is 11.10, training accuracy is 0.125000, time 55.797 samples/sec
+epoch 2, total_step 42440, total loss is 25.78 , inference loss is 14.69, weight deacy loss is 11.10, training accuracy is 0.093750, time 55.999 samples/sec
+epoch 2, total_step 42460, total loss is 25.16 , inference loss is 14.06, weight deacy loss is 11.10, training accuracy is 0.062500, time 55.372 samples/sec
+epoch 2, total_step 42480, total loss is 30.37 , inference loss is 19.27, weight deacy loss is 11.09, training accuracy is 0.062500, time 55.603 samples/sec
+epoch 2, total_step 42500, total loss is 28.66 , inference loss is 17.57, weight deacy loss is 11.09, training accuracy is 0.062500, time 56.627 samples/sec
+epoch 2, total_step 42520, total loss is 24.42 , inference loss is 13.32, weight deacy loss is 11.09, training accuracy is 0.187500, time 55.170 samples/sec
+epoch 2, total_step 42540, total loss is 25.68 , inference loss is 14.59, weight deacy loss is 11.09, training accuracy is 0.125000, time 55.274 samples/sec
+epoch 2, total_step 42560, total loss is 27.31 , inference loss is 16.22, weight deacy loss is 11.09, training accuracy is 0.218750, time 56.451 samples/sec
+epoch 2, total_step 42580, total loss is 26.68 , inference loss is 15.59, weight deacy loss is 11.09, training accuracy is 0.093750, time 53.219 samples/sec
+epoch 2, total_step 42600, total loss is 27.65 , inference loss is 16.57, weight deacy loss is 11.09, training accuracy is 0.125000, time 56.543 samples/sec
+epoch 2, total_step 42620, total loss is 25.74 , inference loss is 14.66, weight deacy loss is 11.08, training accuracy is 0.187500, time 55.688 samples/sec
+epoch 2, total_step 42640, total loss is 26.83 , inference loss is 15.75, weight deacy loss is 11.08, training accuracy is 0.187500, time 55.933 samples/sec
+epoch 2, total_step 42660, total loss is 30.62 , inference loss is 19.54, weight deacy loss is 11.08, training accuracy is 0.031250, time 54.804 samples/sec
+epoch 2, total_step 42680, total loss is 25.68 , inference loss is 14.60, weight deacy loss is 11.08, training accuracy is 0.125000, time 56.152 samples/sec
+epoch 2, total_step 42700, total loss is 24.23 , inference loss is 13.15, weight deacy loss is 11.08, training accuracy is 0.187500, time 55.786 samples/sec
+epoch 2, total_step 42720, total loss is 22.05 , inference loss is 10.97, weight deacy loss is 11.08, training accuracy is 0.218750, time 56.153 samples/sec
+epoch 2, total_step 42740, total loss is 31.22 , inference loss is 20.15, weight deacy loss is 11.08, training accuracy is 0.093750, time 55.896 samples/sec
+epoch 2, total_step 42760, total loss is 28.92 , inference loss is 17.85, weight deacy loss is 11.07, training accuracy is 0.031250, time 55.724 samples/sec
+epoch 2, total_step 42780, total loss is 27.64 , inference loss is 16.56, weight deacy loss is 11.07, training accuracy is 0.156250, time 55.970 samples/sec
+epoch 2, total_step 42800, total loss is 30.91 , inference loss is 19.84, weight deacy loss is 11.07, training accuracy is 0.093750, time 56.312 samples/sec
+epoch 2, total_step 42820, total loss is 25.23 , inference loss is 14.16, weight deacy loss is 11.07, training accuracy is 0.093750, time 55.501 samples/sec
+epoch 2, total_step 42840, total loss is 25.61 , inference loss is 14.54, weight deacy loss is 11.07, training accuracy is 0.031250, time 56.227 samples/sec
+epoch 2, total_step 42860, total loss is 24.87 , inference loss is 13.80, weight deacy loss is 11.07, training accuracy is 0.093750, time 55.826 samples/sec
+epoch 2, total_step 42880, total loss is 25.11 , inference loss is 14.04, weight deacy loss is 11.07, training accuracy is 0.156250, time 56.755 samples/sec
+epoch 2, total_step 42900, total loss is 28.66 , inference loss is 17.60, weight deacy loss is 11.07, training accuracy is 0.125000, time 55.953 samples/sec
+epoch 2, total_step 42920, total loss is 27.68 , inference loss is 16.61, weight deacy loss is 11.06, training accuracy is 0.125000, time 55.749 samples/sec
+epoch 2, total_step 42940, total loss is 25.38 , inference loss is 14.32, weight deacy loss is 11.06, training accuracy is 0.093750, time 53.083 samples/sec
+epoch 2, total_step 42960, total loss is 28.69 , inference loss is 17.63, weight deacy loss is 11.06, training accuracy is 0.031250, time 55.666 samples/sec
+epoch 2, total_step 42980, total loss is 29.27 , inference loss is 18.21, weight deacy loss is 11.06, training accuracy is 0.093750, time 55.988 samples/sec
+epoch 2, total_step 43000, total loss is 24.71 , inference loss is 13.66, weight deacy loss is 11.06, training accuracy is 0.093750, time 56.678 samples/sec
+epoch 2, total_step 43020, total loss is 24.09 , inference loss is 13.03, weight deacy loss is 11.06, training accuracy is 0.062500, time 56.383 samples/sec
+epoch 2, total_step 43040, total loss is 24.54 , inference loss is 13.48, weight deacy loss is 11.06, training accuracy is 0.093750, time 56.336 samples/sec
+epoch 2, total_step 43060, total loss is 24.51 , inference loss is 13.46, weight deacy loss is 11.05, training accuracy is 0.187500, time 56.243 samples/sec
+epoch 2, total_step 43080, total loss is 26.91 , inference loss is 15.86, weight deacy loss is 11.05, training accuracy is 0.062500, time 55.850 samples/sec
+epoch 2, total_step 43100, total loss is 26.87 , inference loss is 15.82, weight deacy loss is 11.05, training accuracy is 0.156250, time 56.406 samples/sec
+epoch 2, total_step 43120, total loss is 28.44 , inference loss is 17.39, weight deacy loss is 11.05, training accuracy is 0.031250, time 54.584 samples/sec
+epoch 2, total_step 43140, total loss is 25.47 , inference loss is 14.43, weight deacy loss is 11.05, training accuracy is 0.156250, time 56.180 samples/sec
+epoch 2, total_step 43160, total loss is 25.74 , inference loss is 14.69, weight deacy loss is 11.05, training accuracy is 0.093750, time 56.502 samples/sec
+epoch 2, total_step 43180, total loss is 27.61 , inference loss is 16.56, weight deacy loss is 11.05, training accuracy is 0.125000, time 55.910 samples/sec
+epoch 2, total_step 43200, total loss is 27.94 , inference loss is 16.90, weight deacy loss is 11.04, training accuracy is 0.125000, time 54.378 samples/sec
+epoch 2, total_step 43220, total loss is 23.49 , inference loss is 12.45, weight deacy loss is 11.04, training accuracy is 0.125000, time 56.279 samples/sec
+epoch 2, total_step 43240, total loss is 26.10 , inference loss is 15.06, weight deacy loss is 11.04, training accuracy is 0.156250, time 55.678 samples/sec
+epoch 2, total_step 43260, total loss is 32.05 , inference loss is 21.01, weight deacy loss is 11.04, training accuracy is 0.156250, time 56.295 samples/sec
+epoch 2, total_step 43280, total loss is 29.76 , inference loss is 18.73, weight deacy loss is 11.04, training accuracy is 0.062500, time 55.879 samples/sec
+epoch 2, total_step 43300, total loss is 26.70 , inference loss is 15.67, weight deacy loss is 11.04, training accuracy is 0.093750, time 56.934 samples/sec
+epoch 2, total_step 43320, total loss is 29.43 , inference loss is 18.39, weight deacy loss is 11.04, training accuracy is 0.125000, time 55.118 samples/sec
+epoch 2, total_step 43340, total loss is 24.00 , inference loss is 12.96, weight deacy loss is 11.03, training accuracy is 0.125000, time 56.208 samples/sec
+epoch 2, total_step 43360, total loss is 31.48 , inference loss is 20.44, weight deacy loss is 11.03, training accuracy is 0.125000, time 55.541 samples/sec
+epoch 2, total_step 43380, total loss is 25.92 , inference loss is 14.89, weight deacy loss is 11.03, training accuracy is 0.062500, time 56.017 samples/sec
+epoch 2, total_step 43400, total loss is 26.04 , inference loss is 15.01, weight deacy loss is 11.03, training accuracy is 0.062500, time 56.538 samples/sec
+epoch 2, total_step 43420, total loss is 24.56 , inference loss is 13.53, weight deacy loss is 11.03, training accuracy is 0.125000, time 56.644 samples/sec
+epoch 2, total_step 43440, total loss is 31.55 , inference loss is 20.52, weight deacy loss is 11.03, training accuracy is 0.125000, time 56.420 samples/sec
+epoch 2, total_step 43460, total loss is 26.34 , inference loss is 15.31, weight deacy loss is 11.03, training accuracy is 0.156250, time 56.167 samples/sec
+epoch 2, total_step 43480, total loss is 23.79 , inference loss is 12.76, weight deacy loss is 11.02, training accuracy is 0.187500, time 55.340 samples/sec
+epoch 2, total_step 43500, total loss is 26.79 , inference loss is 15.77, weight deacy loss is 11.02, training accuracy is 0.031250, time 55.679 samples/sec
+epoch 2, total_step 43520, total loss is 24.03 , inference loss is 13.01, weight deacy loss is 11.02, training accuracy is 0.125000, time 55.595 samples/sec
+epoch 2, total_step 43540, total loss is 28.83 , inference loss is 17.81, weight deacy loss is 11.02, training accuracy is 0.093750, time 57.712 samples/sec
+epoch 2, total_step 43560, total loss is 23.91 , inference loss is 12.89, weight deacy loss is 11.02, training accuracy is 0.125000, time 56.908 samples/sec
+epoch 2, total_step 43580, total loss is 26.02 , inference loss is 15.00, weight deacy loss is 11.02, training accuracy is 0.156250, time 56.575 samples/sec
+epoch 2, total_step 43600, total loss is 28.12 , inference loss is 17.10, weight deacy loss is 11.02, training accuracy is 0.156250, time 56.387 samples/sec
+epoch 2, total_step 43620, total loss is 27.25 , inference loss is 16.23, weight deacy loss is 11.02, training accuracy is 0.156250, time 56.399 samples/sec
+epoch 2, total_step 43640, total loss is 26.38 , inference loss is 15.37, weight deacy loss is 11.01, training accuracy is 0.156250, time 52.892 samples/sec
+epoch 2, total_step 43660, total loss is 28.30 , inference loss is 17.28, weight deacy loss is 11.01, training accuracy is 0.031250, time 56.265 samples/sec
+epoch 2, total_step 43680, total loss is 30.35 , inference loss is 19.33, weight deacy loss is 11.01, training accuracy is 0.125000, time 55.862 samples/sec
+epoch 2, total_step 43700, total loss is 27.01 , inference loss is 16.00, weight deacy loss is 11.01, training accuracy is 0.156250, time 55.790 samples/sec
+epoch 2, total_step 43720, total loss is 24.81 , inference loss is 13.80, weight deacy loss is 11.01, training accuracy is 0.093750, time 55.871 samples/sec
+epoch 2, total_step 43740, total loss is 27.28 , inference loss is 16.27, weight deacy loss is 11.01, training accuracy is 0.031250, time 55.813 samples/sec
+epoch 2, total_step 43760, total loss is 26.16 , inference loss is 15.15, weight deacy loss is 11.01, training accuracy is 0.156250, time 56.745 samples/sec
+epoch 2, total_step 43780, total loss is 31.69 , inference loss is 20.68, weight deacy loss is 11.00, training accuracy is 0.031250, time 56.109 samples/sec
+epoch 2, total_step 43800, total loss is 24.98 , inference loss is 13.97, weight deacy loss is 11.00, training accuracy is 0.093750, time 54.763 samples/sec
+epoch 2, total_step 43820, total loss is 29.92 , inference loss is 18.92, weight deacy loss is 11.00, training accuracy is 0.062500, time 55.819 samples/sec
+epoch 2, total_step 43840, total loss is 26.83 , inference loss is 15.82, weight deacy loss is 11.00, training accuracy is 0.031250, time 57.380 samples/sec
+epoch 2, total_step 43860, total loss is 26.14 , inference loss is 15.14, weight deacy loss is 11.00, training accuracy is 0.062500, time 56.593 samples/sec
+epoch 2, total_step 43880, total loss is 23.78 , inference loss is 12.78, weight deacy loss is 11.00, training accuracy is 0.125000, time 54.276 samples/sec
+epoch 2, total_step 43900, total loss is 25.36 , inference loss is 14.37, weight deacy loss is 11.00, training accuracy is 0.093750, time 55.452 samples/sec
+epoch 2, total_step 43920, total loss is 28.70 , inference loss is 17.70, weight deacy loss is 10.99, training accuracy is 0.093750, time 53.508 samples/sec
+epoch 2, total_step 43940, total loss is 25.20 , inference loss is 14.20, weight deacy loss is 10.99, training accuracy is 0.093750, time 55.644 samples/sec
+epoch 2, total_step 43960, total loss is 25.83 , inference loss is 14.84, weight deacy loss is 10.99, training accuracy is 0.187500, time 56.031 samples/sec
+epoch 2, total_step 43980, total loss is 26.30 , inference loss is 15.31, weight deacy loss is 10.99, training accuracy is 0.125000, time 56.008 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.80726200000007
+best_threshold_index 108 0.9783333333333334
+best_threshold_index 108 0.9777777777777777
+best_threshold_index 108 0.9796296296296296
+best_threshold_index 108 0.9807407407407407
+best_threshold_index 109 0.9783333333333334
+best_threshold_index 108 0.9775925925925926
+best_threshold_index 108 0.9777777777777777
+best_threshold_index 108 0.9774074074074074
+best_threshold_index 109 0.9774074074074074
+best_threshold_index 108 0.977037037037037
+[lfw][44000]XNorm: 23.472826
+[lfw][44000]Accuracy-Flip: 0.97783+-0.00981
+testing verification..
+(14000, 512)
+infer time 85.77983099999993
+best_threshold_index 116 0.9777777777777777
+best_threshold_index 116 0.9768253968253968
+best_threshold_index 113 0.9776190476190476
+best_threshold_index 116 0.9755555555555555
+best_threshold_index 116 0.9763492063492063
+best_threshold_index 116 0.9758730158730159
+best_threshold_index 116 0.976984126984127
+best_threshold_index 116 0.974920634920635
+best_threshold_index 116 0.976984126984127
+best_threshold_index 116 0.976984126984127
+[cfp_ff][44000]XNorm: 22.772775
+[cfp_ff][44000]Accuracy-Flip: 0.97586+-0.00858
+testing verification..
+(14000, 512)
+infer time 85.98138699999998
+best_threshold_index 135 0.8823809523809524
+best_threshold_index 134 0.8812698412698413
+best_threshold_index 134 0.8796825396825397
+best_threshold_index 134 0.8815873015873016
+best_threshold_index 134 0.8838095238095238
+best_threshold_index 135 0.88
+best_threshold_index 134 0.8776190476190476
+best_threshold_index 135 0.8796825396825397
+best_threshold_index 134 0.8812698412698413
+best_threshold_index 134 0.8820634920634921
+[cfp_fp][44000]XNorm: 22.097268
+[cfp_fp][44000]Accuracy-Flip: 0.87971+-0.01482
+testing verification..
+(12000, 512)
+infer time 73.50273700000001
+best_threshold_index 137 0.8690740740740741
+best_threshold_index 137 0.8659259259259259
+best_threshold_index 136 0.8694444444444445
+best_threshold_index 137 0.87
+best_threshold_index 136 0.8751851851851852
+best_threshold_index 136 0.8690740740740741
+best_threshold_index 136 0.872037037037037
+best_threshold_index 136 0.8735185185185185
+best_threshold_index 139 0.872037037037037
+best_threshold_index 137 0.8707407407407407
+[agedb_30][44000]XNorm: 22.973047
+[agedb_30][44000]Accuracy-Flip: 0.86783+-0.02410
+test accuracy is: 0.9778333333333332
+epoch 2, total_step 44000, total loss is 27.99 , inference loss is 17.00, weight deacy loss is 10.99, training accuracy is 0.093750, time 55.419 samples/sec
+epoch 2, total_step 44020, total loss is 28.24 , inference loss is 17.26, weight deacy loss is 10.99, training accuracy is 0.156250, time 55.857 samples/sec
+epoch 2, total_step 44040, total loss is 29.33 , inference loss is 18.34, weight deacy loss is 10.99, training accuracy is 0.062500, time 56.140 samples/sec
+epoch 2, total_step 44060, total loss is 26.50 , inference loss is 15.52, weight deacy loss is 10.98, training accuracy is 0.031250, time 55.613 samples/sec
+epoch 2, total_step 44080, total loss is 25.17 , inference loss is 14.19, weight deacy loss is 10.98, training accuracy is 0.093750, time 56.503 samples/sec
+epoch 2, total_step 44100, total loss is 23.63 , inference loss is 12.64, weight deacy loss is 10.98, training accuracy is 0.156250, time 56.262 samples/sec
+epoch 2, total_step 44120, total loss is 27.81 , inference loss is 16.83, weight deacy loss is 10.98, training accuracy is 0.125000, time 56.462 samples/sec
+epoch 2, total_step 44140, total loss is 30.06 , inference loss is 19.08, weight deacy loss is 10.98, training accuracy is 0.093750, time 56.146 samples/sec
+epoch 2, total_step 44160, total loss is 26.22 , inference loss is 15.24, weight deacy loss is 10.98, training accuracy is 0.093750, time 55.712 samples/sec
+epoch 2, total_step 44180, total loss is 26.00 , inference loss is 15.02, weight deacy loss is 10.98, training accuracy is 0.156250, time 56.454 samples/sec
+epoch 2, total_step 44200, total loss is 28.19 , inference loss is 17.22, weight deacy loss is 10.98, training accuracy is 0.093750, time 55.395 samples/sec
+epoch 2, total_step 44220, total loss is 24.71 , inference loss is 13.74, weight deacy loss is 10.97, training accuracy is 0.156250, time 56.383 samples/sec
+epoch 2, total_step 44240, total loss is 27.14 , inference loss is 16.16, weight deacy loss is 10.97, training accuracy is 0.093750, time 56.196 samples/sec
+epoch 2, total_step 44260, total loss is 31.92 , inference loss is 20.94, weight deacy loss is 10.97, training accuracy is 0.031250, time 55.580 samples/sec
+epoch 2, total_step 44280, total loss is 27.27 , inference loss is 16.30, weight deacy loss is 10.97, training accuracy is 0.093750, time 56.615 samples/sec
+epoch 2, total_step 44300, total loss is 25.43 , inference loss is 14.46, weight deacy loss is 10.97, training accuracy is 0.156250, time 56.760 samples/sec
+epoch 2, total_step 44320, total loss is 26.65 , inference loss is 15.68, weight deacy loss is 10.97, training accuracy is 0.156250, time 56.004 samples/sec
+epoch 2, total_step 44340, total loss is 25.69 , inference loss is 14.73, weight deacy loss is 10.97, training accuracy is 0.218750, time 55.836 samples/sec
+epoch 2, total_step 44360, total loss is 24.66 , inference loss is 13.69, weight deacy loss is 10.96, training accuracy is 0.187500, time 56.815 samples/sec
+epoch 2, total_step 44380, total loss is 30.78 , inference loss is 19.82, weight deacy loss is 10.96, training accuracy is 0.062500, time 55.746 samples/sec
+epoch 2, total_step 44400, total loss is 23.45 , inference loss is 12.48, weight deacy loss is 10.96, training accuracy is 0.187500, time 55.994 samples/sec
+epoch 2, total_step 44420, total loss is 23.27 , inference loss is 12.31, weight deacy loss is 10.96, training accuracy is 0.250000, time 55.684 samples/sec
+epoch 2, total_step 44440, total loss is 24.46 , inference loss is 13.51, weight deacy loss is 10.96, training accuracy is 0.093750, time 56.994 samples/sec
+epoch 2, total_step 44460, total loss is 26.27 , inference loss is 15.31, weight deacy loss is 10.96, training accuracy is 0.250000, time 55.625 samples/sec
+epoch 2, total_step 44480, total loss is 22.65 , inference loss is 11.69, weight deacy loss is 10.96, training accuracy is 0.281250, time 56.179 samples/sec
+epoch 2, total_step 44500, total loss is 23.84 , inference loss is 12.88, weight deacy loss is 10.95, training accuracy is 0.187500, time 56.452 samples/sec
+epoch 2, total_step 44520, total loss is 26.55 , inference loss is 15.60, weight deacy loss is 10.95, training accuracy is 0.156250, time 56.221 samples/sec
+epoch 2, total_step 44540, total loss is 22.78 , inference loss is 11.83, weight deacy loss is 10.95, training accuracy is 0.218750, time 56.692 samples/sec
+epoch 2, total_step 44560, total loss is 29.75 , inference loss is 18.80, weight deacy loss is 10.95, training accuracy is 0.031250, time 56.203 samples/sec
+epoch 2, total_step 44580, total loss is 27.18 , inference loss is 16.23, weight deacy loss is 10.95, training accuracy is 0.062500, time 56.801 samples/sec
+epoch 2, total_step 44600, total loss is 23.86 , inference loss is 12.91, weight deacy loss is 10.95, training accuracy is 0.062500, time 55.830 samples/sec
+epoch 2, total_step 44620, total loss is 29.35 , inference loss is 18.40, weight deacy loss is 10.95, training accuracy is 0.031250, time 56.413 samples/sec
+epoch 2, total_step 44640, total loss is 28.04 , inference loss is 17.09, weight deacy loss is 10.95, training accuracy is 0.062500, time 55.675 samples/sec
+epoch 2, total_step 44660, total loss is 27.15 , inference loss is 16.20, weight deacy loss is 10.94, training accuracy is 0.093750, time 55.781 samples/sec
+epoch 2, total_step 44680, total loss is 30.44 , inference loss is 19.50, weight deacy loss is 10.94, training accuracy is 0.156250, time 56.329 samples/sec
+epoch 2, total_step 44700, total loss is 26.29 , inference loss is 15.35, weight deacy loss is 10.94, training accuracy is 0.093750, time 56.876 samples/sec
+epoch 2, total_step 44720, total loss is 25.20 , inference loss is 14.26, weight deacy loss is 10.94, training accuracy is 0.125000, time 55.534 samples/sec
+epoch 2, total_step 44740, total loss is 26.97 , inference loss is 16.03, weight deacy loss is 10.94, training accuracy is 0.031250, time 56.341 samples/sec
+epoch 2, total_step 44760, total loss is 22.64 , inference loss is 11.71, weight deacy loss is 10.94, training accuracy is 0.250000, time 54.957 samples/sec
+epoch 2, total_step 44780, total loss is 25.95 , inference loss is 15.02, weight deacy loss is 10.94, training accuracy is 0.125000, time 56.386 samples/sec
+epoch 2, total_step 44800, total loss is 28.42 , inference loss is 17.49, weight deacy loss is 10.93, training accuracy is 0.125000, time 56.371 samples/sec
+epoch 2, total_step 44820, total loss is 25.63 , inference loss is 14.70, weight deacy loss is 10.93, training accuracy is 0.156250, time 57.250 samples/sec
+epoch 2, total_step 44840, total loss is 28.42 , inference loss is 17.49, weight deacy loss is 10.93, training accuracy is 0.093750, time 54.931 samples/sec
+epoch 2, total_step 44860, total loss is 28.97 , inference loss is 18.04, weight deacy loss is 10.93, training accuracy is 0.125000, time 56.443 samples/sec
+epoch 2, total_step 44880, total loss is 32.79 , inference loss is 21.86, weight deacy loss is 10.93, training accuracy is 0.093750, time 55.285 samples/sec
+epoch 2, total_step 44900, total loss is 24.11 , inference loss is 13.18, weight deacy loss is 10.93, training accuracy is 0.156250, time 55.885 samples/sec
+epoch 2, total_step 44920, total loss is 29.39 , inference loss is 18.47, weight deacy loss is 10.93, training accuracy is 0.031250, time 56.597 samples/sec
+epoch 2, total_step 44940, total loss is 26.64 , inference loss is 15.71, weight deacy loss is 10.92, training accuracy is 0.062500, time 56.182 samples/sec
+epoch 2, total_step 44960, total loss is 28.87 , inference loss is 17.95, weight deacy loss is 10.92, training accuracy is 0.062500, time 56.379 samples/sec
+epoch 2, total_step 44980, total loss is 29.01 , inference loss is 18.08, weight deacy loss is 10.92, training accuracy is 0.000000, time 53.292 samples/sec
+epoch 2, total_step 45000, total loss is 28.82 , inference loss is 17.90, weight deacy loss is 10.92, training accuracy is 0.062500, time 55.659 samples/sec
+epoch 2, total_step 45020, total loss is 22.76 , inference loss is 11.84, weight deacy loss is 10.92, training accuracy is 0.125000, time 56.061 samples/sec
+epoch 2, total_step 45040, total loss is 27.24 , inference loss is 16.32, weight deacy loss is 10.92, training accuracy is 0.062500, time 56.077 samples/sec
+epoch 2, total_step 45060, total loss is 28.06 , inference loss is 17.14, weight deacy loss is 10.92, training accuracy is 0.093750, time 55.489 samples/sec
+epoch 2, total_step 45080, total loss is 24.67 , inference loss is 13.75, weight deacy loss is 10.92, training accuracy is 0.156250, time 55.450 samples/sec
+epoch 2, total_step 45100, total loss is 28.48 , inference loss is 17.56, weight deacy loss is 10.91, training accuracy is 0.093750, time 54.568 samples/sec
+epoch 2, total_step 45120, total loss is 29.26 , inference loss is 18.35, weight deacy loss is 10.91, training accuracy is 0.062500, time 56.226 samples/sec
+epoch 2, total_step 45140, total loss is 31.14 , inference loss is 20.23, weight deacy loss is 10.91, training accuracy is 0.093750, time 53.117 samples/sec
+epoch 2, total_step 45160, total loss is 26.46 , inference loss is 15.55, weight deacy loss is 10.91, training accuracy is 0.093750, time 56.282 samples/sec
+epoch 2, total_step 45180, total loss is 22.14 , inference loss is 11.23, weight deacy loss is 10.91, training accuracy is 0.156250, time 56.185 samples/sec
+epoch 2, total_step 45200, total loss is 24.31 , inference loss is 13.40, weight deacy loss is 10.91, training accuracy is 0.187500, time 56.275 samples/sec
+epoch 2, total_step 45220, total loss is 27.77 , inference loss is 16.86, weight deacy loss is 10.91, training accuracy is 0.062500, time 55.812 samples/sec
+epoch 2, total_step 45240, total loss is 27.77 , inference loss is 16.86, weight deacy loss is 10.90, training accuracy is 0.125000, time 56.681 samples/sec
+epoch 2, total_step 45260, total loss is 25.05 , inference loss is 14.15, weight deacy loss is 10.90, training accuracy is 0.093750, time 56.283 samples/sec
+epoch 2, total_step 45280, total loss is 27.61 , inference loss is 16.71, weight deacy loss is 10.90, training accuracy is 0.125000, time 56.172 samples/sec
+epoch 2, total_step 45300, total loss is 30.15 , inference loss is 19.25, weight deacy loss is 10.90, training accuracy is 0.031250, time 57.113 samples/sec
+epoch 2, total_step 45320, total loss is 28.01 , inference loss is 17.11, weight deacy loss is 10.90, training accuracy is 0.093750, time 56.462 samples/sec
+epoch 2, total_step 45340, total loss is 23.30 , inference loss is 12.40, weight deacy loss is 10.90, training accuracy is 0.093750, time 55.667 samples/sec
+epoch 2, total_step 45360, total loss is 30.44 , inference loss is 19.55, weight deacy loss is 10.90, training accuracy is 0.125000, time 56.368 samples/sec
+epoch 2, total_step 45380, total loss is 23.56 , inference loss is 12.66, weight deacy loss is 10.90, training accuracy is 0.125000, time 55.967 samples/sec
+epoch 2, total_step 45400, total loss is 30.15 , inference loss is 19.26, weight deacy loss is 10.89, training accuracy is 0.031250, time 56.195 samples/sec
+epoch 2, total_step 45420, total loss is 31.96 , inference loss is 21.07, weight deacy loss is 10.89, training accuracy is 0.031250, time 56.007 samples/sec
+epoch 2, total_step 45440, total loss is 25.75 , inference loss is 14.86, weight deacy loss is 10.89, training accuracy is 0.125000, time 56.537 samples/sec
+epoch 2, total_step 45460, total loss is 26.99 , inference loss is 16.10, weight deacy loss is 10.89, training accuracy is 0.093750, time 56.169 samples/sec
+epoch 2, total_step 45480, total loss is 30.01 , inference loss is 19.12, weight deacy loss is 10.89, training accuracy is 0.031250, time 55.781 samples/sec
+epoch 2, total_step 45500, total loss is 27.27 , inference loss is 16.38, weight deacy loss is 10.89, training accuracy is 0.000000, time 56.415 samples/sec
+epoch 2, total_step 45520, total loss is 26.50 , inference loss is 15.62, weight deacy loss is 10.89, training accuracy is 0.031250, time 56.389 samples/sec
+epoch 2, total_step 45540, total loss is 25.47 , inference loss is 14.59, weight deacy loss is 10.88, training accuracy is 0.156250, time 55.700 samples/sec
+epoch 2, total_step 45560, total loss is 28.91 , inference loss is 18.03, weight deacy loss is 10.88, training accuracy is 0.062500, time 55.124 samples/sec
+epoch 2, total_step 45580, total loss is 27.31 , inference loss is 16.43, weight deacy loss is 10.88, training accuracy is 0.125000, time 54.403 samples/sec
+epoch 2, total_step 45600, total loss is 28.17 , inference loss is 17.29, weight deacy loss is 10.88, training accuracy is 0.062500, time 54.859 samples/sec
+epoch 2, total_step 45620, total loss is 23.54 , inference loss is 12.66, weight deacy loss is 10.88, training accuracy is 0.125000, time 56.174 samples/sec
+epoch 2, total_step 45640, total loss is 29.65 , inference loss is 18.78, weight deacy loss is 10.88, training accuracy is 0.125000, time 55.632 samples/sec
+epoch 2, total_step 45660, total loss is 30.69 , inference loss is 19.81, weight deacy loss is 10.88, training accuracy is 0.062500, time 55.856 samples/sec
+epoch 2, total_step 45680, total loss is 28.08 , inference loss is 17.20, weight deacy loss is 10.87, training accuracy is 0.093750, time 54.776 samples/sec
+epoch 2, total_step 45700, total loss is 27.81 , inference loss is 16.94, weight deacy loss is 10.87, training accuracy is 0.031250, time 55.523 samples/sec
+epoch 2, total_step 45720, total loss is 28.45 , inference loss is 17.58, weight deacy loss is 10.87, training accuracy is 0.062500, time 53.942 samples/sec
+epoch 2, total_step 45740, total loss is 27.85 , inference loss is 16.98, weight deacy loss is 10.87, training accuracy is 0.187500, time 53.855 samples/sec
+epoch 2, total_step 45760, total loss is 26.21 , inference loss is 15.34, weight deacy loss is 10.87, training accuracy is 0.218750, time 56.553 samples/sec
+epoch 2, total_step 45780, total loss is 25.50 , inference loss is 14.63, weight deacy loss is 10.87, training accuracy is 0.218750, time 55.315 samples/sec
+epoch 2, total_step 45800, total loss is 28.51 , inference loss is 17.65, weight deacy loss is 10.87, training accuracy is 0.093750, time 56.679 samples/sec
+epoch 2, total_step 45820, total loss is 26.02 , inference loss is 15.15, weight deacy loss is 10.87, training accuracy is 0.062500, time 56.171 samples/sec
+epoch 2, total_step 45840, total loss is 27.49 , inference loss is 16.62, weight deacy loss is 10.86, training accuracy is 0.187500, time 57.509 samples/sec
+epoch 2, total_step 45860, total loss is 28.87 , inference loss is 18.01, weight deacy loss is 10.86, training accuracy is 0.000000, time 56.261 samples/sec
+epoch 2, total_step 45880, total loss is 27.83 , inference loss is 16.97, weight deacy loss is 10.86, training accuracy is 0.062500, time 56.647 samples/sec
+epoch 2, total_step 45900, total loss is 26.82 , inference loss is 15.96, weight deacy loss is 10.86, training accuracy is 0.156250, time 56.153 samples/sec
+epoch 2, total_step 45920, total loss is 27.55 , inference loss is 16.69, weight deacy loss is 10.86, training accuracy is 0.031250, time 56.590 samples/sec
+epoch 2, total_step 45940, total loss is 26.06 , inference loss is 15.20, weight deacy loss is 10.86, training accuracy is 0.156250, time 55.637 samples/sec
+epoch 2, total_step 45960, total loss is 23.43 , inference loss is 12.57, weight deacy loss is 10.86, training accuracy is 0.062500, time 55.884 samples/sec
+epoch 2, total_step 45980, total loss is 25.64 , inference loss is 14.78, weight deacy loss is 10.86, training accuracy is 0.187500, time 56.634 samples/sec
+
+
+
+
+End of epoch 2
+testing verification..
+(12000, 512)
+infer time 73.24627699999999
+best_threshold_index 114 0.9811111111111112
+best_threshold_index 115 0.98
+best_threshold_index 115 0.9805555555555555
+best_threshold_index 114 0.9816666666666667
+best_threshold_index 114 0.9803703703703703
+best_threshold_index 115 0.9790740740740741
+best_threshold_index 115 0.980925925925926
+best_threshold_index 115 0.9794444444444445
+best_threshold_index 115 0.9787037037037037
+best_threshold_index 115 0.98
+[lfw][46000]XNorm: 22.340609
+[lfw][46000]Accuracy-Flip: 0.97950+-0.00847
+testing verification..
+(14000, 512)
+infer time 85.72998499999996
+best_threshold_index 123 0.9801587301587301
+best_threshold_index 121 0.9788888888888889
+best_threshold_index 123 0.9796825396825397
+best_threshold_index 123 0.9787301587301588
+best_threshold_index 123 0.979047619047619
+best_threshold_index 123 0.9782539682539683
+best_threshold_index 123 0.9793650793650793
+best_threshold_index 123 0.9780952380952381
+best_threshold_index 123 0.9796825396825397
+best_threshold_index 124 0.9798412698412698
+[cfp_ff][46000]XNorm: 21.866058
+[cfp_ff][46000]Accuracy-Flip: 0.97843+-0.00597
+testing verification..
+(14000, 512)
+infer time 85.65134700000006
+best_threshold_index 141 0.8887301587301587
+best_threshold_index 139 0.8893650793650794
+best_threshold_index 141 0.888095238095238
+best_threshold_index 142 0.8895238095238095
+best_threshold_index 142 0.8912698412698413
+best_threshold_index 139 0.8874603174603175
+best_threshold_index 142 0.8852380952380953
+best_threshold_index 139 0.8885714285714286
+best_threshold_index 139 0.8884126984126984
+best_threshold_index 142 0.8890476190476191
+[cfp_fp][46000]XNorm: 21.223604
+[cfp_fp][46000]Accuracy-Flip: 0.88629+-0.01319
+testing verification..
+(12000, 512)
+infer time 74.28391899999993
+best_threshold_index 145 0.8703703703703703
+best_threshold_index 145 0.8690740740740741
+best_threshold_index 145 0.8688888888888889
+best_threshold_index 145 0.8714814814814815
+best_threshold_index 142 0.8761111111111111
+best_threshold_index 142 0.8711111111111111
+best_threshold_index 142 0.8701851851851852
+best_threshold_index 145 0.8770370370370371
+best_threshold_index 142 0.8742592592592593
+best_threshold_index 145 0.8716666666666667
+[agedb_30][46000]XNorm: 21.563179
+[agedb_30][46000]Accuracy-Flip: 0.86883+-0.02477
+test accuracy is: 0.9795000000000001
+epoch 3, total_step 46000, total loss is 40.12 , inference loss is 29.27, weight deacy loss is 10.85, training accuracy is 0.000000, time 52.764 samples/sec
+epoch 3, total_step 46020, total loss is 28.52 , inference loss is 17.66, weight deacy loss is 10.85, training accuracy is 0.125000, time 56.165 samples/sec
+epoch 3, total_step 46040, total loss is 23.75 , inference loss is 12.90, weight deacy loss is 10.85, training accuracy is 0.156250, time 56.205 samples/sec
+epoch 3, total_step 46060, total loss is 21.71 , inference loss is 10.86, weight deacy loss is 10.85, training accuracy is 0.343750, time 55.996 samples/sec
+epoch 3, total_step 46080, total loss is 21.61 , inference loss is 10.76, weight deacy loss is 10.85, training accuracy is 0.375000, time 57.241 samples/sec
+epoch 3, total_step 46100, total loss is 19.01 , inference loss is 8.16, weight deacy loss is 10.85, training accuracy is 0.343750, time 56.635 samples/sec
+epoch 3, total_step 46120, total loss is 22.47 , inference loss is 11.62, weight deacy loss is 10.85, training accuracy is 0.437500, time 55.630 samples/sec
+epoch 3, total_step 46140, total loss is 17.51 , inference loss is 6.67, weight deacy loss is 10.84, training accuracy is 0.437500, time 55.543 samples/sec
+epoch 3, total_step 46160, total loss is 21.28 , inference loss is 10.43, weight deacy loss is 10.84, training accuracy is 0.281250, time 56.701 samples/sec
+epoch 3, total_step 46180, total loss is 23.02 , inference loss is 12.18, weight deacy loss is 10.84, training accuracy is 0.281250, time 58.099 samples/sec
+epoch 3, total_step 46200, total loss is 21.39 , inference loss is 10.55, weight deacy loss is 10.84, training accuracy is 0.468750, time 56.125 samples/sec
+epoch 3, total_step 46220, total loss is 25.63 , inference loss is 14.79, weight deacy loss is 10.84, training accuracy is 0.281250, time 54.773 samples/sec
+epoch 3, total_step 46240, total loss is 23.85 , inference loss is 13.01, weight deacy loss is 10.84, training accuracy is 0.093750, time 55.856 samples/sec
+epoch 3, total_step 46260, total loss is 23.19 , inference loss is 12.35, weight deacy loss is 10.84, training accuracy is 0.312500, time 56.353 samples/sec
+epoch 3, total_step 46280, total loss is 24.56 , inference loss is 13.73, weight deacy loss is 10.83, training accuracy is 0.250000, time 56.130 samples/sec
+epoch 3, total_step 46300, total loss is 28.63 , inference loss is 17.80, weight deacy loss is 10.83, training accuracy is 0.343750, time 56.443 samples/sec
+epoch 3, total_step 46320, total loss is 21.52 , inference loss is 10.69, weight deacy loss is 10.83, training accuracy is 0.250000, time 56.357 samples/sec
+epoch 3, total_step 46340, total loss is 23.12 , inference loss is 12.29, weight deacy loss is 10.83, training accuracy is 0.250000, time 54.469 samples/sec
+epoch 3, total_step 46360, total loss is 20.34 , inference loss is 9.51, weight deacy loss is 10.83, training accuracy is 0.343750, time 55.090 samples/sec
+epoch 3, total_step 46380, total loss is 24.89 , inference loss is 14.06, weight deacy loss is 10.83, training accuracy is 0.250000, time 56.730 samples/sec
+epoch 3, total_step 46400, total loss is 16.96 , inference loss is 6.13, weight deacy loss is 10.83, training accuracy is 0.531250, time 54.294 samples/sec
+epoch 3, total_step 46420, total loss is 23.39 , inference loss is 12.56, weight deacy loss is 10.82, training accuracy is 0.312500, time 55.640 samples/sec
+epoch 3, total_step 46440, total loss is 26.90 , inference loss is 16.08, weight deacy loss is 10.82, training accuracy is 0.250000, time 55.246 samples/sec
+epoch 3, total_step 46460, total loss is 23.28 , inference loss is 12.46, weight deacy loss is 10.82, training accuracy is 0.250000, time 55.280 samples/sec
+epoch 3, total_step 46480, total loss is 22.02 , inference loss is 11.20, weight deacy loss is 10.82, training accuracy is 0.218750, time 56.562 samples/sec
+epoch 3, total_step 46500, total loss is 24.65 , inference loss is 13.83, weight deacy loss is 10.82, training accuracy is 0.187500, time 56.260 samples/sec
+epoch 3, total_step 46520, total loss is 23.85 , inference loss is 13.03, weight deacy loss is 10.82, training accuracy is 0.281250, time 55.343 samples/sec
+epoch 3, total_step 46540, total loss is 22.54 , inference loss is 11.72, weight deacy loss is 10.82, training accuracy is 0.250000, time 56.287 samples/sec
+epoch 3, total_step 46560, total loss is 23.84 , inference loss is 13.02, weight deacy loss is 10.81, training accuracy is 0.125000, time 55.305 samples/sec
+epoch 3, total_step 46580, total loss is 23.27 , inference loss is 12.45, weight deacy loss is 10.81, training accuracy is 0.187500, time 55.345 samples/sec
+epoch 3, total_step 46600, total loss is 28.02 , inference loss is 17.20, weight deacy loss is 10.81, training accuracy is 0.218750, time 56.241 samples/sec
+epoch 3, total_step 46620, total loss is 23.88 , inference loss is 13.07, weight deacy loss is 10.81, training accuracy is 0.343750, time 56.328 samples/sec
+epoch 3, total_step 46640, total loss is 24.22 , inference loss is 13.41, weight deacy loss is 10.81, training accuracy is 0.343750, time 56.294 samples/sec
+epoch 3, total_step 46660, total loss is 25.25 , inference loss is 14.44, weight deacy loss is 10.81, training accuracy is 0.125000, time 54.507 samples/sec
+epoch 3, total_step 46680, total loss is 24.55 , inference loss is 13.75, weight deacy loss is 10.81, training accuracy is 0.343750, time 54.248 samples/sec
+epoch 3, total_step 46700, total loss is 22.33 , inference loss is 11.52, weight deacy loss is 10.81, training accuracy is 0.281250, time 56.608 samples/sec
+epoch 3, total_step 46720, total loss is 25.20 , inference loss is 14.40, weight deacy loss is 10.80, training accuracy is 0.187500, time 56.424 samples/sec
+epoch 3, total_step 46740, total loss is 27.41 , inference loss is 16.61, weight deacy loss is 10.80, training accuracy is 0.218750, time 56.597 samples/sec
+epoch 3, total_step 46760, total loss is 25.03 , inference loss is 14.23, weight deacy loss is 10.80, training accuracy is 0.187500, time 54.184 samples/sec
+epoch 3, total_step 46780, total loss is 22.45 , inference loss is 11.65, weight deacy loss is 10.80, training accuracy is 0.250000, time 56.634 samples/sec
+epoch 3, total_step 46800, total loss is 20.17 , inference loss is 9.37, weight deacy loss is 10.80, training accuracy is 0.343750, time 55.790 samples/sec
+epoch 3, total_step 46820, total loss is 20.65 , inference loss is 9.85, weight deacy loss is 10.80, training accuracy is 0.375000, time 55.526 samples/sec
+epoch 3, total_step 46840, total loss is 23.20 , inference loss is 12.40, weight deacy loss is 10.80, training accuracy is 0.187500, time 56.234 samples/sec
+epoch 3, total_step 46860, total loss is 20.41 , inference loss is 9.62, weight deacy loss is 10.79, training accuracy is 0.437500, time 54.602 samples/sec
+epoch 3, total_step 46880, total loss is 23.46 , inference loss is 12.66, weight deacy loss is 10.79, training accuracy is 0.312500, time 56.791 samples/sec
+epoch 3, total_step 46900, total loss is 28.63 , inference loss is 17.84, weight deacy loss is 10.79, training accuracy is 0.218750, time 56.055 samples/sec
+epoch 3, total_step 46920, total loss is 25.00 , inference loss is 14.21, weight deacy loss is 10.79, training accuracy is 0.343750, time 55.796 samples/sec
+epoch 3, total_step 46940, total loss is 25.31 , inference loss is 14.52, weight deacy loss is 10.79, training accuracy is 0.187500, time 55.839 samples/sec
+epoch 3, total_step 46960, total loss is 20.85 , inference loss is 10.06, weight deacy loss is 10.79, training accuracy is 0.250000, time 55.530 samples/sec
+epoch 3, total_step 46980, total loss is 22.04 , inference loss is 11.25, weight deacy loss is 10.79, training accuracy is 0.312500, time 54.834 samples/sec
+epoch 3, total_step 47000, total loss is 19.96 , inference loss is 9.17, weight deacy loss is 10.79, training accuracy is 0.375000, time 56.642 samples/sec
+epoch 3, total_step 47020, total loss is 24.63 , inference loss is 13.85, weight deacy loss is 10.78, training accuracy is 0.093750, time 56.688 samples/sec
+epoch 3, total_step 47040, total loss is 22.97 , inference loss is 12.19, weight deacy loss is 10.78, training accuracy is 0.343750, time 56.698 samples/sec
+epoch 3, total_step 47060, total loss is 21.47 , inference loss is 10.69, weight deacy loss is 10.78, training accuracy is 0.312500, time 56.112 samples/sec
+epoch 3, total_step 47080, total loss is 24.69 , inference loss is 13.91, weight deacy loss is 10.78, training accuracy is 0.312500, time 54.973 samples/sec
+epoch 3, total_step 47100, total loss is 30.17 , inference loss is 19.39, weight deacy loss is 10.78, training accuracy is 0.218750, time 55.715 samples/sec
+epoch 3, total_step 47120, total loss is 24.05 , inference loss is 13.27, weight deacy loss is 10.78, training accuracy is 0.218750, time 56.533 samples/sec
+epoch 3, total_step 47140, total loss is 22.38 , inference loss is 11.60, weight deacy loss is 10.78, training accuracy is 0.281250, time 55.734 samples/sec
+epoch 3, total_step 47160, total loss is 22.15 , inference loss is 11.37, weight deacy loss is 10.77, training accuracy is 0.250000, time 56.271 samples/sec
+epoch 3, total_step 47180, total loss is 23.00 , inference loss is 12.22, weight deacy loss is 10.77, training accuracy is 0.218750, time 55.814 samples/sec
+epoch 3, total_step 47200, total loss is 24.02 , inference loss is 13.24, weight deacy loss is 10.77, training accuracy is 0.187500, time 56.160 samples/sec
+epoch 3, total_step 47220, total loss is 28.81 , inference loss is 18.04, weight deacy loss is 10.77, training accuracy is 0.218750, time 56.346 samples/sec
+epoch 3, total_step 47240, total loss is 24.54 , inference loss is 13.77, weight deacy loss is 10.77, training accuracy is 0.250000, time 56.128 samples/sec
+epoch 3, total_step 47260, total loss is 26.13 , inference loss is 15.36, weight deacy loss is 10.77, training accuracy is 0.250000, time 56.014 samples/sec
+epoch 3, total_step 47280, total loss is 24.86 , inference loss is 14.09, weight deacy loss is 10.77, training accuracy is 0.187500, time 55.499 samples/sec
+epoch 3, total_step 47300, total loss is 27.61 , inference loss is 16.84, weight deacy loss is 10.77, training accuracy is 0.125000, time 56.244 samples/sec
+epoch 3, total_step 47320, total loss is 26.10 , inference loss is 15.33, weight deacy loss is 10.76, training accuracy is 0.312500, time 56.145 samples/sec
+epoch 3, total_step 47340, total loss is 22.40 , inference loss is 11.64, weight deacy loss is 10.76, training accuracy is 0.218750, time 55.629 samples/sec
+epoch 3, total_step 47360, total loss is 26.77 , inference loss is 16.01, weight deacy loss is 10.76, training accuracy is 0.187500, time 56.419 samples/sec
+epoch 3, total_step 47380, total loss is 25.56 , inference loss is 14.80, weight deacy loss is 10.76, training accuracy is 0.187500, time 56.643 samples/sec
+epoch 3, total_step 47400, total loss is 24.33 , inference loss is 13.57, weight deacy loss is 10.76, training accuracy is 0.187500, time 55.505 samples/sec
+epoch 3, total_step 47420, total loss is 27.45 , inference loss is 16.70, weight deacy loss is 10.76, training accuracy is 0.062500, time 56.234 samples/sec
+epoch 3, total_step 47440, total loss is 24.37 , inference loss is 13.62, weight deacy loss is 10.76, training accuracy is 0.125000, time 55.878 samples/sec
+epoch 3, total_step 47460, total loss is 27.06 , inference loss is 16.31, weight deacy loss is 10.75, training accuracy is 0.062500, time 55.821 samples/sec
+epoch 3, total_step 47480, total loss is 20.90 , inference loss is 10.15, weight deacy loss is 10.75, training accuracy is 0.250000, time 55.270 samples/sec
+epoch 3, total_step 47500, total loss is 22.61 , inference loss is 11.86, weight deacy loss is 10.75, training accuracy is 0.218750, time 56.160 samples/sec
+epoch 3, total_step 47520, total loss is 22.53 , inference loss is 11.78, weight deacy loss is 10.75, training accuracy is 0.250000, time 56.529 samples/sec
+epoch 3, total_step 47540, total loss is 26.63 , inference loss is 15.88, weight deacy loss is 10.75, training accuracy is 0.218750, time 55.796 samples/sec
+epoch 3, total_step 47560, total loss is 28.31 , inference loss is 17.56, weight deacy loss is 10.75, training accuracy is 0.218750, time 56.694 samples/sec
+epoch 3, total_step 47580, total loss is 22.07 , inference loss is 11.33, weight deacy loss is 10.75, training accuracy is 0.250000, time 56.531 samples/sec
+epoch 3, total_step 47600, total loss is 22.94 , inference loss is 12.19, weight deacy loss is 10.74, training accuracy is 0.312500, time 55.728 samples/sec
+epoch 3, total_step 47620, total loss is 26.60 , inference loss is 15.86, weight deacy loss is 10.74, training accuracy is 0.281250, time 56.593 samples/sec
+epoch 3, total_step 47640, total loss is 22.86 , inference loss is 12.12, weight deacy loss is 10.74, training accuracy is 0.218750, time 55.769 samples/sec
+epoch 3, total_step 47660, total loss is 19.92 , inference loss is 9.18, weight deacy loss is 10.74, training accuracy is 0.093750, time 56.485 samples/sec
+epoch 3, total_step 47680, total loss is 26.04 , inference loss is 15.31, weight deacy loss is 10.74, training accuracy is 0.250000, time 56.996 samples/sec
+epoch 3, total_step 47700, total loss is 25.29 , inference loss is 14.56, weight deacy loss is 10.74, training accuracy is 0.250000, time 56.290 samples/sec
+epoch 3, total_step 47720, total loss is 22.62 , inference loss is 11.88, weight deacy loss is 10.74, training accuracy is 0.156250, time 57.136 samples/sec
+epoch 3, total_step 47740, total loss is 23.57 , inference loss is 12.84, weight deacy loss is 10.74, training accuracy is 0.125000, time 56.499 samples/sec
+epoch 3, total_step 47760, total loss is 23.33 , inference loss is 12.60, weight deacy loss is 10.73, training accuracy is 0.281250, time 56.514 samples/sec
+epoch 3, total_step 47780, total loss is 22.18 , inference loss is 11.45, weight deacy loss is 10.73, training accuracy is 0.218750, time 56.692 samples/sec
+epoch 3, total_step 47800, total loss is 26.61 , inference loss is 15.88, weight deacy loss is 10.73, training accuracy is 0.125000, time 56.584 samples/sec
+epoch 3, total_step 47820, total loss is 23.20 , inference loss is 12.47, weight deacy loss is 10.73, training accuracy is 0.406250, time 56.298 samples/sec
+epoch 3, total_step 47840, total loss is 22.58 , inference loss is 11.85, weight deacy loss is 10.73, training accuracy is 0.250000, time 55.555 samples/sec
+epoch 3, total_step 47860, total loss is 19.44 , inference loss is 8.72, weight deacy loss is 10.73, training accuracy is 0.312500, time 55.449 samples/sec
+epoch 3, total_step 47880, total loss is 18.32 , inference loss is 7.59, weight deacy loss is 10.73, training accuracy is 0.406250, time 55.559 samples/sec
+epoch 3, total_step 47900, total loss is 24.07 , inference loss is 13.35, weight deacy loss is 10.72, training accuracy is 0.281250, time 56.597 samples/sec
+epoch 3, total_step 47920, total loss is 24.41 , inference loss is 13.68, weight deacy loss is 10.72, training accuracy is 0.250000, time 56.837 samples/sec
+epoch 3, total_step 47940, total loss is 17.51 , inference loss is 6.79, weight deacy loss is 10.72, training accuracy is 0.375000, time 55.947 samples/sec
+epoch 3, total_step 47960, total loss is 20.92 , inference loss is 10.20, weight deacy loss is 10.72, training accuracy is 0.250000, time 55.703 samples/sec
+epoch 3, total_step 47980, total loss is 24.58 , inference loss is 13.86, weight deacy loss is 10.72, training accuracy is 0.125000, time 56.402 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.070423
+best_threshold_index 108 0.9818518518518519
+best_threshold_index 108 0.9812962962962963
+best_threshold_index 109 0.982037037037037
+best_threshold_index 108 0.9829629629629629
+best_threshold_index 108 0.9814814814814815
+best_threshold_index 108 0.9807407407407407
+best_threshold_index 112 0.9818518518518519
+best_threshold_index 112 0.9812962962962963
+best_threshold_index 112 0.9807407407407407
+best_threshold_index 108 0.9803703703703703
+[lfw][48000]XNorm: 23.073484
+[lfw][48000]Accuracy-Flip: 0.98017+-0.00647
+testing verification..
+(14000, 512)
+infer time 85.91917899999993
+best_threshold_index 113 0.9787301587301588
+best_threshold_index 113 0.9776190476190476
+best_threshold_index 113 0.9774603174603175
+best_threshold_index 113 0.9755555555555555
+best_threshold_index 113 0.976984126984127
+best_threshold_index 113 0.9768253968253968
+best_threshold_index 113 0.9774603174603175
+best_threshold_index 113 0.9758730158730159
+best_threshold_index 113 0.9768253968253968
+best_threshold_index 113 0.9780952380952381
+[cfp_ff][48000]XNorm: 22.186944
+[cfp_ff][48000]Accuracy-Flip: 0.97714+-0.00816
+testing verification..
+(14000, 512)
+infer time 85.91559699999996
+best_threshold_index 136 0.8849206349206349
+best_threshold_index 136 0.8836507936507937
+best_threshold_index 136 0.883015873015873
+best_threshold_index 136 0.8846031746031746
+best_threshold_index 136 0.8852380952380953
+best_threshold_index 136 0.8809523809523809
+best_threshold_index 136 0.8815873015873016
+best_threshold_index 136 0.8819047619047619
+best_threshold_index 136 0.8836507936507937
+best_threshold_index 136 0.8847619047619047
+[cfp_fp][48000]XNorm: 22.401031
+[cfp_fp][48000]Accuracy-Flip: 0.88343+-0.01297
+testing verification..
+(12000, 512)
+infer time 73.43203299999999
+best_threshold_index 133 0.8770370370370371
+best_threshold_index 133 0.8772222222222222
+best_threshold_index 133 0.8783333333333333
+best_threshold_index 129 0.8768518518518519
+best_threshold_index 129 0.8840740740740741
+best_threshold_index 133 0.8762962962962964
+best_threshold_index 131 0.8794444444444445
+best_threshold_index 133 0.8835185185185185
+best_threshold_index 133 0.8820370370370371
+best_threshold_index 133 0.8814814814814815
+[agedb_30][48000]XNorm: 23.039292
+[agedb_30][48000]Accuracy-Flip: 0.87617+-0.02569
+test accuracy is: 0.9801666666666667
+epoch 3, total_step 48000, total loss is 23.25 , inference loss is 12.53, weight deacy loss is 10.72, training accuracy is 0.281250, time 53.019 samples/sec
+epoch 3, total_step 48020, total loss is 18.78 , inference loss is 8.07, weight deacy loss is 10.72, training accuracy is 0.375000, time 56.393 samples/sec
+epoch 3, total_step 48040, total loss is 22.92 , inference loss is 12.21, weight deacy loss is 10.71, training accuracy is 0.343750, time 55.996 samples/sec
+epoch 3, total_step 48060, total loss is 26.00 , inference loss is 15.29, weight deacy loss is 10.71, training accuracy is 0.375000, time 56.687 samples/sec
+epoch 3, total_step 48080, total loss is 22.32 , inference loss is 11.61, weight deacy loss is 10.71, training accuracy is 0.281250, time 56.164 samples/sec
+epoch 3, total_step 48100, total loss is 21.54 , inference loss is 10.83, weight deacy loss is 10.71, training accuracy is 0.343750, time 56.602 samples/sec
+epoch 3, total_step 48120, total loss is 22.03 , inference loss is 11.32, weight deacy loss is 10.71, training accuracy is 0.312500, time 54.208 samples/sec
+epoch 3, total_step 48140, total loss is 22.98 , inference loss is 12.27, weight deacy loss is 10.71, training accuracy is 0.281250, time 57.129 samples/sec
+epoch 3, total_step 48160, total loss is 22.83 , inference loss is 12.12, weight deacy loss is 10.71, training accuracy is 0.250000, time 55.701 samples/sec
+epoch 3, total_step 48180, total loss is 22.26 , inference loss is 11.56, weight deacy loss is 10.70, training accuracy is 0.281250, time 55.013 samples/sec
+epoch 3, total_step 48200, total loss is 25.77 , inference loss is 15.07, weight deacy loss is 10.70, training accuracy is 0.312500, time 56.235 samples/sec
+epoch 3, total_step 48220, total loss is 22.88 , inference loss is 12.18, weight deacy loss is 10.70, training accuracy is 0.250000, time 56.006 samples/sec
+epoch 3, total_step 48240, total loss is 24.08 , inference loss is 13.38, weight deacy loss is 10.70, training accuracy is 0.218750, time 56.492 samples/sec
+epoch 3, total_step 48260, total loss is 23.51 , inference loss is 12.81, weight deacy loss is 10.70, training accuracy is 0.250000, time 56.237 samples/sec
+epoch 3, total_step 48280, total loss is 26.36 , inference loss is 15.66, weight deacy loss is 10.70, training accuracy is 0.187500, time 56.211 samples/sec
+epoch 3, total_step 48300, total loss is 18.74 , inference loss is 8.05, weight deacy loss is 10.70, training accuracy is 0.500000, time 56.157 samples/sec
+epoch 3, total_step 48320, total loss is 21.28 , inference loss is 10.59, weight deacy loss is 10.69, training accuracy is 0.156250, time 56.797 samples/sec
+epoch 3, total_step 48340, total loss is 18.24 , inference loss is 7.54, weight deacy loss is 10.69, training accuracy is 0.218750, time 56.112 samples/sec
+epoch 3, total_step 48360, total loss is 20.38 , inference loss is 9.69, weight deacy loss is 10.69, training accuracy is 0.312500, time 55.942 samples/sec
+epoch 3, total_step 48380, total loss is 20.31 , inference loss is 9.62, weight deacy loss is 10.69, training accuracy is 0.343750, time 55.623 samples/sec
+epoch 3, total_step 48400, total loss is 22.38 , inference loss is 11.70, weight deacy loss is 10.69, training accuracy is 0.281250, time 56.247 samples/sec
+epoch 3, total_step 48420, total loss is 23.83 , inference loss is 13.15, weight deacy loss is 10.69, training accuracy is 0.343750, time 55.574 samples/sec
+epoch 3, total_step 48440, total loss is 24.61 , inference loss is 13.93, weight deacy loss is 10.69, training accuracy is 0.250000, time 56.278 samples/sec
+epoch 3, total_step 48460, total loss is 24.78 , inference loss is 14.09, weight deacy loss is 10.68, training accuracy is 0.375000, time 56.236 samples/sec
+epoch 3, total_step 48480, total loss is 20.65 , inference loss is 9.96, weight deacy loss is 10.68, training accuracy is 0.343750, time 56.487 samples/sec
+epoch 3, total_step 48500, total loss is 22.63 , inference loss is 11.95, weight deacy loss is 10.68, training accuracy is 0.187500, time 56.207 samples/sec
+epoch 3, total_step 48520, total loss is 20.62 , inference loss is 9.94, weight deacy loss is 10.68, training accuracy is 0.218750, time 56.347 samples/sec
+epoch 3, total_step 48540, total loss is 20.22 , inference loss is 9.54, weight deacy loss is 10.68, training accuracy is 0.406250, time 54.103 samples/sec
+epoch 3, total_step 48560, total loss is 26.64 , inference loss is 15.96, weight deacy loss is 10.68, training accuracy is 0.125000, time 56.157 samples/sec
+epoch 3, total_step 48580, total loss is 21.63 , inference loss is 10.95, weight deacy loss is 10.68, training accuracy is 0.375000, time 55.889 samples/sec
+epoch 3, total_step 48600, total loss is 21.91 , inference loss is 11.24, weight deacy loss is 10.68, training accuracy is 0.375000, time 55.726 samples/sec
+epoch 3, total_step 48620, total loss is 24.71 , inference loss is 14.04, weight deacy loss is 10.67, training accuracy is 0.187500, time 56.535 samples/sec
+epoch 3, total_step 48640, total loss is 21.42 , inference loss is 10.75, weight deacy loss is 10.67, training accuracy is 0.250000, time 56.341 samples/sec
+epoch 3, total_step 48660, total loss is 21.29 , inference loss is 10.62, weight deacy loss is 10.67, training accuracy is 0.281250, time 56.032 samples/sec
+epoch 3, total_step 48680, total loss is 22.13 , inference loss is 11.46, weight deacy loss is 10.67, training accuracy is 0.218750, time 54.054 samples/sec
+epoch 3, total_step 48700, total loss is 21.96 , inference loss is 11.30, weight deacy loss is 10.67, training accuracy is 0.250000, time 55.325 samples/sec
+epoch 3, total_step 48720, total loss is 17.13 , inference loss is 6.46, weight deacy loss is 10.67, training accuracy is 0.375000, time 57.602 samples/sec
+epoch 3, total_step 48740, total loss is 24.05 , inference loss is 13.38, weight deacy loss is 10.67, training accuracy is 0.281250, time 56.105 samples/sec
+epoch 3, total_step 48760, total loss is 18.30 , inference loss is 7.64, weight deacy loss is 10.66, training accuracy is 0.375000, time 55.764 samples/sec
+epoch 3, total_step 48780, total loss is 21.79 , inference loss is 11.12, weight deacy loss is 10.66, training accuracy is 0.250000, time 53.861 samples/sec
+epoch 3, total_step 48800, total loss is 25.17 , inference loss is 14.51, weight deacy loss is 10.66, training accuracy is 0.187500, time 54.134 samples/sec
+epoch 3, total_step 48820, total loss is 21.97 , inference loss is 11.31, weight deacy loss is 10.66, training accuracy is 0.218750, time 54.641 samples/sec
+epoch 3, total_step 48840, total loss is 24.07 , inference loss is 13.41, weight deacy loss is 10.66, training accuracy is 0.281250, time 55.398 samples/sec
+epoch 3, total_step 48860, total loss is 16.55 , inference loss is 5.89, weight deacy loss is 10.66, training accuracy is 0.343750, time 56.072 samples/sec
+epoch 3, total_step 48880, total loss is 21.70 , inference loss is 11.05, weight deacy loss is 10.66, training accuracy is 0.281250, time 55.541 samples/sec
+epoch 3, total_step 48900, total loss is 20.47 , inference loss is 9.82, weight deacy loss is 10.65, training accuracy is 0.343750, time 56.473 samples/sec
+epoch 3, total_step 48920, total loss is 23.19 , inference loss is 12.54, weight deacy loss is 10.65, training accuracy is 0.343750, time 56.213 samples/sec
+epoch 3, total_step 48940, total loss is 21.34 , inference loss is 10.69, weight deacy loss is 10.65, training accuracy is 0.312500, time 55.584 samples/sec
+epoch 3, total_step 48960, total loss is 22.03 , inference loss is 11.38, weight deacy loss is 10.65, training accuracy is 0.343750, time 56.069 samples/sec
+epoch 3, total_step 48980, total loss is 19.23 , inference loss is 8.58, weight deacy loss is 10.65, training accuracy is 0.281250, time 56.228 samples/sec
+epoch 3, total_step 49000, total loss is 23.35 , inference loss is 12.70, weight deacy loss is 10.65, training accuracy is 0.218750, time 56.618 samples/sec
+epoch 3, total_step 49020, total loss is 18.66 , inference loss is 8.02, weight deacy loss is 10.65, training accuracy is 0.406250, time 56.104 samples/sec
+epoch 3, total_step 49040, total loss is 26.15 , inference loss is 15.50, weight deacy loss is 10.64, training accuracy is 0.187500, time 55.914 samples/sec
+epoch 3, total_step 49060, total loss is 22.25 , inference loss is 11.61, weight deacy loss is 10.64, training accuracy is 0.343750, time 56.260 samples/sec
+epoch 3, total_step 49080, total loss is 21.14 , inference loss is 10.50, weight deacy loss is 10.64, training accuracy is 0.250000, time 56.658 samples/sec
+epoch 3, total_step 49100, total loss is 23.34 , inference loss is 12.69, weight deacy loss is 10.64, training accuracy is 0.218750, time 56.575 samples/sec
+epoch 3, total_step 49120, total loss is 19.53 , inference loss is 8.89, weight deacy loss is 10.64, training accuracy is 0.406250, time 55.395 samples/sec
+epoch 3, total_step 49140, total loss is 26.16 , inference loss is 15.52, weight deacy loss is 10.64, training accuracy is 0.250000, time 56.186 samples/sec
+epoch 3, total_step 49160, total loss is 24.06 , inference loss is 13.42, weight deacy loss is 10.64, training accuracy is 0.250000, time 56.478 samples/sec
+epoch 3, total_step 49180, total loss is 21.46 , inference loss is 10.83, weight deacy loss is 10.64, training accuracy is 0.312500, time 55.750 samples/sec
+epoch 3, total_step 49200, total loss is 24.99 , inference loss is 14.36, weight deacy loss is 10.63, training accuracy is 0.250000, time 51.693 samples/sec
+epoch 3, total_step 49220, total loss is 23.62 , inference loss is 12.99, weight deacy loss is 10.63, training accuracy is 0.312500, time 55.392 samples/sec
+epoch 3, total_step 49240, total loss is 23.87 , inference loss is 13.23, weight deacy loss is 10.63, training accuracy is 0.281250, time 56.543 samples/sec
+epoch 3, total_step 49260, total loss is 25.25 , inference loss is 14.62, weight deacy loss is 10.63, training accuracy is 0.218750, time 55.803 samples/sec
+epoch 3, total_step 49280, total loss is 27.24 , inference loss is 16.61, weight deacy loss is 10.63, training accuracy is 0.156250, time 56.329 samples/sec
+epoch 3, total_step 49300, total loss is 23.47 , inference loss is 12.84, weight deacy loss is 10.63, training accuracy is 0.218750, time 56.380 samples/sec
+epoch 3, total_step 49320, total loss is 20.23 , inference loss is 9.61, weight deacy loss is 10.63, training accuracy is 0.312500, time 56.198 samples/sec
+epoch 3, total_step 49340, total loss is 23.46 , inference loss is 12.84, weight deacy loss is 10.62, training accuracy is 0.187500, time 56.124 samples/sec
+epoch 3, total_step 49360, total loss is 22.25 , inference loss is 11.63, weight deacy loss is 10.62, training accuracy is 0.312500, time 56.352 samples/sec
+epoch 3, total_step 49380, total loss is 27.41 , inference loss is 16.79, weight deacy loss is 10.62, training accuracy is 0.156250, time 57.096 samples/sec
+epoch 3, total_step 49400, total loss is 23.48 , inference loss is 12.86, weight deacy loss is 10.62, training accuracy is 0.312500, time 56.414 samples/sec
+epoch 3, total_step 49420, total loss is 19.51 , inference loss is 8.90, weight deacy loss is 10.62, training accuracy is 0.343750, time 56.948 samples/sec
+epoch 3, total_step 49440, total loss is 25.48 , inference loss is 14.87, weight deacy loss is 10.62, training accuracy is 0.250000, time 56.130 samples/sec
+epoch 3, total_step 49460, total loss is 22.58 , inference loss is 11.96, weight deacy loss is 10.62, training accuracy is 0.437500, time 54.924 samples/sec
+epoch 3, total_step 49480, total loss is 25.25 , inference loss is 14.64, weight deacy loss is 10.62, training accuracy is 0.312500, time 56.155 samples/sec
+epoch 3, total_step 49500, total loss is 24.01 , inference loss is 13.39, weight deacy loss is 10.61, training accuracy is 0.281250, time 56.399 samples/sec
+epoch 3, total_step 49520, total loss is 23.20 , inference loss is 12.59, weight deacy loss is 10.61, training accuracy is 0.156250, time 55.875 samples/sec
+epoch 3, total_step 49540, total loss is 28.19 , inference loss is 17.58, weight deacy loss is 10.61, training accuracy is 0.218750, time 56.128 samples/sec
+epoch 3, total_step 49560, total loss is 22.41 , inference loss is 11.80, weight deacy loss is 10.61, training accuracy is 0.343750, time 56.099 samples/sec
+epoch 3, total_step 49580, total loss is 25.40 , inference loss is 14.79, weight deacy loss is 10.61, training accuracy is 0.218750, time 56.203 samples/sec
+epoch 3, total_step 49600, total loss is 25.79 , inference loss is 15.18, weight deacy loss is 10.61, training accuracy is 0.187500, time 55.334 samples/sec
+epoch 3, total_step 49620, total loss is 22.31 , inference loss is 11.71, weight deacy loss is 10.61, training accuracy is 0.250000, time 55.209 samples/sec
+epoch 3, total_step 49640, total loss is 22.42 , inference loss is 11.81, weight deacy loss is 10.60, training accuracy is 0.218750, time 56.098 samples/sec
+epoch 3, total_step 49660, total loss is 22.61 , inference loss is 12.01, weight deacy loss is 10.60, training accuracy is 0.281250, time 55.894 samples/sec
+epoch 3, total_step 49680, total loss is 23.67 , inference loss is 13.07, weight deacy loss is 10.60, training accuracy is 0.343750, time 54.382 samples/sec
+epoch 3, total_step 49700, total loss is 25.11 , inference loss is 14.51, weight deacy loss is 10.60, training accuracy is 0.250000, time 55.625 samples/sec
+epoch 3, total_step 49720, total loss is 25.70 , inference loss is 15.10, weight deacy loss is 10.60, training accuracy is 0.125000, time 55.887 samples/sec
+epoch 3, total_step 49740, total loss is 24.50 , inference loss is 13.90, weight deacy loss is 10.60, training accuracy is 0.093750, time 55.659 samples/sec
+epoch 3, total_step 49760, total loss is 28.84 , inference loss is 18.24, weight deacy loss is 10.60, training accuracy is 0.156250, time 56.382 samples/sec
+epoch 3, total_step 49780, total loss is 24.83 , inference loss is 14.23, weight deacy loss is 10.60, training accuracy is 0.125000, time 55.582 samples/sec
+epoch 3, total_step 49800, total loss is 23.63 , inference loss is 13.04, weight deacy loss is 10.59, training accuracy is 0.250000, time 55.666 samples/sec
+epoch 3, total_step 49820, total loss is 23.82 , inference loss is 13.23, weight deacy loss is 10.59, training accuracy is 0.125000, time 56.145 samples/sec
+epoch 3, total_step 49840, total loss is 29.02 , inference loss is 18.43, weight deacy loss is 10.59, training accuracy is 0.156250, time 56.117 samples/sec
+epoch 3, total_step 49860, total loss is 26.37 , inference loss is 15.78, weight deacy loss is 10.59, training accuracy is 0.125000, time 54.822 samples/sec
+epoch 3, total_step 49880, total loss is 22.69 , inference loss is 12.10, weight deacy loss is 10.59, training accuracy is 0.375000, time 55.976 samples/sec
+epoch 3, total_step 49900, total loss is 25.23 , inference loss is 14.64, weight deacy loss is 10.59, training accuracy is 0.187500, time 56.279 samples/sec
+epoch 3, total_step 49920, total loss is 29.20 , inference loss is 18.61, weight deacy loss is 10.59, training accuracy is 0.093750, time 56.577 samples/sec
+epoch 3, total_step 49940, total loss is 24.27 , inference loss is 13.68, weight deacy loss is 10.58, training accuracy is 0.187500, time 56.209 samples/sec
+epoch 3, total_step 49960, total loss is 24.54 , inference loss is 13.96, weight deacy loss is 10.58, training accuracy is 0.156250, time 56.550 samples/sec
+epoch 3, total_step 49980, total loss is 26.38 , inference loss is 15.80, weight deacy loss is 10.58, training accuracy is 0.125000, time 56.263 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.60326300000004
+best_threshold_index 121 0.9835185185185186
+best_threshold_index 121 0.9824074074074074
+best_threshold_index 115 0.9837037037037037
+best_threshold_index 121 0.9838888888888889
+best_threshold_index 116 0.9842592592592593
+best_threshold_index 121 0.9829629629629629
+best_threshold_index 121 0.9833333333333333
+best_threshold_index 117 0.9825925925925926
+best_threshold_index 121 0.9816666666666667
+best_threshold_index 121 0.982037037037037
+[lfw][50000]XNorm: 21.925767
+[lfw][50000]Accuracy-Flip: 0.98133+-0.00836
+testing verification..
+(14000, 512)
+infer time 85.904047
+best_threshold_index 117 0.9823809523809524
+best_threshold_index 117 0.9803174603174604
+best_threshold_index 117 0.9815873015873016
+best_threshold_index 117 0.9804761904761905
+best_threshold_index 117 0.9807936507936508
+best_threshold_index 118 0.9798412698412698
+best_threshold_index 118 0.9823809523809524
+best_threshold_index 117 0.9801587301587301
+best_threshold_index 117 0.9807936507936508
+best_threshold_index 118 0.9817460317460317
+[cfp_ff][50000]XNorm: 21.496391
+[cfp_ff][50000]Accuracy-Flip: 0.98057+-0.00783
+testing verification..
+(14000, 512)
+infer time 86.06371399999993
+best_threshold_index 145 0.8836507936507937
+best_threshold_index 142 0.8822222222222222
+best_threshold_index 145 0.8817460317460317
+best_threshold_index 145 0.8844444444444445
+best_threshold_index 145 0.8838095238095238
+best_threshold_index 142 0.8782539682539683
+best_threshold_index 142 0.8812698412698413
+best_threshold_index 145 0.8807936507936508
+best_threshold_index 145 0.8828571428571429
+best_threshold_index 145 0.8831746031746032
+[cfp_fp][50000]XNorm: 21.943664
+[cfp_fp][50000]Accuracy-Flip: 0.88014+-0.01438
+testing verification..
+(12000, 512)
+infer time 73.877752
+best_threshold_index 144 0.8583333333333333
+best_threshold_index 144 0.8574074074074074
+best_threshold_index 144 0.8592592592592593
+best_threshold_index 144 0.8607407407407407
+best_threshold_index 144 0.8640740740740741
+best_threshold_index 144 0.8587037037037037
+best_threshold_index 144 0.8579629629629629
+best_threshold_index 144 0.8644444444444445
+best_threshold_index 146 0.8625925925925926
+best_threshold_index 144 0.8618518518518519
+[agedb_30][50000]XNorm: 21.399574
+[agedb_30][50000]Accuracy-Flip: 0.85950+-0.02269
+test accuracy is: 0.9813333333333333
+epoch 3, total_step 50000, total loss is 24.13 , inference loss is 13.55, weight deacy loss is 10.58, training accuracy is 0.218750, time 54.658 samples/sec
+epoch 3, total_step 50020, total loss is 28.28 , inference loss is 17.70, weight deacy loss is 10.58, training accuracy is 0.125000, time 56.183 samples/sec
+epoch 3, total_step 50040, total loss is 22.93 , inference loss is 12.35, weight deacy loss is 10.58, training accuracy is 0.218750, time 56.300 samples/sec
+epoch 3, total_step 50060, total loss is 25.46 , inference loss is 14.89, weight deacy loss is 10.58, training accuracy is 0.125000, time 56.430 samples/sec
+epoch 3, total_step 50080, total loss is 23.02 , inference loss is 12.45, weight deacy loss is 10.58, training accuracy is 0.250000, time 56.084 samples/sec
+epoch 3, total_step 50100, total loss is 26.22 , inference loss is 15.65, weight deacy loss is 10.57, training accuracy is 0.156250, time 56.095 samples/sec
+epoch 3, total_step 50120, total loss is 25.44 , inference loss is 14.86, weight deacy loss is 10.57, training accuracy is 0.250000, time 55.536 samples/sec
+epoch 3, total_step 50140, total loss is 24.07 , inference loss is 13.49, weight deacy loss is 10.57, training accuracy is 0.250000, time 56.213 samples/sec
+epoch 3, total_step 50160, total loss is 26.97 , inference loss is 16.40, weight deacy loss is 10.57, training accuracy is 0.406250, time 56.548 samples/sec
+epoch 3, total_step 50180, total loss is 24.96 , inference loss is 14.39, weight deacy loss is 10.57, training accuracy is 0.250000, time 55.070 samples/sec
+epoch 3, total_step 50200, total loss is 23.69 , inference loss is 13.13, weight deacy loss is 10.57, training accuracy is 0.187500, time 56.369 samples/sec
+epoch 3, total_step 50220, total loss is 25.27 , inference loss is 14.70, weight deacy loss is 10.57, training accuracy is 0.156250, time 55.681 samples/sec
+epoch 3, total_step 50240, total loss is 24.67 , inference loss is 14.11, weight deacy loss is 10.57, training accuracy is 0.093750, time 55.611 samples/sec
+epoch 3, total_step 50260, total loss is 30.42 , inference loss is 19.85, weight deacy loss is 10.56, training accuracy is 0.093750, time 56.401 samples/sec
+epoch 3, total_step 50280, total loss is 24.73 , inference loss is 14.16, weight deacy loss is 10.56, training accuracy is 0.125000, time 56.981 samples/sec
+epoch 3, total_step 50300, total loss is 27.35 , inference loss is 16.79, weight deacy loss is 10.56, training accuracy is 0.218750, time 55.595 samples/sec
+epoch 3, total_step 50320, total loss is 25.19 , inference loss is 14.63, weight deacy loss is 10.56, training accuracy is 0.281250, time 56.657 samples/sec
+epoch 3, total_step 50340, total loss is 21.52 , inference loss is 10.97, weight deacy loss is 10.56, training accuracy is 0.218750, time 55.826 samples/sec
+epoch 3, total_step 50360, total loss is 18.06 , inference loss is 7.51, weight deacy loss is 10.56, training accuracy is 0.375000, time 55.896 samples/sec
+epoch 3, total_step 50380, total loss is 27.92 , inference loss is 17.37, weight deacy loss is 10.56, training accuracy is 0.093750, time 56.370 samples/sec
+epoch 3, total_step 50400, total loss is 22.24 , inference loss is 11.68, weight deacy loss is 10.55, training accuracy is 0.250000, time 55.737 samples/sec
+epoch 3, total_step 50420, total loss is 24.95 , inference loss is 14.40, weight deacy loss is 10.55, training accuracy is 0.187500, time 55.657 samples/sec
+epoch 3, total_step 50440, total loss is 24.81 , inference loss is 14.26, weight deacy loss is 10.55, training accuracy is 0.218750, time 56.262 samples/sec
+epoch 3, total_step 50460, total loss is 21.60 , inference loss is 11.05, weight deacy loss is 10.55, training accuracy is 0.281250, time 55.341 samples/sec
+epoch 3, total_step 50480, total loss is 25.51 , inference loss is 14.96, weight deacy loss is 10.55, training accuracy is 0.093750, time 56.537 samples/sec
+epoch 3, total_step 50500, total loss is 22.97 , inference loss is 12.42, weight deacy loss is 10.55, training accuracy is 0.125000, time 55.423 samples/sec
+epoch 3, total_step 50520, total loss is 25.57 , inference loss is 15.03, weight deacy loss is 10.55, training accuracy is 0.250000, time 56.854 samples/sec
+epoch 3, total_step 50540, total loss is 21.67 , inference loss is 11.13, weight deacy loss is 10.55, training accuracy is 0.312500, time 55.860 samples/sec
+epoch 3, total_step 50560, total loss is 23.85 , inference loss is 13.31, weight deacy loss is 10.54, training accuracy is 0.218750, time 56.084 samples/sec
+epoch 3, total_step 50580, total loss is 23.93 , inference loss is 13.38, weight deacy loss is 10.54, training accuracy is 0.250000, time 56.696 samples/sec
+epoch 3, total_step 50600, total loss is 23.49 , inference loss is 12.95, weight deacy loss is 10.54, training accuracy is 0.375000, time 55.680 samples/sec
+epoch 3, total_step 50620, total loss is 24.01 , inference loss is 13.47, weight deacy loss is 10.54, training accuracy is 0.406250, time 56.698 samples/sec
+epoch 3, total_step 50640, total loss is 26.15 , inference loss is 15.61, weight deacy loss is 10.54, training accuracy is 0.156250, time 56.887 samples/sec
+epoch 3, total_step 50660, total loss is 28.36 , inference loss is 17.82, weight deacy loss is 10.54, training accuracy is 0.093750, time 56.732 samples/sec
+epoch 3, total_step 50680, total loss is 24.03 , inference loss is 13.49, weight deacy loss is 10.54, training accuracy is 0.187500, time 56.719 samples/sec
+epoch 3, total_step 50700, total loss is 23.26 , inference loss is 12.72, weight deacy loss is 10.54, training accuracy is 0.250000, time 53.800 samples/sec
+epoch 3, total_step 50720, total loss is 25.82 , inference loss is 15.29, weight deacy loss is 10.53, training accuracy is 0.156250, time 55.638 samples/sec
+epoch 3, total_step 50740, total loss is 21.35 , inference loss is 10.82, weight deacy loss is 10.53, training accuracy is 0.250000, time 56.106 samples/sec
+epoch 3, total_step 50760, total loss is 24.88 , inference loss is 14.35, weight deacy loss is 10.53, training accuracy is 0.218750, time 57.799 samples/sec
+epoch 3, total_step 50780, total loss is 25.46 , inference loss is 14.93, weight deacy loss is 10.53, training accuracy is 0.187500, time 56.199 samples/sec
+epoch 3, total_step 50800, total loss is 19.87 , inference loss is 9.34, weight deacy loss is 10.53, training accuracy is 0.187500, time 56.012 samples/sec
+epoch 3, total_step 50820, total loss is 18.83 , inference loss is 8.30, weight deacy loss is 10.53, training accuracy is 0.312500, time 56.698 samples/sec
+epoch 3, total_step 50840, total loss is 22.17 , inference loss is 11.64, weight deacy loss is 10.53, training accuracy is 0.218750, time 56.462 samples/sec
+epoch 3, total_step 50860, total loss is 25.43 , inference loss is 14.91, weight deacy loss is 10.52, training accuracy is 0.062500, time 57.334 samples/sec
+epoch 3, total_step 50880, total loss is 21.58 , inference loss is 11.05, weight deacy loss is 10.52, training accuracy is 0.250000, time 57.243 samples/sec
+epoch 3, total_step 50900, total loss is 21.86 , inference loss is 11.34, weight deacy loss is 10.52, training accuracy is 0.156250, time 56.043 samples/sec
+epoch 3, total_step 50920, total loss is 21.98 , inference loss is 11.46, weight deacy loss is 10.52, training accuracy is 0.281250, time 55.927 samples/sec
+epoch 3, total_step 50940, total loss is 26.78 , inference loss is 16.26, weight deacy loss is 10.52, training accuracy is 0.281250, time 55.664 samples/sec
+epoch 3, total_step 50960, total loss is 23.84 , inference loss is 13.33, weight deacy loss is 10.52, training accuracy is 0.250000, time 55.543 samples/sec
+epoch 3, total_step 50980, total loss is 24.32 , inference loss is 13.80, weight deacy loss is 10.52, training accuracy is 0.093750, time 56.147 samples/sec
+epoch 3, total_step 51000, total loss is 25.19 , inference loss is 14.67, weight deacy loss is 10.52, training accuracy is 0.250000, time 56.590 samples/sec
+epoch 3, total_step 51020, total loss is 21.05 , inference loss is 10.53, weight deacy loss is 10.51, training accuracy is 0.343750, time 55.628 samples/sec
+epoch 3, total_step 51040, total loss is 21.45 , inference loss is 10.94, weight deacy loss is 10.51, training accuracy is 0.312500, time 56.518 samples/sec
+epoch 3, total_step 51060, total loss is 30.12 , inference loss is 19.61, weight deacy loss is 10.51, training accuracy is 0.187500, time 55.985 samples/sec
+epoch 3, total_step 51080, total loss is 24.89 , inference loss is 14.38, weight deacy loss is 10.51, training accuracy is 0.312500, time 56.332 samples/sec
+epoch 3, total_step 51100, total loss is 22.66 , inference loss is 12.15, weight deacy loss is 10.51, training accuracy is 0.187500, time 56.345 samples/sec
+epoch 3, total_step 51120, total loss is 23.46 , inference loss is 12.95, weight deacy loss is 10.51, training accuracy is 0.218750, time 56.444 samples/sec
+epoch 3, total_step 51140, total loss is 22.15 , inference loss is 11.64, weight deacy loss is 10.51, training accuracy is 0.250000, time 53.718 samples/sec
+epoch 3, total_step 51160, total loss is 22.84 , inference loss is 12.33, weight deacy loss is 10.51, training accuracy is 0.187500, time 55.141 samples/sec
+epoch 3, total_step 51180, total loss is 24.63 , inference loss is 14.13, weight deacy loss is 10.50, training accuracy is 0.187500, time 56.217 samples/sec
+epoch 3, total_step 51200, total loss is 20.62 , inference loss is 10.11, weight deacy loss is 10.50, training accuracy is 0.218750, time 55.694 samples/sec
+epoch 3, total_step 51220, total loss is 24.92 , inference loss is 14.42, weight deacy loss is 10.50, training accuracy is 0.312500, time 55.913 samples/sec
+epoch 3, total_step 51240, total loss is 24.18 , inference loss is 13.68, weight deacy loss is 10.50, training accuracy is 0.156250, time 56.213 samples/sec
+epoch 3, total_step 51260, total loss is 24.52 , inference loss is 14.02, weight deacy loss is 10.50, training accuracy is 0.281250, time 55.627 samples/sec
+epoch 3, total_step 51280, total loss is 24.11 , inference loss is 13.61, weight deacy loss is 10.50, training accuracy is 0.312500, time 55.546 samples/sec
+epoch 3, total_step 51300, total loss is 26.61 , inference loss is 16.12, weight deacy loss is 10.50, training accuracy is 0.156250, time 55.570 samples/sec
+epoch 3, total_step 51320, total loss is 25.02 , inference loss is 14.52, weight deacy loss is 10.49, training accuracy is 0.218750, time 56.433 samples/sec
+epoch 3, total_step 51340, total loss is 21.91 , inference loss is 11.42, weight deacy loss is 10.49, training accuracy is 0.343750, time 56.218 samples/sec
+epoch 3, total_step 51360, total loss is 28.52 , inference loss is 18.03, weight deacy loss is 10.49, training accuracy is 0.156250, time 56.715 samples/sec
+epoch 3, total_step 51380, total loss is 21.02 , inference loss is 10.53, weight deacy loss is 10.49, training accuracy is 0.312500, time 57.121 samples/sec
+epoch 3, total_step 51400, total loss is 25.10 , inference loss is 14.61, weight deacy loss is 10.49, training accuracy is 0.156250, time 55.102 samples/sec
+epoch 3, total_step 51420, total loss is 27.39 , inference loss is 16.90, weight deacy loss is 10.49, training accuracy is 0.156250, time 55.746 samples/sec
+epoch 3, total_step 51440, total loss is 23.35 , inference loss is 12.86, weight deacy loss is 10.49, training accuracy is 0.218750, time 55.784 samples/sec
+epoch 3, total_step 51460, total loss is 19.23 , inference loss is 8.74, weight deacy loss is 10.49, training accuracy is 0.531250, time 56.371 samples/sec
+epoch 3, total_step 51480, total loss is 22.48 , inference loss is 11.99, weight deacy loss is 10.48, training accuracy is 0.343750, time 55.718 samples/sec
+epoch 3, total_step 51500, total loss is 20.18 , inference loss is 9.70, weight deacy loss is 10.48, training accuracy is 0.250000, time 56.227 samples/sec
+epoch 3, total_step 51520, total loss is 24.35 , inference loss is 13.87, weight deacy loss is 10.48, training accuracy is 0.312500, time 56.131 samples/sec
+epoch 3, total_step 51540, total loss is 27.91 , inference loss is 17.43, weight deacy loss is 10.48, training accuracy is 0.187500, time 56.958 samples/sec
+epoch 3, total_step 51560, total loss is 22.47 , inference loss is 11.99, weight deacy loss is 10.48, training accuracy is 0.218750, time 56.096 samples/sec
+epoch 3, total_step 51580, total loss is 23.07 , inference loss is 12.60, weight deacy loss is 10.48, training accuracy is 0.343750, time 55.584 samples/sec
+epoch 3, total_step 51600, total loss is 24.13 , inference loss is 13.65, weight deacy loss is 10.48, training accuracy is 0.281250, time 55.293 samples/sec
+epoch 3, total_step 51620, total loss is 24.02 , inference loss is 13.55, weight deacy loss is 10.48, training accuracy is 0.187500, time 55.941 samples/sec
+epoch 3, total_step 51640, total loss is 29.02 , inference loss is 18.55, weight deacy loss is 10.47, training accuracy is 0.187500, time 56.318 samples/sec
+epoch 3, total_step 51660, total loss is 25.28 , inference loss is 14.81, weight deacy loss is 10.47, training accuracy is 0.312500, time 56.212 samples/sec
+epoch 3, total_step 51680, total loss is 26.37 , inference loss is 15.90, weight deacy loss is 10.47, training accuracy is 0.156250, time 55.586 samples/sec
+epoch 3, total_step 51700, total loss is 22.27 , inference loss is 11.80, weight deacy loss is 10.47, training accuracy is 0.312500, time 56.712 samples/sec
+epoch 3, total_step 51720, total loss is 20.27 , inference loss is 9.81, weight deacy loss is 10.47, training accuracy is 0.312500, time 55.343 samples/sec
+epoch 3, total_step 51740, total loss is 25.27 , inference loss is 14.80, weight deacy loss is 10.47, training accuracy is 0.218750, time 55.862 samples/sec
+epoch 3, total_step 51760, total loss is 23.75 , inference loss is 13.29, weight deacy loss is 10.47, training accuracy is 0.281250, time 56.315 samples/sec
+epoch 3, total_step 51780, total loss is 23.65 , inference loss is 13.18, weight deacy loss is 10.46, training accuracy is 0.156250, time 56.694 samples/sec
+epoch 3, total_step 51800, total loss is 25.96 , inference loss is 15.49, weight deacy loss is 10.46, training accuracy is 0.125000, time 56.714 samples/sec
+epoch 3, total_step 51820, total loss is 23.61 , inference loss is 13.15, weight deacy loss is 10.46, training accuracy is 0.187500, time 56.087 samples/sec
+epoch 3, total_step 51840, total loss is 24.49 , inference loss is 14.03, weight deacy loss is 10.46, training accuracy is 0.281250, time 56.425 samples/sec
+epoch 3, total_step 51860, total loss is 23.88 , inference loss is 13.42, weight deacy loss is 10.46, training accuracy is 0.218750, time 55.945 samples/sec
+epoch 3, total_step 51880, total loss is 22.51 , inference loss is 12.05, weight deacy loss is 10.46, training accuracy is 0.250000, time 56.666 samples/sec
+epoch 3, total_step 51900, total loss is 24.28 , inference loss is 13.82, weight deacy loss is 10.46, training accuracy is 0.125000, time 55.793 samples/sec
+epoch 3, total_step 51920, total loss is 23.62 , inference loss is 13.16, weight deacy loss is 10.46, training accuracy is 0.250000, time 55.960 samples/sec
+epoch 3, total_step 51940, total loss is 23.65 , inference loss is 13.19, weight deacy loss is 10.45, training accuracy is 0.281250, time 57.296 samples/sec
+epoch 3, total_step 51960, total loss is 23.41 , inference loss is 12.96, weight deacy loss is 10.45, training accuracy is 0.187500, time 56.367 samples/sec
+epoch 3, total_step 51980, total loss is 21.52 , inference loss is 11.06, weight deacy loss is 10.45, training accuracy is 0.250000, time 56.818 samples/sec
+testing verification..
+(12000, 512)
+infer time 75.07564399999994
+best_threshold_index 110 0.9842592592592593
+best_threshold_index 110 0.9833333333333333
+best_threshold_index 112 0.9846296296296296
+best_threshold_index 110 0.9844444444444445
+best_threshold_index 110 0.9840740740740741
+best_threshold_index 110 0.9835185185185186
+best_threshold_index 110 0.9838888888888889
+best_threshold_index 110 0.9829629629629629
+best_threshold_index 110 0.9831481481481481
+best_threshold_index 110 0.9825925925925926
+[lfw][52000]XNorm: 22.154313
+[lfw][52000]Accuracy-Flip: 0.98350+-0.00580
+testing verification..
+(14000, 512)
+infer time 87.12550500000003
+best_threshold_index 116 0.9801587301587301
+best_threshold_index 117 0.9784126984126984
+best_threshold_index 117 0.9785714285714285
+best_threshold_index 116 0.9782539682539683
+best_threshold_index 117 0.9787301587301588
+best_threshold_index 117 0.9780952380952381
+best_threshold_index 117 0.9787301587301588
+best_threshold_index 115 0.9782539682539683
+best_threshold_index 117 0.9792063492063492
+best_threshold_index 117 0.9801587301587301
+[cfp_ff][52000]XNorm: 21.583516
+[cfp_ff][52000]Accuracy-Flip: 0.97814+-0.00636
+testing verification..
+(14000, 512)
+infer time 87.24456999999997
+best_threshold_index 134 0.8971428571428571
+best_threshold_index 134 0.8960317460317461
+best_threshold_index 134 0.8955555555555555
+best_threshold_index 134 0.8958730158730158
+best_threshold_index 134 0.8980952380952381
+best_threshold_index 129 0.8914285714285715
+best_threshold_index 129 0.8928571428571429
+best_threshold_index 134 0.8931746031746032
+best_threshold_index 134 0.8955555555555555
+best_threshold_index 134 0.8963492063492063
+[cfp_fp][52000]XNorm: 21.186407
+[cfp_fp][52000]Accuracy-Flip: 0.89343+-0.01599
+testing verification..
+(12000, 512)
+infer time 75.12434600000014
+best_threshold_index 134 0.8666666666666667
+best_threshold_index 135 0.8642592592592593
+best_threshold_index 133 0.8668518518518519
+best_threshold_index 133 0.8661111111111112
+best_threshold_index 133 0.8725925925925926
+best_threshold_index 133 0.8672222222222222
+best_threshold_index 133 0.8655555555555555
+best_threshold_index 135 0.8742592592592593
+best_threshold_index 134 0.8694444444444445
+best_threshold_index 133 0.8679629629629629
+[agedb_30][52000]XNorm: 21.940746
+[agedb_30][52000]Accuracy-Flip: 0.86517+-0.02706
+test accuracy is: 0.9835
+epoch 3, total_step 52000, total loss is 26.00 , inference loss is 15.55, weight deacy loss is 10.45, training accuracy is 0.125000, time 52.730 samples/sec
+epoch 3, total_step 52020, total loss is 26.97 , inference loss is 16.52, weight deacy loss is 10.45, training accuracy is 0.156250, time 56.441 samples/sec
+epoch 3, total_step 52040, total loss is 28.15 , inference loss is 17.70, weight deacy loss is 10.45, training accuracy is 0.187500, time 56.247 samples/sec
+epoch 3, total_step 52060, total loss is 24.34 , inference loss is 13.89, weight deacy loss is 10.45, training accuracy is 0.062500, time 55.643 samples/sec
+epoch 3, total_step 52080, total loss is 25.92 , inference loss is 15.47, weight deacy loss is 10.45, training accuracy is 0.031250, time 54.981 samples/sec
+epoch 3, total_step 52100, total loss is 24.00 , inference loss is 13.56, weight deacy loss is 10.44, training accuracy is 0.218750, time 56.498 samples/sec
+epoch 3, total_step 52120, total loss is 26.10 , inference loss is 15.66, weight deacy loss is 10.44, training accuracy is 0.218750, time 56.462 samples/sec
+epoch 3, total_step 52140, total loss is 24.86 , inference loss is 14.41, weight deacy loss is 10.44, training accuracy is 0.187500, time 56.362 samples/sec
+epoch 3, total_step 52160, total loss is 25.69 , inference loss is 15.25, weight deacy loss is 10.44, training accuracy is 0.125000, time 56.178 samples/sec
+epoch 3, total_step 52180, total loss is 23.82 , inference loss is 13.38, weight deacy loss is 10.44, training accuracy is 0.281250, time 55.614 samples/sec
+epoch 3, total_step 52200, total loss is 21.48 , inference loss is 11.04, weight deacy loss is 10.44, training accuracy is 0.156250, time 56.493 samples/sec
+epoch 3, total_step 52220, total loss is 25.08 , inference loss is 14.65, weight deacy loss is 10.44, training accuracy is 0.156250, time 55.605 samples/sec
+epoch 3, total_step 52240, total loss is 22.53 , inference loss is 12.09, weight deacy loss is 10.43, training accuracy is 0.250000, time 55.190 samples/sec
+epoch 3, total_step 52260, total loss is 20.25 , inference loss is 9.82, weight deacy loss is 10.43, training accuracy is 0.312500, time 56.118 samples/sec
+epoch 3, total_step 52280, total loss is 26.23 , inference loss is 15.80, weight deacy loss is 10.43, training accuracy is 0.125000, time 55.895 samples/sec
+epoch 3, total_step 52300, total loss is 23.98 , inference loss is 13.54, weight deacy loss is 10.43, training accuracy is 0.187500, time 55.072 samples/sec
+epoch 3, total_step 52320, total loss is 21.79 , inference loss is 11.36, weight deacy loss is 10.43, training accuracy is 0.406250, time 56.561 samples/sec
+epoch 3, total_step 52340, total loss is 21.32 , inference loss is 10.89, weight deacy loss is 10.43, training accuracy is 0.312500, time 53.376 samples/sec
+epoch 3, total_step 52360, total loss is 24.32 , inference loss is 13.89, weight deacy loss is 10.43, training accuracy is 0.187500, time 55.443 samples/sec
+epoch 3, total_step 52380, total loss is 18.02 , inference loss is 7.60, weight deacy loss is 10.43, training accuracy is 0.281250, time 55.266 samples/sec
+epoch 3, total_step 52400, total loss is 26.71 , inference loss is 16.28, weight deacy loss is 10.42, training accuracy is 0.312500, time 55.961 samples/sec
+epoch 3, total_step 52420, total loss is 25.48 , inference loss is 15.06, weight deacy loss is 10.42, training accuracy is 0.156250, time 56.638 samples/sec
+epoch 3, total_step 52440, total loss is 26.04 , inference loss is 15.62, weight deacy loss is 10.42, training accuracy is 0.250000, time 57.088 samples/sec
+epoch 3, total_step 52460, total loss is 24.01 , inference loss is 13.58, weight deacy loss is 10.42, training accuracy is 0.093750, time 55.619 samples/sec
+epoch 3, total_step 52480, total loss is 26.64 , inference loss is 16.22, weight deacy loss is 10.42, training accuracy is 0.250000, time 56.063 samples/sec
+epoch 3, total_step 52500, total loss is 21.55 , inference loss is 11.13, weight deacy loss is 10.42, training accuracy is 0.250000, time 54.961 samples/sec
+epoch 3, total_step 52520, total loss is 22.05 , inference loss is 11.63, weight deacy loss is 10.42, training accuracy is 0.281250, time 54.818 samples/sec
+epoch 3, total_step 52540, total loss is 25.11 , inference loss is 14.69, weight deacy loss is 10.42, training accuracy is 0.187500, time 54.185 samples/sec
+epoch 3, total_step 52560, total loss is 28.03 , inference loss is 17.61, weight deacy loss is 10.41, training accuracy is 0.218750, time 56.470 samples/sec
+epoch 3, total_step 52580, total loss is 28.43 , inference loss is 18.01, weight deacy loss is 10.41, training accuracy is 0.187500, time 56.266 samples/sec
+epoch 3, total_step 52600, total loss is 24.41 , inference loss is 14.00, weight deacy loss is 10.41, training accuracy is 0.187500, time 55.896 samples/sec
+epoch 3, total_step 52620, total loss is 27.87 , inference loss is 17.46, weight deacy loss is 10.41, training accuracy is 0.156250, time 55.985 samples/sec
+epoch 3, total_step 52640, total loss is 26.59 , inference loss is 16.18, weight deacy loss is 10.41, training accuracy is 0.156250, time 55.450 samples/sec
+epoch 3, total_step 52660, total loss is 24.48 , inference loss is 14.08, weight deacy loss is 10.41, training accuracy is 0.187500, time 55.830 samples/sec
+epoch 3, total_step 52680, total loss is 26.03 , inference loss is 15.62, weight deacy loss is 10.41, training accuracy is 0.218750, time 56.926 samples/sec
+epoch 3, total_step 52700, total loss is 23.02 , inference loss is 12.62, weight deacy loss is 10.41, training accuracy is 0.156250, time 56.284 samples/sec
+epoch 3, total_step 52720, total loss is 22.88 , inference loss is 12.47, weight deacy loss is 10.40, training accuracy is 0.281250, time 55.818 samples/sec
+epoch 3, total_step 52740, total loss is 28.01 , inference loss is 17.61, weight deacy loss is 10.40, training accuracy is 0.093750, time 56.246 samples/sec
+epoch 3, total_step 52760, total loss is 21.39 , inference loss is 10.99, weight deacy loss is 10.40, training accuracy is 0.343750, time 56.799 samples/sec
+epoch 3, total_step 52780, total loss is 24.71 , inference loss is 14.31, weight deacy loss is 10.40, training accuracy is 0.187500, time 56.823 samples/sec
+epoch 3, total_step 52800, total loss is 28.32 , inference loss is 17.92, weight deacy loss is 10.40, training accuracy is 0.125000, time 55.917 samples/sec
+epoch 3, total_step 52820, total loss is 25.53 , inference loss is 15.13, weight deacy loss is 10.40, training accuracy is 0.093750, time 55.710 samples/sec
+epoch 3, total_step 52840, total loss is 23.98 , inference loss is 13.59, weight deacy loss is 10.40, training accuracy is 0.156250, time 56.730 samples/sec
+epoch 3, total_step 52860, total loss is 22.77 , inference loss is 12.38, weight deacy loss is 10.40, training accuracy is 0.218750, time 56.051 samples/sec
+epoch 3, total_step 52880, total loss is 24.16 , inference loss is 13.77, weight deacy loss is 10.39, training accuracy is 0.250000, time 53.913 samples/sec
+epoch 3, total_step 52900, total loss is 27.95 , inference loss is 17.56, weight deacy loss is 10.39, training accuracy is 0.187500, time 56.464 samples/sec
+epoch 3, total_step 52920, total loss is 22.86 , inference loss is 12.47, weight deacy loss is 10.39, training accuracy is 0.187500, time 54.479 samples/sec
+epoch 3, total_step 52940, total loss is 23.17 , inference loss is 12.78, weight deacy loss is 10.39, training accuracy is 0.281250, time 56.438 samples/sec
+epoch 3, total_step 52960, total loss is 23.88 , inference loss is 13.50, weight deacy loss is 10.39, training accuracy is 0.125000, time 56.097 samples/sec
+epoch 3, total_step 52980, total loss is 25.59 , inference loss is 15.20, weight deacy loss is 10.39, training accuracy is 0.156250, time 56.890 samples/sec
+epoch 3, total_step 53000, total loss is 27.49 , inference loss is 17.11, weight deacy loss is 10.39, training accuracy is 0.250000, time 55.993 samples/sec
+epoch 3, total_step 53020, total loss is 23.33 , inference loss is 12.94, weight deacy loss is 10.39, training accuracy is 0.125000, time 56.801 samples/sec
+epoch 3, total_step 53040, total loss is 19.86 , inference loss is 9.47, weight deacy loss is 10.38, training accuracy is 0.406250, time 56.522 samples/sec
+epoch 3, total_step 53060, total loss is 20.94 , inference loss is 10.56, weight deacy loss is 10.38, training accuracy is 0.093750, time 56.335 samples/sec
+epoch 3, total_step 53080, total loss is 19.51 , inference loss is 9.13, weight deacy loss is 10.38, training accuracy is 0.343750, time 56.324 samples/sec
+epoch 3, total_step 53100, total loss is 20.10 , inference loss is 9.72, weight deacy loss is 10.38, training accuracy is 0.218750, time 56.014 samples/sec
+epoch 3, total_step 53120, total loss is 28.23 , inference loss is 17.85, weight deacy loss is 10.38, training accuracy is 0.125000, time 55.847 samples/sec
+epoch 3, total_step 53140, total loss is 23.18 , inference loss is 12.80, weight deacy loss is 10.38, training accuracy is 0.156250, time 56.248 samples/sec
+epoch 3, total_step 53160, total loss is 23.80 , inference loss is 13.43, weight deacy loss is 10.38, training accuracy is 0.187500, time 55.594 samples/sec
+epoch 3, total_step 53180, total loss is 22.92 , inference loss is 12.55, weight deacy loss is 10.38, training accuracy is 0.156250, time 55.989 samples/sec
+epoch 3, total_step 53200, total loss is 24.91 , inference loss is 14.54, weight deacy loss is 10.37, training accuracy is 0.093750, time 56.192 samples/sec
+epoch 3, total_step 53220, total loss is 20.30 , inference loss is 9.92, weight deacy loss is 10.37, training accuracy is 0.281250, time 56.180 samples/sec
+epoch 3, total_step 53240, total loss is 19.84 , inference loss is 9.47, weight deacy loss is 10.37, training accuracy is 0.312500, time 56.499 samples/sec
+epoch 3, total_step 53260, total loss is 23.56 , inference loss is 13.19, weight deacy loss is 10.37, training accuracy is 0.250000, time 56.050 samples/sec
+epoch 3, total_step 53280, total loss is 24.72 , inference loss is 14.35, weight deacy loss is 10.37, training accuracy is 0.218750, time 56.738 samples/sec
+epoch 3, total_step 53300, total loss is 21.69 , inference loss is 11.32, weight deacy loss is 10.37, training accuracy is 0.156250, time 55.996 samples/sec
+epoch 3, total_step 53320, total loss is 21.82 , inference loss is 11.45, weight deacy loss is 10.37, training accuracy is 0.218750, time 56.465 samples/sec
+epoch 3, total_step 53340, total loss is 23.29 , inference loss is 12.93, weight deacy loss is 10.37, training accuracy is 0.187500, time 56.055 samples/sec
+epoch 3, total_step 53360, total loss is 25.05 , inference loss is 14.68, weight deacy loss is 10.36, training accuracy is 0.187500, time 56.783 samples/sec
+epoch 3, total_step 53380, total loss is 23.95 , inference loss is 13.58, weight deacy loss is 10.36, training accuracy is 0.187500, time 56.838 samples/sec
+epoch 3, total_step 53400, total loss is 26.18 , inference loss is 15.82, weight deacy loss is 10.36, training accuracy is 0.312500, time 55.715 samples/sec
+epoch 3, total_step 53420, total loss is 23.97 , inference loss is 13.61, weight deacy loss is 10.36, training accuracy is 0.218750, time 56.353 samples/sec
+epoch 3, total_step 53440, total loss is 21.61 , inference loss is 11.26, weight deacy loss is 10.36, training accuracy is 0.281250, time 56.853 samples/sec
+epoch 3, total_step 53460, total loss is 28.04 , inference loss is 17.68, weight deacy loss is 10.36, training accuracy is 0.156250, time 55.147 samples/sec
+epoch 3, total_step 53480, total loss is 27.93 , inference loss is 17.58, weight deacy loss is 10.36, training accuracy is 0.281250, time 56.411 samples/sec
+epoch 3, total_step 53500, total loss is 25.50 , inference loss is 15.15, weight deacy loss is 10.35, training accuracy is 0.281250, time 56.602 samples/sec
+epoch 3, total_step 53520, total loss is 22.70 , inference loss is 12.35, weight deacy loss is 10.35, training accuracy is 0.375000, time 56.709 samples/sec
+epoch 3, total_step 53540, total loss is 24.38 , inference loss is 14.02, weight deacy loss is 10.35, training accuracy is 0.343750, time 53.665 samples/sec
+epoch 3, total_step 53560, total loss is 20.52 , inference loss is 10.17, weight deacy loss is 10.35, training accuracy is 0.281250, time 56.389 samples/sec
+epoch 3, total_step 53580, total loss is 23.83 , inference loss is 13.48, weight deacy loss is 10.35, training accuracy is 0.187500, time 55.822 samples/sec
+epoch 3, total_step 53600, total loss is 26.09 , inference loss is 15.74, weight deacy loss is 10.35, training accuracy is 0.250000, time 54.887 samples/sec
+epoch 3, total_step 53620, total loss is 24.42 , inference loss is 14.07, weight deacy loss is 10.35, training accuracy is 0.343750, time 56.259 samples/sec
+epoch 3, total_step 53640, total loss is 22.14 , inference loss is 11.80, weight deacy loss is 10.35, training accuracy is 0.281250, time 53.777 samples/sec
+epoch 3, total_step 53660, total loss is 20.65 , inference loss is 10.30, weight deacy loss is 10.34, training accuracy is 0.281250, time 56.377 samples/sec
+epoch 3, total_step 53680, total loss is 25.73 , inference loss is 15.39, weight deacy loss is 10.34, training accuracy is 0.218750, time 55.679 samples/sec
+epoch 3, total_step 53700, total loss is 27.44 , inference loss is 17.10, weight deacy loss is 10.34, training accuracy is 0.218750, time 56.073 samples/sec
+epoch 3, total_step 53720, total loss is 22.40 , inference loss is 12.06, weight deacy loss is 10.34, training accuracy is 0.218750, time 55.174 samples/sec
+epoch 3, total_step 53740, total loss is 24.12 , inference loss is 13.78, weight deacy loss is 10.34, training accuracy is 0.218750, time 55.565 samples/sec
+epoch 3, total_step 53760, total loss is 18.77 , inference loss is 8.44, weight deacy loss is 10.34, training accuracy is 0.406250, time 56.235 samples/sec
+epoch 3, total_step 53780, total loss is 21.55 , inference loss is 11.21, weight deacy loss is 10.34, training accuracy is 0.218750, time 56.106 samples/sec
+epoch 3, total_step 53800, total loss is 24.02 , inference loss is 13.69, weight deacy loss is 10.34, training accuracy is 0.312500, time 55.407 samples/sec
+epoch 3, total_step 53820, total loss is 25.99 , inference loss is 15.65, weight deacy loss is 10.33, training accuracy is 0.156250, time 56.385 samples/sec
+epoch 3, total_step 53840, total loss is 26.08 , inference loss is 15.75, weight deacy loss is 10.33, training accuracy is 0.187500, time 56.695 samples/sec
+epoch 3, total_step 53860, total loss is 22.43 , inference loss is 12.10, weight deacy loss is 10.33, training accuracy is 0.187500, time 56.108 samples/sec
+epoch 3, total_step 53880, total loss is 25.09 , inference loss is 14.76, weight deacy loss is 10.33, training accuracy is 0.125000, time 55.421 samples/sec
+epoch 3, total_step 53900, total loss is 22.33 , inference loss is 12.00, weight deacy loss is 10.33, training accuracy is 0.250000, time 55.935 samples/sec
+epoch 3, total_step 53920, total loss is 18.35 , inference loss is 8.03, weight deacy loss is 10.33, training accuracy is 0.250000, time 56.165 samples/sec
+epoch 3, total_step 53940, total loss is 23.13 , inference loss is 12.81, weight deacy loss is 10.33, training accuracy is 0.187500, time 56.056 samples/sec
+epoch 3, total_step 53960, total loss is 21.24 , inference loss is 10.92, weight deacy loss is 10.32, training accuracy is 0.187500, time 55.434 samples/sec
+epoch 3, total_step 53980, total loss is 21.61 , inference loss is 11.28, weight deacy loss is 10.32, training accuracy is 0.312500, time 56.828 samples/sec
+testing verification..
+(12000, 512)
+infer time 75.14281200000006
+best_threshold_index 117 0.9844444444444445
+best_threshold_index 117 0.9835185185185186
+best_threshold_index 117 0.9846296296296296
+best_threshold_index 117 0.9846296296296296
+best_threshold_index 117 0.9837037037037037
+best_threshold_index 117 0.9831481481481481
+best_threshold_index 112 0.9837037037037037
+best_threshold_index 117 0.9833333333333333
+best_threshold_index 117 0.9824074074074074
+best_threshold_index 117 0.9831481481481481
+[lfw][54000]XNorm: 22.884348
+[lfw][54000]Accuracy-Flip: 0.98300+-0.00653
+testing verification..
+(14000, 512)
+infer time 87.45654999999996
+best_threshold_index 120 0.9792063492063492
+best_threshold_index 114 0.9773015873015873
+best_threshold_index 115 0.9784126984126984
+best_threshold_index 115 0.976984126984127
+best_threshold_index 120 0.9771428571428571
+best_threshold_index 120 0.9768253968253968
+best_threshold_index 115 0.9779365079365079
+best_threshold_index 115 0.9766666666666667
+best_threshold_index 114 0.9774603174603175
+best_threshold_index 115 0.9784126984126984
+[cfp_ff][54000]XNorm: 22.259668
+[cfp_ff][54000]Accuracy-Flip: 0.97629+-0.00706
+testing verification..
+(14000, 512)
+infer time 87.1533139999999
+best_threshold_index 139 0.8896825396825396
+best_threshold_index 140 0.8885714285714286
+best_threshold_index 139 0.8879365079365079
+best_threshold_index 139 0.8895238095238095
+best_threshold_index 143 0.8912698412698413
+best_threshold_index 139 0.8873015873015873
+best_threshold_index 143 0.8879365079365079
+best_threshold_index 143 0.888095238095238
+best_threshold_index 143 0.8882539682539683
+best_threshold_index 139 0.8896825396825396
+[cfp_fp][54000]XNorm: 22.302265
+[cfp_fp][54000]Accuracy-Flip: 0.88600+-0.01014
+testing verification..
+(12000, 512)
+infer time 74.76576699999995
+best_threshold_index 143 0.8790740740740741
+best_threshold_index 143 0.8781481481481481
+best_threshold_index 143 0.8822222222222222
+best_threshold_index 143 0.8814814814814815
+best_threshold_index 143 0.8857407407407407
+best_threshold_index 143 0.8807407407407407
+best_threshold_index 143 0.8785185185185185
+best_threshold_index 143 0.8851851851851852
+best_threshold_index 143 0.882962962962963
+best_threshold_index 143 0.8859259259259259
+[agedb_30][54000]XNorm: 22.705210
+[agedb_30][54000]Accuracy-Flip: 0.88200+-0.02512
+test accuracy is: 0.983
+epoch 3, total_step 54000, total loss is 25.84 , inference loss is 15.52, weight deacy loss is 10.32, training accuracy is 0.187500, time 53.236 samples/sec
+epoch 3, total_step 54020, total loss is 23.78 , inference loss is 13.45, weight deacy loss is 10.32, training accuracy is 0.156250, time 55.882 samples/sec
+epoch 3, total_step 54040, total loss is 26.08 , inference loss is 15.76, weight deacy loss is 10.32, training accuracy is 0.218750, time 55.470 samples/sec
+epoch 3, total_step 54060, total loss is 19.18 , inference loss is 8.86, weight deacy loss is 10.32, training accuracy is 0.281250, time 55.772 samples/sec
+epoch 3, total_step 54080, total loss is 26.23 , inference loss is 15.92, weight deacy loss is 10.32, training accuracy is 0.187500, time 56.426 samples/sec
+epoch 3, total_step 54100, total loss is 26.25 , inference loss is 15.93, weight deacy loss is 10.32, training accuracy is 0.187500, time 56.227 samples/sec
+epoch 3, total_step 54120, total loss is 23.14 , inference loss is 12.83, weight deacy loss is 10.31, training accuracy is 0.250000, time 55.647 samples/sec
+epoch 3, total_step 54140, total loss is 17.85 , inference loss is 7.53, weight deacy loss is 10.31, training accuracy is 0.312500, time 55.940 samples/sec
+epoch 3, total_step 54160, total loss is 27.67 , inference loss is 17.36, weight deacy loss is 10.31, training accuracy is 0.218750, time 55.705 samples/sec
+epoch 3, total_step 54180, total loss is 21.66 , inference loss is 11.35, weight deacy loss is 10.31, training accuracy is 0.281250, time 56.882 samples/sec
+epoch 3, total_step 54200, total loss is 24.85 , inference loss is 14.54, weight deacy loss is 10.31, training accuracy is 0.187500, time 55.718 samples/sec
+epoch 3, total_step 54220, total loss is 24.71 , inference loss is 14.40, weight deacy loss is 10.31, training accuracy is 0.250000, time 56.518 samples/sec
+epoch 3, total_step 54240, total loss is 22.02 , inference loss is 11.72, weight deacy loss is 10.31, training accuracy is 0.250000, time 55.901 samples/sec
+epoch 3, total_step 54260, total loss is 19.53 , inference loss is 9.23, weight deacy loss is 10.31, training accuracy is 0.343750, time 57.684 samples/sec
+epoch 3, total_step 54280, total loss is 23.20 , inference loss is 12.90, weight deacy loss is 10.30, training accuracy is 0.187500, time 56.384 samples/sec
+epoch 3, total_step 54300, total loss is 27.13 , inference loss is 16.82, weight deacy loss is 10.30, training accuracy is 0.156250, time 55.760 samples/sec
+epoch 3, total_step 54320, total loss is 21.98 , inference loss is 11.67, weight deacy loss is 10.30, training accuracy is 0.312500, time 56.390 samples/sec
+epoch 3, total_step 54340, total loss is 25.67 , inference loss is 15.36, weight deacy loss is 10.30, training accuracy is 0.250000, time 56.454 samples/sec
+epoch 3, total_step 54360, total loss is 26.17 , inference loss is 15.87, weight deacy loss is 10.30, training accuracy is 0.125000, time 56.488 samples/sec
+epoch 3, total_step 54380, total loss is 28.33 , inference loss is 18.04, weight deacy loss is 10.30, training accuracy is 0.093750, time 55.950 samples/sec
+epoch 3, total_step 54400, total loss is 21.37 , inference loss is 11.07, weight deacy loss is 10.30, training accuracy is 0.250000, time 56.142 samples/sec
+epoch 3, total_step 54420, total loss is 23.56 , inference loss is 13.26, weight deacy loss is 10.30, training accuracy is 0.312500, time 56.342 samples/sec
+epoch 3, total_step 54440, total loss is 24.46 , inference loss is 14.16, weight deacy loss is 10.29, training accuracy is 0.156250, time 56.488 samples/sec
+epoch 3, total_step 54460, total loss is 29.13 , inference loss is 18.84, weight deacy loss is 10.29, training accuracy is 0.093750, time 56.972 samples/sec
+epoch 3, total_step 54480, total loss is 26.40 , inference loss is 16.10, weight deacy loss is 10.29, training accuracy is 0.250000, time 57.054 samples/sec
+epoch 3, total_step 54500, total loss is 24.06 , inference loss is 13.77, weight deacy loss is 10.29, training accuracy is 0.250000, time 56.169 samples/sec
+epoch 3, total_step 54520, total loss is 29.22 , inference loss is 18.93, weight deacy loss is 10.29, training accuracy is 0.218750, time 56.562 samples/sec
+epoch 3, total_step 54540, total loss is 20.85 , inference loss is 10.56, weight deacy loss is 10.29, training accuracy is 0.312500, time 55.337 samples/sec
+epoch 3, total_step 54560, total loss is 28.42 , inference loss is 18.14, weight deacy loss is 10.29, training accuracy is 0.093750, time 56.282 samples/sec
+epoch 3, total_step 54580, total loss is 22.63 , inference loss is 12.35, weight deacy loss is 10.29, training accuracy is 0.156250, time 56.392 samples/sec
+epoch 3, total_step 54600, total loss is 22.35 , inference loss is 12.07, weight deacy loss is 10.28, training accuracy is 0.312500, time 55.508 samples/sec
+epoch 3, total_step 54620, total loss is 26.28 , inference loss is 16.00, weight deacy loss is 10.28, training accuracy is 0.156250, time 56.349 samples/sec
+epoch 3, total_step 54640, total loss is 26.62 , inference loss is 16.34, weight deacy loss is 10.28, training accuracy is 0.093750, time 57.113 samples/sec
+epoch 3, total_step 54660, total loss is 23.11 , inference loss is 12.83, weight deacy loss is 10.28, training accuracy is 0.312500, time 56.109 samples/sec
+epoch 3, total_step 54680, total loss is 21.70 , inference loss is 11.42, weight deacy loss is 10.28, training accuracy is 0.250000, time 56.483 samples/sec
+epoch 3, total_step 54700, total loss is 23.16 , inference loss is 12.89, weight deacy loss is 10.28, training accuracy is 0.187500, time 55.814 samples/sec
+epoch 3, total_step 54720, total loss is 21.11 , inference loss is 10.83, weight deacy loss is 10.28, training accuracy is 0.187500, time 56.253 samples/sec
+epoch 3, total_step 54740, total loss is 22.79 , inference loss is 12.51, weight deacy loss is 10.28, training accuracy is 0.250000, time 56.137 samples/sec
+epoch 3, total_step 54760, total loss is 25.94 , inference loss is 15.66, weight deacy loss is 10.27, training accuracy is 0.187500, time 55.909 samples/sec
+epoch 3, total_step 54780, total loss is 24.68 , inference loss is 14.41, weight deacy loss is 10.27, training accuracy is 0.218750, time 55.888 samples/sec
+epoch 3, total_step 54800, total loss is 22.29 , inference loss is 12.02, weight deacy loss is 10.27, training accuracy is 0.093750, time 55.680 samples/sec
+epoch 3, total_step 54820, total loss is 23.08 , inference loss is 12.81, weight deacy loss is 10.27, training accuracy is 0.250000, time 57.139 samples/sec
+epoch 3, total_step 54840, total loss is 20.11 , inference loss is 9.84, weight deacy loss is 10.27, training accuracy is 0.187500, time 55.452 samples/sec
+epoch 3, total_step 54860, total loss is 24.86 , inference loss is 14.60, weight deacy loss is 10.27, training accuracy is 0.187500, time 56.863 samples/sec
+epoch 3, total_step 54880, total loss is 22.01 , inference loss is 11.74, weight deacy loss is 10.27, training accuracy is 0.312500, time 56.425 samples/sec
+epoch 3, total_step 54900, total loss is 23.30 , inference loss is 13.04, weight deacy loss is 10.27, training accuracy is 0.187500, time 56.173 samples/sec
+epoch 3, total_step 54920, total loss is 23.87 , inference loss is 13.60, weight deacy loss is 10.26, training accuracy is 0.187500, time 56.630 samples/sec
+epoch 3, total_step 54940, total loss is 25.71 , inference loss is 15.45, weight deacy loss is 10.26, training accuracy is 0.187500, time 55.842 samples/sec
+epoch 3, total_step 54960, total loss is 25.04 , inference loss is 14.78, weight deacy loss is 10.26, training accuracy is 0.156250, time 55.975 samples/sec
+epoch 3, total_step 54980, total loss is 28.24 , inference loss is 17.98, weight deacy loss is 10.26, training accuracy is 0.031250, time 55.542 samples/sec
+epoch 3, total_step 55000, total loss is 27.82 , inference loss is 17.56, weight deacy loss is 10.26, training accuracy is 0.125000, time 53.882 samples/sec
+epoch 3, total_step 55020, total loss is 23.20 , inference loss is 12.94, weight deacy loss is 10.26, training accuracy is 0.250000, time 56.777 samples/sec
+epoch 3, total_step 55040, total loss is 21.06 , inference loss is 10.80, weight deacy loss is 10.26, training accuracy is 0.218750, time 57.123 samples/sec
+epoch 3, total_step 55060, total loss is 21.79 , inference loss is 11.54, weight deacy loss is 10.26, training accuracy is 0.156250, time 55.756 samples/sec
+epoch 3, total_step 55080, total loss is 27.35 , inference loss is 17.10, weight deacy loss is 10.25, training accuracy is 0.156250, time 55.925 samples/sec
+epoch 3, total_step 55100, total loss is 21.96 , inference loss is 11.71, weight deacy loss is 10.25, training accuracy is 0.312500, time 56.071 samples/sec
+epoch 3, total_step 55120, total loss is 29.53 , inference loss is 19.27, weight deacy loss is 10.25, training accuracy is 0.062500, time 56.444 samples/sec
+epoch 3, total_step 55140, total loss is 21.86 , inference loss is 11.61, weight deacy loss is 10.25, training accuracy is 0.218750, time 56.959 samples/sec
+epoch 3, total_step 55160, total loss is 19.97 , inference loss is 9.72, weight deacy loss is 10.25, training accuracy is 0.250000, time 55.576 samples/sec
+epoch 3, total_step 55180, total loss is 24.02 , inference loss is 13.77, weight deacy loss is 10.25, training accuracy is 0.156250, time 56.253 samples/sec
+epoch 3, total_step 55200, total loss is 26.54 , inference loss is 16.30, weight deacy loss is 10.25, training accuracy is 0.125000, time 56.771 samples/sec
+epoch 3, total_step 55220, total loss is 23.43 , inference loss is 13.18, weight deacy loss is 10.25, training accuracy is 0.187500, time 55.308 samples/sec
+epoch 3, total_step 55240, total loss is 24.87 , inference loss is 14.63, weight deacy loss is 10.24, training accuracy is 0.281250, time 56.256 samples/sec
+epoch 3, total_step 55260, total loss is 22.81 , inference loss is 12.57, weight deacy loss is 10.24, training accuracy is 0.312500, time 56.227 samples/sec
+epoch 3, total_step 55280, total loss is 23.41 , inference loss is 13.17, weight deacy loss is 10.24, training accuracy is 0.125000, time 55.158 samples/sec
+epoch 3, total_step 55300, total loss is 21.91 , inference loss is 11.67, weight deacy loss is 10.24, training accuracy is 0.218750, time 56.444 samples/sec
+epoch 3, total_step 55320, total loss is 19.66 , inference loss is 9.42, weight deacy loss is 10.24, training accuracy is 0.312500, time 56.086 samples/sec
+epoch 3, total_step 55340, total loss is 20.52 , inference loss is 10.28, weight deacy loss is 10.24, training accuracy is 0.312500, time 54.575 samples/sec
+epoch 3, total_step 55360, total loss is 18.64 , inference loss is 8.40, weight deacy loss is 10.24, training accuracy is 0.187500, time 56.424 samples/sec
+epoch 3, total_step 55380, total loss is 25.90 , inference loss is 15.66, weight deacy loss is 10.24, training accuracy is 0.093750, time 56.840 samples/sec
+epoch 3, total_step 55400, total loss is 23.94 , inference loss is 13.71, weight deacy loss is 10.23, training accuracy is 0.250000, time 56.695 samples/sec
+epoch 3, total_step 55420, total loss is 21.99 , inference loss is 11.75, weight deacy loss is 10.23, training accuracy is 0.218750, time 56.593 samples/sec
+epoch 3, total_step 55440, total loss is 22.56 , inference loss is 12.33, weight deacy loss is 10.23, training accuracy is 0.156250, time 56.643 samples/sec
+epoch 3, total_step 55460, total loss is 17.68 , inference loss is 7.45, weight deacy loss is 10.23, training accuracy is 0.250000, time 56.185 samples/sec
+epoch 3, total_step 55480, total loss is 22.55 , inference loss is 12.32, weight deacy loss is 10.23, training accuracy is 0.375000, time 57.031 samples/sec
+epoch 3, total_step 55500, total loss is 24.90 , inference loss is 14.67, weight deacy loss is 10.23, training accuracy is 0.187500, time 54.260 samples/sec
+epoch 3, total_step 55520, total loss is 23.19 , inference loss is 12.96, weight deacy loss is 10.23, training accuracy is 0.156250, time 56.249 samples/sec
+epoch 3, total_step 55540, total loss is 22.98 , inference loss is 12.76, weight deacy loss is 10.23, training accuracy is 0.125000, time 55.680 samples/sec
+epoch 3, total_step 55560, total loss is 22.55 , inference loss is 12.32, weight deacy loss is 10.23, training accuracy is 0.156250, time 55.594 samples/sec
+epoch 3, total_step 55580, total loss is 21.19 , inference loss is 10.96, weight deacy loss is 10.22, training accuracy is 0.187500, time 55.847 samples/sec
+epoch 3, total_step 55600, total loss is 25.40 , inference loss is 15.18, weight deacy loss is 10.22, training accuracy is 0.187500, time 56.310 samples/sec
+epoch 3, total_step 55620, total loss is 23.24 , inference loss is 13.02, weight deacy loss is 10.22, training accuracy is 0.218750, time 56.460 samples/sec
+epoch 3, total_step 55640, total loss is 22.63 , inference loss is 12.41, weight deacy loss is 10.22, training accuracy is 0.250000, time 56.806 samples/sec
+epoch 3, total_step 55660, total loss is 23.55 , inference loss is 13.33, weight deacy loss is 10.22, training accuracy is 0.125000, time 55.982 samples/sec
+epoch 3, total_step 55680, total loss is 26.18 , inference loss is 15.96, weight deacy loss is 10.22, training accuracy is 0.250000, time 56.132 samples/sec
+epoch 3, total_step 55700, total loss is 22.85 , inference loss is 12.63, weight deacy loss is 10.22, training accuracy is 0.406250, time 56.316 samples/sec
+epoch 3, total_step 55720, total loss is 20.09 , inference loss is 9.88, weight deacy loss is 10.22, training accuracy is 0.281250, time 55.844 samples/sec
+epoch 3, total_step 55740, total loss is 21.67 , inference loss is 11.46, weight deacy loss is 10.21, training accuracy is 0.187500, time 56.495 samples/sec
+epoch 3, total_step 55760, total loss is 21.82 , inference loss is 11.61, weight deacy loss is 10.21, training accuracy is 0.343750, time 54.367 samples/sec
+epoch 3, total_step 55780, total loss is 21.54 , inference loss is 11.33, weight deacy loss is 10.21, training accuracy is 0.312500, time 56.276 samples/sec
+epoch 3, total_step 55800, total loss is 28.74 , inference loss is 18.53, weight deacy loss is 10.21, training accuracy is 0.125000, time 55.551 samples/sec
+epoch 3, total_step 55820, total loss is 25.00 , inference loss is 14.79, weight deacy loss is 10.21, training accuracy is 0.218750, time 56.222 samples/sec
+epoch 3, total_step 55840, total loss is 25.00 , inference loss is 14.79, weight deacy loss is 10.21, training accuracy is 0.156250, time 56.312 samples/sec
+epoch 3, total_step 55860, total loss is 21.81 , inference loss is 11.60, weight deacy loss is 10.21, training accuracy is 0.125000, time 56.212 samples/sec
+epoch 3, total_step 55880, total loss is 18.29 , inference loss is 8.09, weight deacy loss is 10.21, training accuracy is 0.375000, time 56.324 samples/sec
+epoch 3, total_step 55900, total loss is 21.73 , inference loss is 11.52, weight deacy loss is 10.20, training accuracy is 0.218750, time 55.802 samples/sec
+epoch 3, total_step 55920, total loss is 22.25 , inference loss is 12.05, weight deacy loss is 10.20, training accuracy is 0.187500, time 56.600 samples/sec
+epoch 3, total_step 55940, total loss is 22.84 , inference loss is 12.64, weight deacy loss is 10.20, training accuracy is 0.250000, time 55.971 samples/sec
+epoch 3, total_step 55960, total loss is 21.55 , inference loss is 11.35, weight deacy loss is 10.20, training accuracy is 0.218750, time 56.748 samples/sec
+epoch 3, total_step 55980, total loss is 18.91 , inference loss is 8.71, weight deacy loss is 10.20, training accuracy is 0.281250, time 56.735 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.46343700000001
+best_threshold_index 109 0.9829629629629629
+best_threshold_index 109 0.9824074074074074
+best_threshold_index 109 0.9833333333333333
+best_threshold_index 109 0.9846296296296296
+best_threshold_index 109 0.9829629629629629
+best_threshold_index 109 0.9814814814814815
+best_threshold_index 109 0.9824074074074074
+best_threshold_index 109 0.9816666666666667
+best_threshold_index 109 0.9812962962962963
+best_threshold_index 109 0.9818518518518519
+[lfw][56000]XNorm: 23.527493
+[lfw][56000]Accuracy-Flip: 0.98250+-0.00867
+testing verification..
+(14000, 512)
+infer time 87.51939800000001
+best_threshold_index 120 0.9811111111111112
+best_threshold_index 120 0.98
+best_threshold_index 120 0.98
+best_threshold_index 116 0.979047619047619
+best_threshold_index 120 0.9811111111111112
+best_threshold_index 120 0.9796825396825397
+best_threshold_index 120 0.9812698412698413
+best_threshold_index 120 0.9793650793650793
+best_threshold_index 120 0.9811111111111112
+best_threshold_index 120 0.9803174603174604
+[cfp_ff][56000]XNorm: 22.594505
+[cfp_ff][56000]Accuracy-Flip: 0.97943+-0.00603
+testing verification..
+(14000, 512)
+infer time 86.79167600000002
+best_threshold_index 135 0.8920634920634921
+best_threshold_index 131 0.8904761904761904
+best_threshold_index 131 0.8893650793650794
+best_threshold_index 135 0.8933333333333333
+best_threshold_index 135 0.8938095238095238
+best_threshold_index 131 0.8895238095238095
+best_threshold_index 135 0.8892063492063492
+best_threshold_index 131 0.8901587301587301
+best_threshold_index 135 0.8904761904761904
+best_threshold_index 135 0.8909523809523809
+[cfp_fp][56000]XNorm: 22.181308
+[cfp_fp][56000]Accuracy-Flip: 0.88900+-0.01314
+testing verification..
+(12000, 512)
+infer time 74.81265700000007
+best_threshold_index 139 0.8742592592592593
+best_threshold_index 139 0.8709259259259259
+best_threshold_index 139 0.8738888888888889
+best_threshold_index 139 0.8735185185185185
+best_threshold_index 139 0.8807407407407407
+best_threshold_index 139 0.8729629629629629
+best_threshold_index 139 0.8711111111111111
+best_threshold_index 139 0.8781481481481481
+best_threshold_index 139 0.8785185185185185
+best_threshold_index 139 0.8759259259259259
+[agedb_30][56000]XNorm: 23.032732
+[agedb_30][56000]Accuracy-Flip: 0.87500+-0.02788
+test accuracy is: 0.9825000000000002
+epoch 3, total_step 56000, total loss is 20.51 , inference loss is 10.31, weight deacy loss is 10.20, training accuracy is 0.312500, time 52.822 samples/sec
+epoch 3, total_step 56020, total loss is 24.30 , inference loss is 14.11, weight deacy loss is 10.20, training accuracy is 0.156250, time 55.653 samples/sec
+epoch 3, total_step 56040, total loss is 23.56 , inference loss is 13.37, weight deacy loss is 10.20, training accuracy is 0.156250, time 55.565 samples/sec
+epoch 3, total_step 56060, total loss is 23.58 , inference loss is 13.39, weight deacy loss is 10.19, training accuracy is 0.375000, time 56.372 samples/sec
+epoch 3, total_step 56080, total loss is 25.46 , inference loss is 15.26, weight deacy loss is 10.19, training accuracy is 0.250000, time 56.753 samples/sec
+epoch 3, total_step 56100, total loss is 20.44 , inference loss is 10.25, weight deacy loss is 10.19, training accuracy is 0.156250, time 55.709 samples/sec
+epoch 3, total_step 56120, total loss is 20.52 , inference loss is 10.33, weight deacy loss is 10.19, training accuracy is 0.281250, time 56.808 samples/sec
+epoch 3, total_step 56140, total loss is 20.59 , inference loss is 10.40, weight deacy loss is 10.19, training accuracy is 0.343750, time 55.915 samples/sec
+epoch 3, total_step 56160, total loss is 21.00 , inference loss is 10.81, weight deacy loss is 10.19, training accuracy is 0.343750, time 55.658 samples/sec
+epoch 3, total_step 56180, total loss is 24.69 , inference loss is 14.51, weight deacy loss is 10.19, training accuracy is 0.218750, time 56.281 samples/sec
+epoch 3, total_step 56200, total loss is 21.17 , inference loss is 10.98, weight deacy loss is 10.19, training accuracy is 0.250000, time 54.103 samples/sec
+epoch 3, total_step 56220, total loss is 26.33 , inference loss is 16.14, weight deacy loss is 10.18, training accuracy is 0.125000, time 56.670 samples/sec
+epoch 3, total_step 56240, total loss is 23.79 , inference loss is 13.61, weight deacy loss is 10.18, training accuracy is 0.218750, time 56.412 samples/sec
+epoch 3, total_step 56260, total loss is 24.62 , inference loss is 14.44, weight deacy loss is 10.18, training accuracy is 0.125000, time 56.643 samples/sec
+epoch 3, total_step 56280, total loss is 23.23 , inference loss is 13.05, weight deacy loss is 10.18, training accuracy is 0.250000, time 55.525 samples/sec
+epoch 3, total_step 56300, total loss is 23.35 , inference loss is 13.17, weight deacy loss is 10.18, training accuracy is 0.250000, time 56.395 samples/sec
+epoch 3, total_step 56320, total loss is 20.35 , inference loss is 10.17, weight deacy loss is 10.18, training accuracy is 0.218750, time 56.462 samples/sec
+epoch 3, total_step 56340, total loss is 24.59 , inference loss is 14.42, weight deacy loss is 10.18, training accuracy is 0.062500, time 56.384 samples/sec
+epoch 3, total_step 56360, total loss is 25.78 , inference loss is 15.61, weight deacy loss is 10.18, training accuracy is 0.312500, time 56.486 samples/sec
+epoch 3, total_step 56380, total loss is 23.14 , inference loss is 12.97, weight deacy loss is 10.17, training accuracy is 0.125000, time 55.979 samples/sec
+epoch 3, total_step 56400, total loss is 22.53 , inference loss is 12.36, weight deacy loss is 10.17, training accuracy is 0.218750, time 56.617 samples/sec
+epoch 3, total_step 56420, total loss is 19.82 , inference loss is 9.65, weight deacy loss is 10.17, training accuracy is 0.250000, time 56.347 samples/sec
+epoch 3, total_step 56440, total loss is 21.57 , inference loss is 11.40, weight deacy loss is 10.17, training accuracy is 0.250000, time 56.869 samples/sec
+epoch 3, total_step 56460, total loss is 21.87 , inference loss is 11.70, weight deacy loss is 10.17, training accuracy is 0.312500, time 56.445 samples/sec
+epoch 3, total_step 56480, total loss is 23.29 , inference loss is 13.12, weight deacy loss is 10.17, training accuracy is 0.218750, time 54.817 samples/sec
+epoch 3, total_step 56500, total loss is 23.53 , inference loss is 13.36, weight deacy loss is 10.17, training accuracy is 0.187500, time 55.972 samples/sec
+epoch 3, total_step 56520, total loss is 25.43 , inference loss is 15.27, weight deacy loss is 10.17, training accuracy is 0.062500, time 57.007 samples/sec
+epoch 3, total_step 56540, total loss is 27.09 , inference loss is 16.92, weight deacy loss is 10.16, training accuracy is 0.281250, time 56.516 samples/sec
+epoch 3, total_step 56560, total loss is 24.11 , inference loss is 13.95, weight deacy loss is 10.16, training accuracy is 0.281250, time 56.173 samples/sec
+epoch 3, total_step 56580, total loss is 23.79 , inference loss is 13.63, weight deacy loss is 10.16, training accuracy is 0.156250, time 55.862 samples/sec
+epoch 3, total_step 56600, total loss is 22.88 , inference loss is 12.72, weight deacy loss is 10.16, training accuracy is 0.281250, time 55.863 samples/sec
+epoch 3, total_step 56620, total loss is 20.56 , inference loss is 10.40, weight deacy loss is 10.16, training accuracy is 0.281250, time 55.129 samples/sec
+epoch 3, total_step 56640, total loss is 22.00 , inference loss is 11.84, weight deacy loss is 10.16, training accuracy is 0.156250, time 56.052 samples/sec
+epoch 3, total_step 56660, total loss is 26.42 , inference loss is 16.26, weight deacy loss is 10.16, training accuracy is 0.156250, time 56.105 samples/sec
+epoch 3, total_step 56680, total loss is 26.22 , inference loss is 16.07, weight deacy loss is 10.16, training accuracy is 0.187500, time 56.696 samples/sec
+epoch 3, total_step 56700, total loss is 25.07 , inference loss is 14.91, weight deacy loss is 10.15, training accuracy is 0.156250, time 57.127 samples/sec
+epoch 3, total_step 56720, total loss is 27.29 , inference loss is 17.14, weight deacy loss is 10.15, training accuracy is 0.093750, time 56.375 samples/sec
+epoch 3, total_step 56740, total loss is 25.38 , inference loss is 15.23, weight deacy loss is 10.15, training accuracy is 0.156250, time 57.166 samples/sec
+epoch 3, total_step 56760, total loss is 23.79 , inference loss is 13.64, weight deacy loss is 10.15, training accuracy is 0.218750, time 56.177 samples/sec
+epoch 3, total_step 56780, total loss is 23.60 , inference loss is 13.45, weight deacy loss is 10.15, training accuracy is 0.250000, time 56.498 samples/sec
+epoch 3, total_step 56800, total loss is 25.87 , inference loss is 15.72, weight deacy loss is 10.15, training accuracy is 0.187500, time 56.423 samples/sec
+epoch 3, total_step 56820, total loss is 22.89 , inference loss is 12.74, weight deacy loss is 10.15, training accuracy is 0.187500, time 56.368 samples/sec
+epoch 3, total_step 56840, total loss is 21.40 , inference loss is 11.25, weight deacy loss is 10.15, training accuracy is 0.187500, time 55.528 samples/sec
+epoch 3, total_step 56860, total loss is 22.40 , inference loss is 12.26, weight deacy loss is 10.14, training accuracy is 0.343750, time 56.104 samples/sec
+epoch 3, total_step 56880, total loss is 18.57 , inference loss is 8.43, weight deacy loss is 10.14, training accuracy is 0.250000, time 56.254 samples/sec
+epoch 3, total_step 56900, total loss is 20.25 , inference loss is 10.11, weight deacy loss is 10.14, training accuracy is 0.156250, time 56.560 samples/sec
+epoch 3, total_step 56920, total loss is 25.45 , inference loss is 15.31, weight deacy loss is 10.14, training accuracy is 0.125000, time 56.734 samples/sec
+epoch 3, total_step 56940, total loss is 18.00 , inference loss is 7.86, weight deacy loss is 10.14, training accuracy is 0.437500, time 55.548 samples/sec
+epoch 3, total_step 56960, total loss is 24.23 , inference loss is 14.09, weight deacy loss is 10.14, training accuracy is 0.187500, time 55.487 samples/sec
+epoch 3, total_step 56980, total loss is 22.73 , inference loss is 12.59, weight deacy loss is 10.14, training accuracy is 0.281250, time 55.878 samples/sec
+epoch 3, total_step 57000, total loss is 21.04 , inference loss is 10.91, weight deacy loss is 10.14, training accuracy is 0.187500, time 55.820 samples/sec
+epoch 3, total_step 57020, total loss is 22.86 , inference loss is 12.72, weight deacy loss is 10.14, training accuracy is 0.093750, time 56.552 samples/sec
+epoch 3, total_step 57040, total loss is 23.65 , inference loss is 13.52, weight deacy loss is 10.13, training accuracy is 0.218750, time 56.879 samples/sec
+epoch 3, total_step 57060, total loss is 19.60 , inference loss is 9.47, weight deacy loss is 10.13, training accuracy is 0.218750, time 55.616 samples/sec
+epoch 3, total_step 57080, total loss is 22.45 , inference loss is 12.32, weight deacy loss is 10.13, training accuracy is 0.187500, time 55.476 samples/sec
+epoch 3, total_step 57100, total loss is 27.49 , inference loss is 17.36, weight deacy loss is 10.13, training accuracy is 0.187500, time 56.056 samples/sec
+epoch 3, total_step 57120, total loss is 22.80 , inference loss is 12.67, weight deacy loss is 10.13, training accuracy is 0.156250, time 54.874 samples/sec
+epoch 3, total_step 57140, total loss is 20.99 , inference loss is 10.86, weight deacy loss is 10.13, training accuracy is 0.218750, time 56.036 samples/sec
+epoch 3, total_step 57160, total loss is 23.15 , inference loss is 13.03, weight deacy loss is 10.13, training accuracy is 0.187500, time 56.611 samples/sec
+epoch 3, total_step 57180, total loss is 26.25 , inference loss is 16.13, weight deacy loss is 10.13, training accuracy is 0.093750, time 56.346 samples/sec
+epoch 3, total_step 57200, total loss is 25.37 , inference loss is 15.24, weight deacy loss is 10.12, training accuracy is 0.125000, time 55.669 samples/sec
+epoch 3, total_step 57220, total loss is 23.04 , inference loss is 12.92, weight deacy loss is 10.12, training accuracy is 0.156250, time 57.017 samples/sec
+epoch 3, total_step 57240, total loss is 24.70 , inference loss is 14.58, weight deacy loss is 10.12, training accuracy is 0.187500, time 56.478 samples/sec
+epoch 3, total_step 57260, total loss is 20.44 , inference loss is 10.32, weight deacy loss is 10.12, training accuracy is 0.250000, time 55.471 samples/sec
+epoch 3, total_step 57280, total loss is 21.20 , inference loss is 11.08, weight deacy loss is 10.12, training accuracy is 0.281250, time 56.322 samples/sec
+epoch 3, total_step 57300, total loss is 21.00 , inference loss is 10.88, weight deacy loss is 10.12, training accuracy is 0.187500, time 55.793 samples/sec
+epoch 3, total_step 57320, total loss is 24.34 , inference loss is 14.22, weight deacy loss is 10.12, training accuracy is 0.187500, time 56.854 samples/sec
+epoch 3, total_step 57340, total loss is 21.71 , inference loss is 11.59, weight deacy loss is 10.12, training accuracy is 0.218750, time 56.011 samples/sec
+epoch 3, total_step 57360, total loss is 27.27 , inference loss is 17.15, weight deacy loss is 10.11, training accuracy is 0.156250, time 56.401 samples/sec
+epoch 3, total_step 57380, total loss is 20.40 , inference loss is 10.29, weight deacy loss is 10.11, training accuracy is 0.281250, time 55.690 samples/sec
+epoch 3, total_step 57400, total loss is 22.52 , inference loss is 12.41, weight deacy loss is 10.11, training accuracy is 0.125000, time 53.856 samples/sec
+epoch 3, total_step 57420, total loss is 20.18 , inference loss is 10.07, weight deacy loss is 10.11, training accuracy is 0.250000, time 56.499 samples/sec
+epoch 3, total_step 57440, total loss is 25.45 , inference loss is 15.34, weight deacy loss is 10.11, training accuracy is 0.062500, time 56.407 samples/sec
+epoch 3, total_step 57460, total loss is 22.34 , inference loss is 12.23, weight deacy loss is 10.11, training accuracy is 0.156250, time 56.552 samples/sec
+epoch 3, total_step 57480, total loss is 26.18 , inference loss is 16.08, weight deacy loss is 10.11, training accuracy is 0.156250, time 56.418 samples/sec
+epoch 3, total_step 57500, total loss is 22.85 , inference loss is 12.75, weight deacy loss is 10.11, training accuracy is 0.250000, time 56.538 samples/sec
+epoch 3, total_step 57520, total loss is 21.71 , inference loss is 11.61, weight deacy loss is 10.11, training accuracy is 0.125000, time 56.512 samples/sec
+epoch 3, total_step 57540, total loss is 28.98 , inference loss is 18.87, weight deacy loss is 10.10, training accuracy is 0.125000, time 56.518 samples/sec
+epoch 3, total_step 57560, total loss is 23.87 , inference loss is 13.76, weight deacy loss is 10.10, training accuracy is 0.156250, time 57.245 samples/sec
+epoch 3, total_step 57580, total loss is 22.20 , inference loss is 12.10, weight deacy loss is 10.10, training accuracy is 0.250000, time 56.576 samples/sec
+epoch 3, total_step 57600, total loss is 22.73 , inference loss is 12.63, weight deacy loss is 10.10, training accuracy is 0.187500, time 55.348 samples/sec
+epoch 3, total_step 57620, total loss is 22.06 , inference loss is 11.96, weight deacy loss is 10.10, training accuracy is 0.156250, time 55.874 samples/sec
+epoch 3, total_step 57640, total loss is 22.71 , inference loss is 12.61, weight deacy loss is 10.10, training accuracy is 0.125000, time 56.109 samples/sec
+epoch 3, total_step 57660, total loss is 18.26 , inference loss is 8.16, weight deacy loss is 10.10, training accuracy is 0.312500, time 56.416 samples/sec
+epoch 3, total_step 57680, total loss is 24.36 , inference loss is 14.26, weight deacy loss is 10.10, training accuracy is 0.187500, time 54.015 samples/sec
+epoch 3, total_step 57700, total loss is 25.93 , inference loss is 15.83, weight deacy loss is 10.09, training accuracy is 0.156250, time 56.794 samples/sec
+epoch 3, total_step 57720, total loss is 23.81 , inference loss is 13.71, weight deacy loss is 10.09, training accuracy is 0.125000, time 55.953 samples/sec
+epoch 3, total_step 57740, total loss is 27.90 , inference loss is 17.81, weight deacy loss is 10.09, training accuracy is 0.156250, time 56.431 samples/sec
+epoch 3, total_step 57760, total loss is 22.87 , inference loss is 12.78, weight deacy loss is 10.09, training accuracy is 0.250000, time 56.919 samples/sec
+epoch 3, total_step 57780, total loss is 25.22 , inference loss is 15.13, weight deacy loss is 10.09, training accuracy is 0.187500, time 55.683 samples/sec
+epoch 3, total_step 57800, total loss is 25.48 , inference loss is 15.39, weight deacy loss is 10.09, training accuracy is 0.062500, time 55.851 samples/sec
+epoch 3, total_step 57820, total loss is 21.45 , inference loss is 11.36, weight deacy loss is 10.09, training accuracy is 0.250000, time 55.938 samples/sec
+epoch 3, total_step 57840, total loss is 24.34 , inference loss is 14.26, weight deacy loss is 10.09, training accuracy is 0.187500, time 56.545 samples/sec
+epoch 3, total_step 57860, total loss is 23.84 , inference loss is 13.75, weight deacy loss is 10.08, training accuracy is 0.156250, time 56.019 samples/sec
+epoch 3, total_step 57880, total loss is 28.01 , inference loss is 17.92, weight deacy loss is 10.08, training accuracy is 0.062500, time 52.964 samples/sec
+epoch 3, total_step 57900, total loss is 24.53 , inference loss is 14.45, weight deacy loss is 10.08, training accuracy is 0.156250, time 55.251 samples/sec
+epoch 3, total_step 57920, total loss is 27.65 , inference loss is 17.57, weight deacy loss is 10.08, training accuracy is 0.093750, time 55.695 samples/sec
+epoch 3, total_step 57940, total loss is 22.81 , inference loss is 12.73, weight deacy loss is 10.08, training accuracy is 0.218750, time 55.561 samples/sec
+epoch 3, total_step 57960, total loss is 24.88 , inference loss is 14.80, weight deacy loss is 10.08, training accuracy is 0.031250, time 55.120 samples/sec
+epoch 3, total_step 57980, total loss is 24.73 , inference loss is 14.65, weight deacy loss is 10.08, training accuracy is 0.187500, time 55.993 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.64854100000004
+best_threshold_index 112 0.9825925925925926
+best_threshold_index 116 0.9818518518518519
+best_threshold_index 112 0.9835185185185186
+best_threshold_index 112 0.9842592592592593
+best_threshold_index 116 0.9833333333333333
+best_threshold_index 116 0.9824074074074074
+best_threshold_index 115 0.982037037037037
+best_threshold_index 112 0.9818518518518519
+best_threshold_index 116 0.9811111111111112
+best_threshold_index 112 0.982037037037037
+[lfw][58000]XNorm: 23.360172
+[lfw][58000]Accuracy-Flip: 0.98067+-0.00800
+testing verification..
+(14000, 512)
+infer time 86.48279700000008
+best_threshold_index 116 0.9825396825396825
+best_threshold_index 116 0.981904761904762
+best_threshold_index 122 0.981904761904762
+best_threshold_index 116 0.9809523809523809
+best_threshold_index 122 0.981904761904762
+best_threshold_index 116 0.9806349206349206
+best_threshold_index 122 0.9820634920634921
+best_threshold_index 116 0.9812698412698413
+best_threshold_index 116 0.9820634920634921
+best_threshold_index 116 0.9820634920634921
+[cfp_ff][58000]XNorm: 22.642918
+[cfp_ff][58000]Accuracy-Flip: 0.98014+-0.00501
+testing verification..
+(14000, 512)
+infer time 86.35597599999993
+best_threshold_index 141 0.8988888888888888
+best_threshold_index 140 0.8990476190476191
+best_threshold_index 140 0.8984126984126984
+best_threshold_index 141 0.8977777777777778
+best_threshold_index 141 0.9001587301587302
+best_threshold_index 140 0.8961904761904762
+best_threshold_index 138 0.8968253968253969
+best_threshold_index 140 0.8979365079365079
+best_threshold_index 140 0.8976190476190476
+best_threshold_index 140 0.8998412698412699
+[cfp_fp][58000]XNorm: 22.665178
+[cfp_fp][58000]Accuracy-Flip: 0.89629+-0.01007
+testing verification..
+(12000, 512)
+infer time 74.209304
+best_threshold_index 135 0.8787037037037037
+best_threshold_index 135 0.8781481481481481
+best_threshold_index 134 0.8801851851851852
+best_threshold_index 135 0.8812962962962962
+best_threshold_index 134 0.8855555555555555
+best_threshold_index 134 0.8792592592592593
+best_threshold_index 135 0.8818518518518519
+best_threshold_index 134 0.8851851851851852
+best_threshold_index 133 0.8831481481481481
+best_threshold_index 135 0.8814814814814815
+[agedb_30][58000]XNorm: 23.080417
+[agedb_30][58000]Accuracy-Flip: 0.87967+-0.02206
+test accuracy is: 0.9806666666666667
+epoch 3, total_step 58000, total loss is 25.57 , inference loss is 15.50, weight deacy loss is 10.08, training accuracy is 0.156250, time 54.156 samples/sec
+epoch 3, total_step 58020, total loss is 24.70 , inference loss is 14.62, weight deacy loss is 10.08, training accuracy is 0.156250, time 56.327 samples/sec
+epoch 3, total_step 58040, total loss is 23.89 , inference loss is 13.82, weight deacy loss is 10.07, training accuracy is 0.312500, time 56.345 samples/sec
+epoch 3, total_step 58060, total loss is 23.42 , inference loss is 13.35, weight deacy loss is 10.07, training accuracy is 0.062500, time 56.213 samples/sec
+epoch 3, total_step 58080, total loss is 23.41 , inference loss is 13.33, weight deacy loss is 10.07, training accuracy is 0.125000, time 56.559 samples/sec
+epoch 3, total_step 58100, total loss is 25.80 , inference loss is 15.73, weight deacy loss is 10.07, training accuracy is 0.093750, time 55.640 samples/sec
+epoch 3, total_step 58120, total loss is 20.81 , inference loss is 10.74, weight deacy loss is 10.07, training accuracy is 0.343750, time 56.235 samples/sec
+epoch 3, total_step 58140, total loss is 23.21 , inference loss is 13.14, weight deacy loss is 10.07, training accuracy is 0.187500, time 55.478 samples/sec
+epoch 3, total_step 58160, total loss is 19.52 , inference loss is 9.46, weight deacy loss is 10.07, training accuracy is 0.343750, time 55.838 samples/sec
+epoch 3, total_step 58180, total loss is 25.70 , inference loss is 15.63, weight deacy loss is 10.07, training accuracy is 0.187500, time 56.274 samples/sec
+epoch 3, total_step 58200, total loss is 26.60 , inference loss is 16.53, weight deacy loss is 10.06, training accuracy is 0.187500, time 56.159 samples/sec
+epoch 3, total_step 58220, total loss is 27.02 , inference loss is 16.96, weight deacy loss is 10.06, training accuracy is 0.062500, time 55.888 samples/sec
+epoch 3, total_step 58240, total loss is 23.77 , inference loss is 13.71, weight deacy loss is 10.06, training accuracy is 0.093750, time 56.091 samples/sec
+epoch 3, total_step 58260, total loss is 22.32 , inference loss is 12.26, weight deacy loss is 10.06, training accuracy is 0.187500, time 56.144 samples/sec
+epoch 3, total_step 58280, total loss is 25.48 , inference loss is 15.42, weight deacy loss is 10.06, training accuracy is 0.218750, time 56.374 samples/sec
+epoch 3, total_step 58300, total loss is 27.25 , inference loss is 17.19, weight deacy loss is 10.06, training accuracy is 0.187500, time 56.741 samples/sec
+epoch 3, total_step 58320, total loss is 23.78 , inference loss is 13.72, weight deacy loss is 10.06, training accuracy is 0.281250, time 56.127 samples/sec
+epoch 3, total_step 58340, total loss is 24.27 , inference loss is 14.21, weight deacy loss is 10.06, training accuracy is 0.125000, time 55.457 samples/sec
+epoch 3, total_step 58360, total loss is 23.79 , inference loss is 13.73, weight deacy loss is 10.05, training accuracy is 0.250000, time 56.789 samples/sec
+epoch 3, total_step 58380, total loss is 24.48 , inference loss is 14.43, weight deacy loss is 10.05, training accuracy is 0.156250, time 54.256 samples/sec
+epoch 3, total_step 58400, total loss is 26.00 , inference loss is 15.94, weight deacy loss is 10.05, training accuracy is 0.218750, time 56.240 samples/sec
+epoch 3, total_step 58420, total loss is 25.04 , inference loss is 14.99, weight deacy loss is 10.05, training accuracy is 0.156250, time 56.128 samples/sec
+epoch 3, total_step 58440, total loss is 29.35 , inference loss is 19.30, weight deacy loss is 10.05, training accuracy is 0.125000, time 56.150 samples/sec
+epoch 3, total_step 58460, total loss is 26.63 , inference loss is 16.58, weight deacy loss is 10.05, training accuracy is 0.093750, time 55.576 samples/sec
+epoch 3, total_step 58480, total loss is 25.38 , inference loss is 15.33, weight deacy loss is 10.05, training accuracy is 0.062500, time 55.618 samples/sec
+epoch 3, total_step 58500, total loss is 25.93 , inference loss is 15.88, weight deacy loss is 10.05, training accuracy is 0.218750, time 55.947 samples/sec
+epoch 3, total_step 58520, total loss is 22.05 , inference loss is 12.01, weight deacy loss is 10.05, training accuracy is 0.187500, time 56.194 samples/sec
+epoch 3, total_step 58540, total loss is 28.04 , inference loss is 18.00, weight deacy loss is 10.04, training accuracy is 0.156250, time 55.951 samples/sec
+epoch 3, total_step 58560, total loss is 23.88 , inference loss is 13.83, weight deacy loss is 10.04, training accuracy is 0.187500, time 56.059 samples/sec
+epoch 3, total_step 58580, total loss is 22.99 , inference loss is 12.95, weight deacy loss is 10.04, training accuracy is 0.062500, time 56.973 samples/sec
+epoch 3, total_step 58600, total loss is 26.95 , inference loss is 16.91, weight deacy loss is 10.04, training accuracy is 0.156250, time 56.307 samples/sec
+epoch 3, total_step 58620, total loss is 18.94 , inference loss is 8.90, weight deacy loss is 10.04, training accuracy is 0.281250, time 56.852 samples/sec
+epoch 3, total_step 58640, total loss is 26.43 , inference loss is 16.39, weight deacy loss is 10.04, training accuracy is 0.093750, time 55.987 samples/sec
+epoch 3, total_step 58660, total loss is 23.63 , inference loss is 13.60, weight deacy loss is 10.04, training accuracy is 0.156250, time 55.907 samples/sec
+epoch 3, total_step 58680, total loss is 20.04 , inference loss is 10.01, weight deacy loss is 10.04, training accuracy is 0.218750, time 56.488 samples/sec
+epoch 3, total_step 58700, total loss is 23.45 , inference loss is 13.41, weight deacy loss is 10.03, training accuracy is 0.218750, time 56.256 samples/sec
+epoch 3, total_step 58720, total loss is 25.63 , inference loss is 15.60, weight deacy loss is 10.03, training accuracy is 0.187500, time 55.660 samples/sec
+epoch 3, total_step 58740, total loss is 23.25 , inference loss is 13.22, weight deacy loss is 10.03, training accuracy is 0.218750, time 56.619 samples/sec
+epoch 3, total_step 58760, total loss is 22.62 , inference loss is 12.59, weight deacy loss is 10.03, training accuracy is 0.187500, time 55.823 samples/sec
+epoch 3, total_step 58780, total loss is 22.70 , inference loss is 12.67, weight deacy loss is 10.03, training accuracy is 0.218750, time 55.621 samples/sec
+epoch 3, total_step 58800, total loss is 22.72 , inference loss is 12.69, weight deacy loss is 10.03, training accuracy is 0.093750, time 55.895 samples/sec
+epoch 3, total_step 58820, total loss is 21.48 , inference loss is 11.45, weight deacy loss is 10.03, training accuracy is 0.156250, time 56.887 samples/sec
+epoch 3, total_step 58840, total loss is 26.65 , inference loss is 16.62, weight deacy loss is 10.03, training accuracy is 0.093750, time 55.419 samples/sec
+epoch 3, total_step 58860, total loss is 25.96 , inference loss is 15.93, weight deacy loss is 10.03, training accuracy is 0.156250, time 56.231 samples/sec
+epoch 3, total_step 58880, total loss is 19.57 , inference loss is 9.55, weight deacy loss is 10.02, training accuracy is 0.250000, time 56.196 samples/sec
+epoch 3, total_step 58900, total loss is 24.89 , inference loss is 14.87, weight deacy loss is 10.02, training accuracy is 0.187500, time 55.950 samples/sec
+epoch 3, total_step 58920, total loss is 25.27 , inference loss is 15.25, weight deacy loss is 10.02, training accuracy is 0.156250, time 55.812 samples/sec
+epoch 3, total_step 58940, total loss is 24.40 , inference loss is 14.38, weight deacy loss is 10.02, training accuracy is 0.156250, time 55.928 samples/sec
+epoch 3, total_step 58960, total loss is 25.84 , inference loss is 15.82, weight deacy loss is 10.02, training accuracy is 0.062500, time 55.642 samples/sec
+epoch 3, total_step 58980, total loss is 25.57 , inference loss is 15.55, weight deacy loss is 10.02, training accuracy is 0.093750, time 55.827 samples/sec
+epoch 3, total_step 59000, total loss is 22.65 , inference loss is 12.63, weight deacy loss is 10.02, training accuracy is 0.031250, time 55.999 samples/sec
+epoch 3, total_step 59020, total loss is 25.53 , inference loss is 15.51, weight deacy loss is 10.02, training accuracy is 0.125000, time 56.409 samples/sec
+epoch 3, total_step 59040, total loss is 23.18 , inference loss is 13.16, weight deacy loss is 10.01, training accuracy is 0.156250, time 56.466 samples/sec
+epoch 3, total_step 59060, total loss is 24.02 , inference loss is 14.01, weight deacy loss is 10.01, training accuracy is 0.125000, time 56.624 samples/sec
+epoch 3, total_step 59080, total loss is 22.63 , inference loss is 12.62, weight deacy loss is 10.01, training accuracy is 0.125000, time 56.169 samples/sec
+epoch 3, total_step 59100, total loss is 21.59 , inference loss is 11.58, weight deacy loss is 10.01, training accuracy is 0.343750, time 55.980 samples/sec
+epoch 3, total_step 59120, total loss is 25.31 , inference loss is 15.30, weight deacy loss is 10.01, training accuracy is 0.125000, time 56.097 samples/sec
+epoch 3, total_step 59140, total loss is 24.16 , inference loss is 14.15, weight deacy loss is 10.01, training accuracy is 0.187500, time 55.747 samples/sec
+epoch 3, total_step 59160, total loss is 23.83 , inference loss is 13.82, weight deacy loss is 10.01, training accuracy is 0.187500, time 54.921 samples/sec
+epoch 3, total_step 59180, total loss is 21.08 , inference loss is 11.07, weight deacy loss is 10.01, training accuracy is 0.187500, time 55.945 samples/sec
+epoch 3, total_step 59200, total loss is 25.18 , inference loss is 15.18, weight deacy loss is 10.01, training accuracy is 0.156250, time 55.832 samples/sec
+epoch 3, total_step 59220, total loss is 24.59 , inference loss is 14.59, weight deacy loss is 10.00, training accuracy is 0.093750, time 56.575 samples/sec
+epoch 3, total_step 59240, total loss is 18.75 , inference loss is 8.74, weight deacy loss is 10.00, training accuracy is 0.312500, time 55.544 samples/sec
+epoch 3, total_step 59260, total loss is 23.70 , inference loss is 13.70, weight deacy loss is 10.00, training accuracy is 0.093750, time 56.632 samples/sec
+epoch 3, total_step 59280, total loss is 25.85 , inference loss is 15.85, weight deacy loss is 10.00, training accuracy is 0.125000, time 56.206 samples/sec
+epoch 3, total_step 59300, total loss is 20.76 , inference loss is 10.76, weight deacy loss is 10.00, training accuracy is 0.312500, time 57.079 samples/sec
+epoch 3, total_step 59320, total loss is 22.55 , inference loss is 12.55, weight deacy loss is 10.00, training accuracy is 0.125000, time 56.591 samples/sec
+epoch 3, total_step 59340, total loss is 27.14 , inference loss is 17.14, weight deacy loss is 10.00, training accuracy is 0.062500, time 56.679 samples/sec
+epoch 3, total_step 59360, total loss is 25.34 , inference loss is 15.34, weight deacy loss is 10.00, training accuracy is 0.093750, time 57.043 samples/sec
+epoch 3, total_step 59380, total loss is 25.97 , inference loss is 15.98, weight deacy loss is 9.99, training accuracy is 0.156250, time 56.774 samples/sec
+epoch 3, total_step 59400, total loss is 21.83 , inference loss is 11.84, weight deacy loss is 9.99, training accuracy is 0.312500, time 55.892 samples/sec
+epoch 3, total_step 59420, total loss is 25.51 , inference loss is 15.52, weight deacy loss is 9.99, training accuracy is 0.250000, time 56.557 samples/sec
+epoch 3, total_step 59440, total loss is 22.18 , inference loss is 12.19, weight deacy loss is 9.99, training accuracy is 0.093750, time 56.410 samples/sec
+epoch 3, total_step 59460, total loss is 22.56 , inference loss is 12.57, weight deacy loss is 9.99, training accuracy is 0.312500, time 55.723 samples/sec
+epoch 3, total_step 59480, total loss is 20.81 , inference loss is 10.82, weight deacy loss is 9.99, training accuracy is 0.187500, time 56.342 samples/sec
+epoch 3, total_step 59500, total loss is 23.79 , inference loss is 13.80, weight deacy loss is 9.99, training accuracy is 0.093750, time 56.563 samples/sec
+epoch 3, total_step 59520, total loss is 25.52 , inference loss is 15.54, weight deacy loss is 9.99, training accuracy is 0.156250, time 56.030 samples/sec
+epoch 3, total_step 59540, total loss is 22.95 , inference loss is 12.96, weight deacy loss is 9.99, training accuracy is 0.156250, time 56.406 samples/sec
+epoch 3, total_step 59560, total loss is 23.29 , inference loss is 13.30, weight deacy loss is 9.98, training accuracy is 0.156250, time 56.646 samples/sec
+epoch 3, total_step 59580, total loss is 22.66 , inference loss is 12.68, weight deacy loss is 9.98, training accuracy is 0.250000, time 55.949 samples/sec
+epoch 3, total_step 59600, total loss is 22.34 , inference loss is 12.36, weight deacy loss is 9.98, training accuracy is 0.093750, time 56.507 samples/sec
+epoch 3, total_step 59620, total loss is 22.68 , inference loss is 12.70, weight deacy loss is 9.98, training accuracy is 0.125000, time 56.164 samples/sec
+epoch 3, total_step 59640, total loss is 22.74 , inference loss is 12.76, weight deacy loss is 9.98, training accuracy is 0.218750, time 56.222 samples/sec
+epoch 3, total_step 59660, total loss is 25.00 , inference loss is 15.02, weight deacy loss is 9.98, training accuracy is 0.156250, time 56.952 samples/sec
+epoch 3, total_step 59680, total loss is 21.60 , inference loss is 11.63, weight deacy loss is 9.98, training accuracy is 0.250000, time 56.777 samples/sec
+epoch 3, total_step 59700, total loss is 25.87 , inference loss is 15.89, weight deacy loss is 9.98, training accuracy is 0.187500, time 55.899 samples/sec
+epoch 3, total_step 59720, total loss is 22.51 , inference loss is 12.53, weight deacy loss is 9.97, training accuracy is 0.125000, time 56.299 samples/sec
+epoch 3, total_step 59740, total loss is 23.07 , inference loss is 13.10, weight deacy loss is 9.97, training accuracy is 0.093750, time 56.388 samples/sec
+epoch 3, total_step 59760, total loss is 26.24 , inference loss is 16.27, weight deacy loss is 9.97, training accuracy is 0.125000, time 56.580 samples/sec
+epoch 3, total_step 59780, total loss is 20.76 , inference loss is 10.78, weight deacy loss is 9.97, training accuracy is 0.281250, time 56.135 samples/sec
+epoch 3, total_step 59800, total loss is 31.51 , inference loss is 21.54, weight deacy loss is 9.97, training accuracy is 0.093750, time 56.878 samples/sec
+epoch 3, total_step 59820, total loss is 25.77 , inference loss is 15.80, weight deacy loss is 9.97, training accuracy is 0.093750, time 57.115 samples/sec
+epoch 3, total_step 59840, total loss is 21.71 , inference loss is 11.74, weight deacy loss is 9.97, training accuracy is 0.281250, time 55.656 samples/sec
+epoch 3, total_step 59860, total loss is 21.53 , inference loss is 11.56, weight deacy loss is 9.97, training accuracy is 0.187500, time 56.259 samples/sec
+epoch 3, total_step 59880, total loss is 21.86 , inference loss is 11.89, weight deacy loss is 9.97, training accuracy is 0.156250, time 55.384 samples/sec
+epoch 3, total_step 59900, total loss is 22.79 , inference loss is 12.83, weight deacy loss is 9.96, training accuracy is 0.187500, time 56.271 samples/sec
+epoch 3, total_step 59920, total loss is 25.33 , inference loss is 15.37, weight deacy loss is 9.96, training accuracy is 0.093750, time 56.164 samples/sec
+epoch 3, total_step 59940, total loss is 19.65 , inference loss is 9.69, weight deacy loss is 9.96, training accuracy is 0.218750, time 55.895 samples/sec
+epoch 3, total_step 59960, total loss is 26.82 , inference loss is 16.86, weight deacy loss is 9.96, training accuracy is 0.125000, time 56.569 samples/sec
+epoch 3, total_step 59980, total loss is 24.76 , inference loss is 14.80, weight deacy loss is 9.96, training accuracy is 0.250000, time 55.843 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.91141499999999
+best_threshold_index 105 0.980925925925926
+best_threshold_index 105 0.9798148148148148
+best_threshold_index 108 0.9812962962962963
+best_threshold_index 105 0.9824074074074074
+best_threshold_index 106 0.9812962962962963
+best_threshold_index 105 0.9798148148148148
+best_threshold_index 105 0.98
+best_threshold_index 105 0.98
+best_threshold_index 105 0.9788888888888889
+best_threshold_index 105 0.9792592592592593
+[lfw][60000]XNorm: 23.064352
+[lfw][60000]Accuracy-Flip: 0.97983+-0.00938
+testing verification..
+(14000, 512)
+infer time 85.91767600000003
+best_threshold_index 116 0.9828571428571429
+best_threshold_index 116 0.9825396825396825
+best_threshold_index 116 0.9822222222222222
+best_threshold_index 116 0.9822222222222222
+best_threshold_index 116 0.9822222222222222
+best_threshold_index 116 0.9817460317460317
+best_threshold_index 116 0.9820634920634921
+best_threshold_index 116 0.9806349206349206
+best_threshold_index 116 0.9825396825396825
+best_threshold_index 116 0.9823809523809524
+[cfp_ff][60000]XNorm: 22.127951
+[cfp_ff][60000]Accuracy-Flip: 0.98214+-0.00520
+testing verification..
+(14000, 512)
+infer time 85.82091900000003
+best_threshold_index 135 0.8823809523809524
+best_threshold_index 135 0.8817460317460317
+best_threshold_index 135 0.88
+best_threshold_index 135 0.8844444444444445
+best_threshold_index 135 0.8838095238095238
+best_threshold_index 135 0.8788888888888889
+best_threshold_index 135 0.8795238095238095
+best_threshold_index 136 0.8809523809523809
+best_threshold_index 135 0.8825396825396825
+best_threshold_index 135 0.883015873015873
+[cfp_fp][60000]XNorm: 22.114211
+[cfp_fp][60000]Accuracy-Flip: 0.88143+-0.01578
+testing verification..
+(12000, 512)
+infer time 73.59422199999999
+best_threshold_index 136 0.8790740740740741
+best_threshold_index 136 0.8770370370370371
+best_threshold_index 137 0.8787037037037037
+best_threshold_index 136 0.8803703703703704
+best_threshold_index 136 0.8853703703703704
+best_threshold_index 136 0.8787037037037037
+best_threshold_index 136 0.8803703703703704
+best_threshold_index 136 0.8859259259259259
+best_threshold_index 136 0.8825925925925926
+best_threshold_index 136 0.8805555555555555
+[agedb_30][60000]XNorm: 22.924247
+[agedb_30][60000]Accuracy-Flip: 0.88017+-0.02468
+test accuracy is: 0.9798333333333333
+epoch 3, total_step 60000, total loss is 27.44 , inference loss is 17.48, weight deacy loss is 9.96, training accuracy is 0.218750, time 54.312 samples/sec
+epoch 3, total_step 60020, total loss is 22.74 , inference loss is 12.79, weight deacy loss is 9.96, training accuracy is 0.156250, time 55.900 samples/sec
+epoch 3, total_step 60040, total loss is 24.61 , inference loss is 14.65, weight deacy loss is 9.96, training accuracy is 0.093750, time 56.100 samples/sec
+epoch 3, total_step 60060, total loss is 21.85 , inference loss is 11.89, weight deacy loss is 9.95, training accuracy is 0.187500, time 56.468 samples/sec
+epoch 3, total_step 60080, total loss is 22.92 , inference loss is 12.96, weight deacy loss is 9.95, training accuracy is 0.218750, time 57.517 samples/sec
+epoch 3, total_step 60100, total loss is 22.30 , inference loss is 12.35, weight deacy loss is 9.95, training accuracy is 0.187500, time 56.591 samples/sec
+epoch 3, total_step 60120, total loss is 27.21 , inference loss is 17.26, weight deacy loss is 9.95, training accuracy is 0.156250, time 55.507 samples/sec
+epoch 3, total_step 60140, total loss is 26.68 , inference loss is 16.73, weight deacy loss is 9.95, training accuracy is 0.125000, time 56.521 samples/sec
+epoch 3, total_step 60160, total loss is 21.90 , inference loss is 11.95, weight deacy loss is 9.95, training accuracy is 0.250000, time 55.902 samples/sec
+epoch 3, total_step 60180, total loss is 20.57 , inference loss is 10.62, weight deacy loss is 9.95, training accuracy is 0.218750, time 55.701 samples/sec
+epoch 3, total_step 60200, total loss is 28.80 , inference loss is 18.85, weight deacy loss is 9.95, training accuracy is 0.156250, time 55.760 samples/sec
+epoch 3, total_step 60220, total loss is 24.28 , inference loss is 14.34, weight deacy loss is 9.95, training accuracy is 0.250000, time 56.171 samples/sec
+epoch 3, total_step 60240, total loss is 24.42 , inference loss is 14.48, weight deacy loss is 9.94, training accuracy is 0.093750, time 55.543 samples/sec
+epoch 3, total_step 60260, total loss is 21.07 , inference loss is 11.13, weight deacy loss is 9.94, training accuracy is 0.218750, time 56.573 samples/sec
+epoch 3, total_step 60280, total loss is 21.82 , inference loss is 11.88, weight deacy loss is 9.94, training accuracy is 0.218750, time 55.349 samples/sec
+epoch 3, total_step 60300, total loss is 26.42 , inference loss is 16.48, weight deacy loss is 9.94, training accuracy is 0.125000, time 56.233 samples/sec
+epoch 3, total_step 60320, total loss is 28.02 , inference loss is 18.08, weight deacy loss is 9.94, training accuracy is 0.125000, time 56.622 samples/sec
+epoch 3, total_step 60340, total loss is 22.20 , inference loss is 12.26, weight deacy loss is 9.94, training accuracy is 0.187500, time 56.188 samples/sec
+epoch 3, total_step 60360, total loss is 22.27 , inference loss is 12.33, weight deacy loss is 9.94, training accuracy is 0.250000, time 55.735 samples/sec
+epoch 3, total_step 60380, total loss is 20.87 , inference loss is 10.94, weight deacy loss is 9.94, training accuracy is 0.187500, time 55.531 samples/sec
+epoch 3, total_step 60400, total loss is 22.86 , inference loss is 12.93, weight deacy loss is 9.94, training accuracy is 0.093750, time 55.932 samples/sec
+epoch 3, total_step 60420, total loss is 28.64 , inference loss is 18.70, weight deacy loss is 9.93, training accuracy is 0.093750, time 56.802 samples/sec
+epoch 3, total_step 60440, total loss is 29.40 , inference loss is 19.47, weight deacy loss is 9.93, training accuracy is 0.062500, time 54.101 samples/sec
+epoch 3, total_step 60460, total loss is 23.99 , inference loss is 14.06, weight deacy loss is 9.93, training accuracy is 0.156250, time 54.425 samples/sec
+epoch 3, total_step 60480, total loss is 26.53 , inference loss is 16.60, weight deacy loss is 9.93, training accuracy is 0.062500, time 55.679 samples/sec
+epoch 3, total_step 60500, total loss is 25.42 , inference loss is 15.49, weight deacy loss is 9.93, training accuracy is 0.187500, time 55.970 samples/sec
+epoch 3, total_step 60520, total loss is 26.19 , inference loss is 16.26, weight deacy loss is 9.93, training accuracy is 0.187500, time 57.169 samples/sec
+epoch 3, total_step 60540, total loss is 25.88 , inference loss is 15.95, weight deacy loss is 9.93, training accuracy is 0.156250, time 55.945 samples/sec
+epoch 3, total_step 60560, total loss is 29.25 , inference loss is 19.32, weight deacy loss is 9.93, training accuracy is 0.093750, time 55.678 samples/sec
+epoch 3, total_step 60580, total loss is 26.37 , inference loss is 16.44, weight deacy loss is 9.92, training accuracy is 0.031250, time 55.848 samples/sec
+epoch 3, total_step 60600, total loss is 26.83 , inference loss is 16.90, weight deacy loss is 9.92, training accuracy is 0.062500, time 56.146 samples/sec
+epoch 3, total_step 60620, total loss is 21.84 , inference loss is 11.92, weight deacy loss is 9.92, training accuracy is 0.250000, time 56.160 samples/sec
+epoch 3, total_step 60640, total loss is 23.60 , inference loss is 13.68, weight deacy loss is 9.92, training accuracy is 0.187500, time 57.304 samples/sec
+epoch 3, total_step 60660, total loss is 22.80 , inference loss is 12.88, weight deacy loss is 9.92, training accuracy is 0.218750, time 56.563 samples/sec
+epoch 3, total_step 60680, total loss is 23.23 , inference loss is 13.31, weight deacy loss is 9.92, training accuracy is 0.218750, time 56.418 samples/sec
+epoch 3, total_step 60700, total loss is 24.33 , inference loss is 14.42, weight deacy loss is 9.92, training accuracy is 0.031250, time 55.987 samples/sec
+epoch 3, total_step 60720, total loss is 24.94 , inference loss is 15.03, weight deacy loss is 9.92, training accuracy is 0.031250, time 57.162 samples/sec
+epoch 3, total_step 60740, total loss is 25.85 , inference loss is 15.93, weight deacy loss is 9.92, training accuracy is 0.031250, time 56.392 samples/sec
+epoch 3, total_step 60760, total loss is 23.61 , inference loss is 13.69, weight deacy loss is 9.91, training accuracy is 0.062500, time 55.351 samples/sec
+epoch 3, total_step 60780, total loss is 20.62 , inference loss is 10.70, weight deacy loss is 9.91, training accuracy is 0.187500, time 56.246 samples/sec
+epoch 3, total_step 60800, total loss is 25.32 , inference loss is 15.41, weight deacy loss is 9.91, training accuracy is 0.125000, time 55.757 samples/sec
+epoch 3, total_step 60820, total loss is 22.64 , inference loss is 12.73, weight deacy loss is 9.91, training accuracy is 0.250000, time 56.373 samples/sec
+epoch 3, total_step 60840, total loss is 24.17 , inference loss is 14.26, weight deacy loss is 9.91, training accuracy is 0.125000, time 56.992 samples/sec
+epoch 3, total_step 60860, total loss is 22.97 , inference loss is 13.06, weight deacy loss is 9.91, training accuracy is 0.187500, time 56.744 samples/sec
+epoch 3, total_step 60880, total loss is 24.35 , inference loss is 14.44, weight deacy loss is 9.91, training accuracy is 0.062500, time 55.876 samples/sec
+epoch 3, total_step 60900, total loss is 26.54 , inference loss is 16.63, weight deacy loss is 9.91, training accuracy is 0.062500, time 56.396 samples/sec
+epoch 3, total_step 60920, total loss is 24.02 , inference loss is 14.11, weight deacy loss is 9.91, training accuracy is 0.062500, time 56.681 samples/sec
+epoch 3, total_step 60940, total loss is 30.10 , inference loss is 20.20, weight deacy loss is 9.90, training accuracy is 0.156250, time 56.683 samples/sec
+epoch 3, total_step 60960, total loss is 27.04 , inference loss is 17.13, weight deacy loss is 9.90, training accuracy is 0.062500, time 55.799 samples/sec
+epoch 3, total_step 60980, total loss is 23.88 , inference loss is 13.98, weight deacy loss is 9.90, training accuracy is 0.093750, time 56.737 samples/sec
+epoch 3, total_step 61000, total loss is 27.93 , inference loss is 18.03, weight deacy loss is 9.90, training accuracy is 0.187500, time 56.565 samples/sec
+epoch 3, total_step 61020, total loss is 22.57 , inference loss is 12.67, weight deacy loss is 9.90, training accuracy is 0.187500, time 55.653 samples/sec
+epoch 3, total_step 61040, total loss is 24.24 , inference loss is 14.35, weight deacy loss is 9.90, training accuracy is 0.156250, time 56.406 samples/sec
+epoch 3, total_step 61060, total loss is 23.54 , inference loss is 13.64, weight deacy loss is 9.90, training accuracy is 0.125000, time 55.756 samples/sec
+epoch 3, total_step 61080, total loss is 27.15 , inference loss is 17.25, weight deacy loss is 9.90, training accuracy is 0.156250, time 55.626 samples/sec
+epoch 3, total_step 61100, total loss is 24.60 , inference loss is 14.70, weight deacy loss is 9.90, training accuracy is 0.125000, time 56.002 samples/sec
+epoch 3, total_step 61120, total loss is 23.68 , inference loss is 13.78, weight deacy loss is 9.89, training accuracy is 0.156250, time 55.552 samples/sec
+epoch 3, total_step 61140, total loss is 25.97 , inference loss is 16.08, weight deacy loss is 9.89, training accuracy is 0.156250, time 56.010 samples/sec
+epoch 3, total_step 61160, total loss is 22.70 , inference loss is 12.80, weight deacy loss is 9.89, training accuracy is 0.156250, time 55.860 samples/sec
+epoch 3, total_step 61180, total loss is 23.82 , inference loss is 13.93, weight deacy loss is 9.89, training accuracy is 0.187500, time 55.474 samples/sec
+epoch 3, total_step 61200, total loss is 23.87 , inference loss is 13.98, weight deacy loss is 9.89, training accuracy is 0.187500, time 55.175 samples/sec
+epoch 3, total_step 61220, total loss is 24.01 , inference loss is 14.12, weight deacy loss is 9.89, training accuracy is 0.187500, time 56.144 samples/sec
+epoch 3, total_step 61240, total loss is 23.45 , inference loss is 13.57, weight deacy loss is 9.89, training accuracy is 0.187500, time 56.754 samples/sec
+epoch 3, total_step 61260, total loss is 26.73 , inference loss is 16.85, weight deacy loss is 9.89, training accuracy is 0.156250, time 55.850 samples/sec
+epoch 3, total_step 61280, total loss is 21.44 , inference loss is 11.55, weight deacy loss is 9.89, training accuracy is 0.312500, time 56.383 samples/sec
+epoch 3, total_step 61300, total loss is 26.02 , inference loss is 16.14, weight deacy loss is 9.88, training accuracy is 0.218750, time 55.896 samples/sec
+epoch 3, total_step 61320, total loss is 21.29 , inference loss is 11.41, weight deacy loss is 9.88, training accuracy is 0.187500, time 55.935 samples/sec
+
+
+
+
+End of epoch 3
+epoch 4, total_step 61340, total loss is 28.63 , inference loss is 18.75, weight deacy loss is 9.88, training accuracy is 0.000000, time 55.869 samples/sec
+epoch 4, total_step 61360, total loss is 19.11 , inference loss is 9.23, weight deacy loss is 9.88, training accuracy is 0.375000, time 56.954 samples/sec
+epoch 4, total_step 61380, total loss is 22.07 , inference loss is 12.19, weight deacy loss is 9.88, training accuracy is 0.312500, time 54.656 samples/sec
+epoch 4, total_step 61400, total loss is 19.11 , inference loss is 9.23, weight deacy loss is 9.88, training accuracy is 0.468750, time 56.351 samples/sec
+epoch 4, total_step 61420, total loss is 16.62 , inference loss is 6.75, weight deacy loss is 9.88, training accuracy is 0.343750, time 56.192 samples/sec
+epoch 4, total_step 61440, total loss is 18.08 , inference loss is 8.21, weight deacy loss is 9.88, training accuracy is 0.375000, time 55.706 samples/sec
+epoch 4, total_step 61460, total loss is 22.60 , inference loss is 12.73, weight deacy loss is 9.87, training accuracy is 0.343750, time 54.812 samples/sec
+epoch 4, total_step 61480, total loss is 18.56 , inference loss is 8.69, weight deacy loss is 9.87, training accuracy is 0.468750, time 56.486 samples/sec
+epoch 4, total_step 61500, total loss is 20.92 , inference loss is 11.05, weight deacy loss is 9.87, training accuracy is 0.281250, time 55.633 samples/sec
+epoch 4, total_step 61520, total loss is 20.57 , inference loss is 10.69, weight deacy loss is 9.87, training accuracy is 0.406250, time 53.385 samples/sec
+epoch 4, total_step 61540, total loss is 28.46 , inference loss is 18.59, weight deacy loss is 9.87, training accuracy is 0.312500, time 56.091 samples/sec
+epoch 4, total_step 61560, total loss is 18.61 , inference loss is 8.74, weight deacy loss is 9.87, training accuracy is 0.531250, time 55.349 samples/sec
+epoch 4, total_step 61580, total loss is 24.63 , inference loss is 14.77, weight deacy loss is 9.87, training accuracy is 0.250000, time 55.871 samples/sec
+epoch 4, total_step 61600, total loss is 21.94 , inference loss is 12.07, weight deacy loss is 9.87, training accuracy is 0.468750, time 56.563 samples/sec
+epoch 4, total_step 61620, total loss is 19.14 , inference loss is 9.28, weight deacy loss is 9.86, training accuracy is 0.343750, time 56.670 samples/sec
+epoch 4, total_step 61640, total loss is 22.65 , inference loss is 12.79, weight deacy loss is 9.86, training accuracy is 0.375000, time 55.805 samples/sec
+epoch 4, total_step 61660, total loss is 19.15 , inference loss is 9.29, weight deacy loss is 9.86, training accuracy is 0.500000, time 55.839 samples/sec
+epoch 4, total_step 61680, total loss is 21.03 , inference loss is 11.17, weight deacy loss is 9.86, training accuracy is 0.375000, time 56.059 samples/sec
+epoch 4, total_step 61700, total loss is 16.67 , inference loss is 6.81, weight deacy loss is 9.86, training accuracy is 0.343750, time 56.149 samples/sec
+epoch 4, total_step 61720, total loss is 22.19 , inference loss is 12.33, weight deacy loss is 9.86, training accuracy is 0.343750, time 55.817 samples/sec
+epoch 4, total_step 61740, total loss is 20.11 , inference loss is 10.25, weight deacy loss is 9.86, training accuracy is 0.281250, time 56.487 samples/sec
+epoch 4, total_step 61760, total loss is 16.98 , inference loss is 7.13, weight deacy loss is 9.86, training accuracy is 0.406250, time 56.484 samples/sec
+epoch 4, total_step 61780, total loss is 19.22 , inference loss is 9.36, weight deacy loss is 9.86, training accuracy is 0.406250, time 55.182 samples/sec
+epoch 4, total_step 61800, total loss is 20.98 , inference loss is 11.13, weight deacy loss is 9.85, training accuracy is 0.406250, time 55.484 samples/sec
+epoch 4, total_step 61820, total loss is 18.81 , inference loss is 8.95, weight deacy loss is 9.85, training accuracy is 0.343750, time 56.043 samples/sec
+epoch 4, total_step 61840, total loss is 19.75 , inference loss is 9.90, weight deacy loss is 9.85, training accuracy is 0.250000, time 56.808 samples/sec
+epoch 4, total_step 61860, total loss is 21.42 , inference loss is 11.57, weight deacy loss is 9.85, training accuracy is 0.312500, time 56.861 samples/sec
+epoch 4, total_step 61880, total loss is 23.84 , inference loss is 13.99, weight deacy loss is 9.85, training accuracy is 0.343750, time 56.442 samples/sec
+epoch 4, total_step 61900, total loss is 20.64 , inference loss is 10.79, weight deacy loss is 9.85, training accuracy is 0.375000, time 56.878 samples/sec
+epoch 4, total_step 61920, total loss is 22.70 , inference loss is 12.85, weight deacy loss is 9.85, training accuracy is 0.281250, time 56.678 samples/sec
+epoch 4, total_step 61940, total loss is 18.30 , inference loss is 8.45, weight deacy loss is 9.85, training accuracy is 0.531250, time 56.079 samples/sec
+epoch 4, total_step 61960, total loss is 20.75 , inference loss is 10.91, weight deacy loss is 9.85, training accuracy is 0.343750, time 56.439 samples/sec
+epoch 4, total_step 61980, total loss is 22.99 , inference loss is 13.14, weight deacy loss is 9.84, training accuracy is 0.312500, time 56.310 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.47836600000004
+best_threshold_index 110 0.9827777777777778
+best_threshold_index 110 0.9829629629629629
+best_threshold_index 111 0.9827777777777778
+best_threshold_index 110 0.9838888888888889
+best_threshold_index 110 0.9833333333333333
+best_threshold_index 111 0.9818518518518519
+best_threshold_index 110 0.9833333333333333
+best_threshold_index 110 0.9825925925925926
+best_threshold_index 110 0.9818518518518519
+best_threshold_index 110 0.9816666666666667
+[lfw][62000]XNorm: 21.860537
+[lfw][62000]Accuracy-Flip: 0.98200+-0.00609
+testing verification..
+(14000, 512)
+infer time 85.53581399999992
+best_threshold_index 110 0.976984126984127
+best_threshold_index 110 0.9744444444444444
+best_threshold_index 106 0.9747619047619047
+best_threshold_index 106 0.9731746031746031
+best_threshold_index 106 0.9734920634920635
+best_threshold_index 112 0.9741269841269842
+best_threshold_index 111 0.9755555555555555
+best_threshold_index 112 0.9739682539682539
+best_threshold_index 106 0.9746031746031746
+best_threshold_index 108 0.9755555555555555
+[cfp_ff][62000]XNorm: 21.708693
+[cfp_ff][62000]Accuracy-Flip: 0.97200+-0.00990
+testing verification..
+(14000, 512)
+infer time 86.6819300000001
+best_threshold_index 133 0.8860317460317461
+best_threshold_index 128 0.8823809523809524
+best_threshold_index 133 0.8825396825396825
+best_threshold_index 133 0.8841269841269841
+best_threshold_index 133 0.8853968253968254
+best_threshold_index 131 0.8803174603174603
+best_threshold_index 131 0.8826984126984126
+best_threshold_index 133 0.8841269841269841
+best_threshold_index 133 0.8842857142857142
+best_threshold_index 132 0.883015873015873
+[cfp_fp][62000]XNorm: 21.323154
+[cfp_fp][62000]Accuracy-Flip: 0.88143+-0.01311
+testing verification..
+(12000, 512)
+infer time 74.90680600000002
+best_threshold_index 134 0.8781481481481481
+best_threshold_index 135 0.8775925925925926
+best_threshold_index 134 0.8809259259259259
+best_threshold_index 134 0.8807407407407407
+best_threshold_index 134 0.8842592592592593
+best_threshold_index 134 0.8770370370370371
+best_threshold_index 134 0.8818518518518519
+best_threshold_index 135 0.8838888888888888
+best_threshold_index 134 0.8824074074074074
+best_threshold_index 134 0.8818518518518519
+[agedb_30][62000]XNorm: 22.269869
+[agedb_30][62000]Accuracy-Flip: 0.88017+-0.02162
+test accuracy is: 0.982
+epoch 4, total_step 62000, total loss is 20.98 , inference loss is 11.14, weight deacy loss is 9.84, training accuracy is 0.281250, time 53.283 samples/sec
+epoch 4, total_step 62020, total loss is 19.55 , inference loss is 9.70, weight deacy loss is 9.84, training accuracy is 0.437500, time 55.817 samples/sec
+epoch 4, total_step 62040, total loss is 22.68 , inference loss is 12.84, weight deacy loss is 9.84, training accuracy is 0.406250, time 56.274 samples/sec
+epoch 4, total_step 62060, total loss is 20.03 , inference loss is 10.19, weight deacy loss is 9.84, training accuracy is 0.437500, time 55.960 samples/sec
+epoch 4, total_step 62080, total loss is 20.47 , inference loss is 10.64, weight deacy loss is 9.84, training accuracy is 0.468750, time 56.611 samples/sec
+epoch 4, total_step 62100, total loss is 21.80 , inference loss is 11.96, weight deacy loss is 9.84, training accuracy is 0.468750, time 55.256 samples/sec
+epoch 4, total_step 62120, total loss is 24.07 , inference loss is 14.24, weight deacy loss is 9.84, training accuracy is 0.093750, time 56.513 samples/sec
+epoch 4, total_step 62140, total loss is 27.58 , inference loss is 17.75, weight deacy loss is 9.84, training accuracy is 0.281250, time 56.038 samples/sec
+epoch 4, total_step 62160, total loss is 18.30 , inference loss is 8.47, weight deacy loss is 9.83, training accuracy is 0.406250, time 54.968 samples/sec
+epoch 4, total_step 62180, total loss is 19.23 , inference loss is 9.39, weight deacy loss is 9.83, training accuracy is 0.312500, time 56.002 samples/sec
+epoch 4, total_step 62200, total loss is 21.04 , inference loss is 11.21, weight deacy loss is 9.83, training accuracy is 0.312500, time 56.152 samples/sec
+epoch 4, total_step 62220, total loss is 22.27 , inference loss is 12.44, weight deacy loss is 9.83, training accuracy is 0.312500, time 56.349 samples/sec
+epoch 4, total_step 62240, total loss is 24.71 , inference loss is 14.88, weight deacy loss is 9.83, training accuracy is 0.218750, time 54.721 samples/sec
+epoch 4, total_step 62260, total loss is 18.85 , inference loss is 9.02, weight deacy loss is 9.83, training accuracy is 0.343750, time 56.745 samples/sec
+epoch 4, total_step 62280, total loss is 25.13 , inference loss is 15.30, weight deacy loss is 9.83, training accuracy is 0.343750, time 57.054 samples/sec
+epoch 4, total_step 62300, total loss is 21.56 , inference loss is 11.73, weight deacy loss is 9.83, training accuracy is 0.187500, time 55.638 samples/sec
+epoch 4, total_step 62320, total loss is 26.44 , inference loss is 16.62, weight deacy loss is 9.83, training accuracy is 0.312500, time 56.554 samples/sec
+epoch 4, total_step 62340, total loss is 21.63 , inference loss is 11.81, weight deacy loss is 9.82, training accuracy is 0.281250, time 54.572 samples/sec
+epoch 4, total_step 62360, total loss is 19.40 , inference loss is 9.58, weight deacy loss is 9.82, training accuracy is 0.375000, time 56.749 samples/sec
+epoch 4, total_step 62380, total loss is 21.25 , inference loss is 11.43, weight deacy loss is 9.82, training accuracy is 0.250000, time 54.969 samples/sec
+epoch 4, total_step 62400, total loss is 17.86 , inference loss is 8.04, weight deacy loss is 9.82, training accuracy is 0.468750, time 56.422 samples/sec
+epoch 4, total_step 62420, total loss is 21.66 , inference loss is 11.84, weight deacy loss is 9.82, training accuracy is 0.406250, time 57.023 samples/sec
+epoch 4, total_step 62440, total loss is 21.18 , inference loss is 11.37, weight deacy loss is 9.82, training accuracy is 0.281250, time 55.908 samples/sec
+epoch 4, total_step 62460, total loss is 20.89 , inference loss is 11.07, weight deacy loss is 9.82, training accuracy is 0.218750, time 56.000 samples/sec
+epoch 4, total_step 62480, total loss is 22.71 , inference loss is 12.90, weight deacy loss is 9.82, training accuracy is 0.406250, time 56.728 samples/sec
+epoch 4, total_step 62500, total loss is 24.24 , inference loss is 14.43, weight deacy loss is 9.82, training accuracy is 0.312500, time 56.323 samples/sec
+epoch 4, total_step 62520, total loss is 23.72 , inference loss is 13.91, weight deacy loss is 9.81, training accuracy is 0.250000, time 56.756 samples/sec
+epoch 4, total_step 62540, total loss is 24.97 , inference loss is 15.16, weight deacy loss is 9.81, training accuracy is 0.281250, time 57.022 samples/sec
+epoch 4, total_step 62560, total loss is 21.53 , inference loss is 11.72, weight deacy loss is 9.81, training accuracy is 0.093750, time 57.605 samples/sec
+epoch 4, total_step 62580, total loss is 19.02 , inference loss is 9.21, weight deacy loss is 9.81, training accuracy is 0.437500, time 55.309 samples/sec
+epoch 4, total_step 62600, total loss is 19.50 , inference loss is 9.69, weight deacy loss is 9.81, training accuracy is 0.312500, time 55.887 samples/sec
+epoch 4, total_step 62620, total loss is 26.54 , inference loss is 16.73, weight deacy loss is 9.81, training accuracy is 0.156250, time 55.955 samples/sec
+epoch 4, total_step 62640, total loss is 23.89 , inference loss is 14.09, weight deacy loss is 9.81, training accuracy is 0.187500, time 54.333 samples/sec
+epoch 4, total_step 62660, total loss is 20.92 , inference loss is 11.11, weight deacy loss is 9.81, training accuracy is 0.250000, time 56.414 samples/sec
+epoch 4, total_step 62680, total loss is 17.26 , inference loss is 7.46, weight deacy loss is 9.81, training accuracy is 0.312500, time 56.437 samples/sec
+epoch 4, total_step 62700, total loss is 23.72 , inference loss is 13.91, weight deacy loss is 9.80, training accuracy is 0.250000, time 55.737 samples/sec
+epoch 4, total_step 62720, total loss is 26.57 , inference loss is 16.76, weight deacy loss is 9.80, training accuracy is 0.218750, time 56.545 samples/sec
+epoch 4, total_step 62740, total loss is 24.23 , inference loss is 14.43, weight deacy loss is 9.80, training accuracy is 0.156250, time 56.801 samples/sec
+epoch 4, total_step 62760, total loss is 22.29 , inference loss is 12.49, weight deacy loss is 9.80, training accuracy is 0.093750, time 56.636 samples/sec
+epoch 4, total_step 62780, total loss is 22.14 , inference loss is 12.34, weight deacy loss is 9.80, training accuracy is 0.218750, time 54.663 samples/sec
+epoch 4, total_step 62800, total loss is 24.83 , inference loss is 15.03, weight deacy loss is 9.80, training accuracy is 0.187500, time 56.481 samples/sec
+epoch 4, total_step 62820, total loss is 24.12 , inference loss is 14.32, weight deacy loss is 9.80, training accuracy is 0.218750, time 56.734 samples/sec
+epoch 4, total_step 62840, total loss is 19.52 , inference loss is 9.73, weight deacy loss is 9.80, training accuracy is 0.312500, time 55.968 samples/sec
+epoch 4, total_step 62860, total loss is 18.79 , inference loss is 8.99, weight deacy loss is 9.80, training accuracy is 0.312500, time 56.573 samples/sec
+epoch 4, total_step 62880, total loss is 25.12 , inference loss is 15.32, weight deacy loss is 9.79, training accuracy is 0.156250, time 55.454 samples/sec
+epoch 4, total_step 62900, total loss is 18.24 , inference loss is 8.45, weight deacy loss is 9.79, training accuracy is 0.437500, time 56.172 samples/sec
+epoch 4, total_step 62920, total loss is 20.07 , inference loss is 10.28, weight deacy loss is 9.79, training accuracy is 0.187500, time 56.004 samples/sec
+epoch 4, total_step 62940, total loss is 22.08 , inference loss is 12.29, weight deacy loss is 9.79, training accuracy is 0.281250, time 55.670 samples/sec
+epoch 4, total_step 62960, total loss is 22.13 , inference loss is 12.34, weight deacy loss is 9.79, training accuracy is 0.250000, time 56.042 samples/sec
+epoch 4, total_step 62980, total loss is 24.35 , inference loss is 14.56, weight deacy loss is 9.79, training accuracy is 0.218750, time 56.087 samples/sec
+epoch 4, total_step 63000, total loss is 20.91 , inference loss is 11.12, weight deacy loss is 9.79, training accuracy is 0.250000, time 54.190 samples/sec
+epoch 4, total_step 63020, total loss is 22.41 , inference loss is 12.63, weight deacy loss is 9.79, training accuracy is 0.468750, time 56.147 samples/sec
+epoch 4, total_step 63040, total loss is 25.01 , inference loss is 15.23, weight deacy loss is 9.79, training accuracy is 0.281250, time 57.052 samples/sec
+epoch 4, total_step 63060, total loss is 21.25 , inference loss is 11.46, weight deacy loss is 9.78, training accuracy is 0.250000, time 56.402 samples/sec
+epoch 4, total_step 63080, total loss is 24.01 , inference loss is 14.22, weight deacy loss is 9.78, training accuracy is 0.187500, time 55.937 samples/sec
+epoch 4, total_step 63100, total loss is 23.49 , inference loss is 13.71, weight deacy loss is 9.78, training accuracy is 0.281250, time 54.893 samples/sec
+epoch 4, total_step 63120, total loss is 22.07 , inference loss is 12.29, weight deacy loss is 9.78, training accuracy is 0.187500, time 56.068 samples/sec
+epoch 4, total_step 63140, total loss is 22.64 , inference loss is 12.86, weight deacy loss is 9.78, training accuracy is 0.281250, time 56.454 samples/sec
+epoch 4, total_step 63160, total loss is 21.97 , inference loss is 12.19, weight deacy loss is 9.78, training accuracy is 0.250000, time 56.315 samples/sec
+epoch 4, total_step 63180, total loss is 19.74 , inference loss is 9.96, weight deacy loss is 9.78, training accuracy is 0.250000, time 56.819 samples/sec
+epoch 4, total_step 63200, total loss is 20.32 , inference loss is 10.54, weight deacy loss is 9.78, training accuracy is 0.375000, time 55.775 samples/sec
+epoch 4, total_step 63220, total loss is 21.17 , inference loss is 11.40, weight deacy loss is 9.77, training accuracy is 0.406250, time 56.582 samples/sec
+epoch 4, total_step 63240, total loss is 20.84 , inference loss is 11.07, weight deacy loss is 9.77, training accuracy is 0.281250, time 56.374 samples/sec
+epoch 4, total_step 63260, total loss is 18.56 , inference loss is 8.79, weight deacy loss is 9.77, training accuracy is 0.531250, time 55.540 samples/sec
+epoch 4, total_step 63280, total loss is 22.07 , inference loss is 12.30, weight deacy loss is 9.77, training accuracy is 0.250000, time 56.140 samples/sec
+epoch 4, total_step 63300, total loss is 18.25 , inference loss is 8.48, weight deacy loss is 9.77, training accuracy is 0.343750, time 56.408 samples/sec
+epoch 4, total_step 63320, total loss is 23.55 , inference loss is 13.78, weight deacy loss is 9.77, training accuracy is 0.218750, time 56.582 samples/sec
+epoch 4, total_step 63340, total loss is 21.35 , inference loss is 11.58, weight deacy loss is 9.77, training accuracy is 0.375000, time 55.974 samples/sec
+epoch 4, total_step 63360, total loss is 17.64 , inference loss is 7.87, weight deacy loss is 9.77, training accuracy is 0.468750, time 53.843 samples/sec
+epoch 4, total_step 63380, total loss is 22.09 , inference loss is 12.32, weight deacy loss is 9.77, training accuracy is 0.406250, time 56.321 samples/sec
+epoch 4, total_step 63400, total loss is 24.25 , inference loss is 14.48, weight deacy loss is 9.76, training accuracy is 0.187500, time 55.674 samples/sec
+epoch 4, total_step 63420, total loss is 20.55 , inference loss is 10.79, weight deacy loss is 9.76, training accuracy is 0.312500, time 57.591 samples/sec
+epoch 4, total_step 63440, total loss is 24.69 , inference loss is 14.92, weight deacy loss is 9.76, training accuracy is 0.187500, time 56.584 samples/sec
+epoch 4, total_step 63460, total loss is 21.17 , inference loss is 11.41, weight deacy loss is 9.76, training accuracy is 0.281250, time 56.248 samples/sec
+epoch 4, total_step 63480, total loss is 22.83 , inference loss is 13.07, weight deacy loss is 9.76, training accuracy is 0.156250, time 56.479 samples/sec
+epoch 4, total_step 63500, total loss is 19.90 , inference loss is 10.15, weight deacy loss is 9.76, training accuracy is 0.250000, time 56.562 samples/sec
+epoch 4, total_step 63520, total loss is 22.02 , inference loss is 12.26, weight deacy loss is 9.76, training accuracy is 0.312500, time 55.631 samples/sec
+epoch 4, total_step 63540, total loss is 17.75 , inference loss is 7.99, weight deacy loss is 9.76, training accuracy is 0.343750, time 56.126 samples/sec
+epoch 4, total_step 63560, total loss is 22.77 , inference loss is 13.02, weight deacy loss is 9.75, training accuracy is 0.218750, time 56.548 samples/sec
+epoch 4, total_step 63580, total loss is 19.19 , inference loss is 9.43, weight deacy loss is 9.75, training accuracy is 0.218750, time 55.535 samples/sec
+epoch 4, total_step 63600, total loss is 18.18 , inference loss is 8.43, weight deacy loss is 9.75, training accuracy is 0.406250, time 55.925 samples/sec
+epoch 4, total_step 63620, total loss is 17.58 , inference loss is 7.83, weight deacy loss is 9.75, training accuracy is 0.437500, time 55.017 samples/sec
+epoch 4, total_step 63640, total loss is 20.06 , inference loss is 10.31, weight deacy loss is 9.75, training accuracy is 0.312500, time 56.532 samples/sec
+epoch 4, total_step 63660, total loss is 25.42 , inference loss is 15.67, weight deacy loss is 9.75, training accuracy is 0.187500, time 56.416 samples/sec
+epoch 4, total_step 63680, total loss is 15.53 , inference loss is 5.78, weight deacy loss is 9.75, training accuracy is 0.437500, time 55.465 samples/sec
+epoch 4, total_step 63700, total loss is 16.78 , inference loss is 7.04, weight deacy loss is 9.75, training accuracy is 0.437500, time 55.837 samples/sec
+epoch 4, total_step 63720, total loss is 21.66 , inference loss is 11.91, weight deacy loss is 9.75, training accuracy is 0.281250, time 56.418 samples/sec
+epoch 4, total_step 63740, total loss is 15.99 , inference loss is 6.25, weight deacy loss is 9.74, training accuracy is 0.468750, time 56.605 samples/sec
+epoch 4, total_step 63760, total loss is 16.13 , inference loss is 6.38, weight deacy loss is 9.74, training accuracy is 0.375000, time 56.141 samples/sec
+epoch 4, total_step 63780, total loss is 16.78 , inference loss is 7.04, weight deacy loss is 9.74, training accuracy is 0.406250, time 56.718 samples/sec
+epoch 4, total_step 63800, total loss is 18.95 , inference loss is 9.21, weight deacy loss is 9.74, training accuracy is 0.500000, time 55.960 samples/sec
+epoch 4, total_step 63820, total loss is 21.03 , inference loss is 11.29, weight deacy loss is 9.74, training accuracy is 0.312500, time 56.501 samples/sec
+epoch 4, total_step 63840, total loss is 20.13 , inference loss is 10.39, weight deacy loss is 9.74, training accuracy is 0.437500, time 55.026 samples/sec
+epoch 4, total_step 63860, total loss is 21.45 , inference loss is 11.72, weight deacy loss is 9.74, training accuracy is 0.375000, time 56.683 samples/sec
+epoch 4, total_step 63880, total loss is 19.25 , inference loss is 9.52, weight deacy loss is 9.74, training accuracy is 0.312500, time 55.660 samples/sec
+epoch 4, total_step 63900, total loss is 15.11 , inference loss is 5.37, weight deacy loss is 9.73, training accuracy is 0.500000, time 54.015 samples/sec
+epoch 4, total_step 63920, total loss is 16.99 , inference loss is 7.26, weight deacy loss is 9.73, training accuracy is 0.343750, time 56.339 samples/sec
+epoch 4, total_step 63940, total loss is 17.83 , inference loss is 8.10, weight deacy loss is 9.73, training accuracy is 0.281250, time 53.889 samples/sec
+epoch 4, total_step 63960, total loss is 21.25 , inference loss is 11.52, weight deacy loss is 9.73, training accuracy is 0.281250, time 56.973 samples/sec
+epoch 4, total_step 63980, total loss is 19.94 , inference loss is 10.21, weight deacy loss is 9.73, training accuracy is 0.312500, time 56.261 samples/sec
+testing verification..
+(12000, 512)
+infer time 75.32356900000002
+best_threshold_index 115 0.9814814814814815
+best_threshold_index 116 0.9811111111111112
+best_threshold_index 116 0.9816666666666667
+best_threshold_index 108 0.9822222222222222
+best_threshold_index 116 0.9816666666666667
+best_threshold_index 116 0.9805555555555555
+best_threshold_index 116 0.980925925925926
+best_threshold_index 116 0.980925925925926
+best_threshold_index 110 0.9805555555555555
+best_threshold_index 115 0.9805555555555555
+[lfw][64000]XNorm: 22.267259
+[lfw][64000]Accuracy-Flip: 0.97983+-0.00555
+testing verification..
+(14000, 512)
+infer time 86.91874500000002
+best_threshold_index 116 0.9831746031746031
+best_threshold_index 116 0.9826984126984127
+best_threshold_index 116 0.9820634920634921
+best_threshold_index 116 0.9823809523809524
+best_threshold_index 116 0.9815873015873016
+best_threshold_index 116 0.9809523809523809
+best_threshold_index 116 0.9823809523809524
+best_threshold_index 116 0.9806349206349206
+best_threshold_index 116 0.9817460317460317
+best_threshold_index 116 0.9823809523809524
+[cfp_ff][64000]XNorm: 21.451641
+[cfp_ff][64000]Accuracy-Flip: 0.98200+-0.00668
+testing verification..
+(14000, 512)
+infer time 86.75229600000002
+best_threshold_index 133 0.8909523809523809
+best_threshold_index 133 0.8885714285714286
+best_threshold_index 133 0.8884126984126984
+best_threshold_index 133 0.8895238095238095
+best_threshold_index 133 0.8907936507936508
+best_threshold_index 133 0.8847619047619047
+best_threshold_index 133 0.8868253968253968
+best_threshold_index 133 0.8888888888888888
+best_threshold_index 133 0.8885714285714286
+best_threshold_index 133 0.8884126984126984
+[cfp_fp][64000]XNorm: 21.831577
+[cfp_fp][64000]Accuracy-Flip: 0.88857+-0.01539
+testing verification..
+(12000, 512)
+infer time 74.16355100000001
+best_threshold_index 135 0.8820370370370371
+best_threshold_index 135 0.8822222222222222
+best_threshold_index 134 0.8861111111111111
+best_threshold_index 135 0.882962962962963
+best_threshold_index 135 0.8918518518518519
+best_threshold_index 134 0.8864814814814815
+best_threshold_index 135 0.885
+best_threshold_index 134 0.8898148148148148
+best_threshold_index 135 0.887962962962963
+best_threshold_index 135 0.8855555555555555
+[agedb_30][64000]XNorm: 22.163086
+[agedb_30][64000]Accuracy-Flip: 0.88500+-0.02801
+test accuracy is: 0.9798333333333333
+epoch 4, total_step 64000, total loss is 18.46 , inference loss is 8.73, weight deacy loss is 9.73, training accuracy is 0.437500, time 53.846 samples/sec
+epoch 4, total_step 64020, total loss is 20.26 , inference loss is 10.54, weight deacy loss is 9.73, training accuracy is 0.343750, time 56.277 samples/sec
+epoch 4, total_step 64040, total loss is 19.79 , inference loss is 10.07, weight deacy loss is 9.73, training accuracy is 0.312500, time 56.989 samples/sec
+epoch 4, total_step 64060, total loss is 18.05 , inference loss is 8.33, weight deacy loss is 9.72, training accuracy is 0.343750, time 57.308 samples/sec
+epoch 4, total_step 64080, total loss is 21.19 , inference loss is 11.47, weight deacy loss is 9.72, training accuracy is 0.312500, time 55.418 samples/sec
+epoch 4, total_step 64100, total loss is 20.59 , inference loss is 10.87, weight deacy loss is 9.72, training accuracy is 0.375000, time 56.781 samples/sec
+epoch 4, total_step 64120, total loss is 17.47 , inference loss is 7.75, weight deacy loss is 9.72, training accuracy is 0.437500, time 55.650 samples/sec
+epoch 4, total_step 64140, total loss is 18.86 , inference loss is 9.14, weight deacy loss is 9.72, training accuracy is 0.281250, time 56.415 samples/sec
+epoch 4, total_step 64160, total loss is 17.27 , inference loss is 7.55, weight deacy loss is 9.72, training accuracy is 0.500000, time 56.691 samples/sec
+epoch 4, total_step 64180, total loss is 19.13 , inference loss is 9.41, weight deacy loss is 9.72, training accuracy is 0.437500, time 56.452 samples/sec
+epoch 4, total_step 64200, total loss is 17.57 , inference loss is 7.85, weight deacy loss is 9.72, training accuracy is 0.281250, time 55.874 samples/sec
+epoch 4, total_step 64220, total loss is 16.28 , inference loss is 6.56, weight deacy loss is 9.72, training accuracy is 0.531250, time 55.167 samples/sec
+epoch 4, total_step 64240, total loss is 19.29 , inference loss is 9.58, weight deacy loss is 9.71, training accuracy is 0.406250, time 56.419 samples/sec
+epoch 4, total_step 64260, total loss is 18.83 , inference loss is 9.11, weight deacy loss is 9.71, training accuracy is 0.437500, time 55.120 samples/sec
+epoch 4, total_step 64280, total loss is 15.74 , inference loss is 6.03, weight deacy loss is 9.71, training accuracy is 0.406250, time 55.721 samples/sec
+epoch 4, total_step 64300, total loss is 15.20 , inference loss is 5.49, weight deacy loss is 9.71, training accuracy is 0.437500, time 55.730 samples/sec
+epoch 4, total_step 64320, total loss is 17.52 , inference loss is 7.81, weight deacy loss is 9.71, training accuracy is 0.437500, time 56.142 samples/sec
+epoch 4, total_step 64340, total loss is 20.31 , inference loss is 10.60, weight deacy loss is 9.71, training accuracy is 0.375000, time 56.324 samples/sec
+epoch 4, total_step 64360, total loss is 17.31 , inference loss is 7.61, weight deacy loss is 9.71, training accuracy is 0.375000, time 55.664 samples/sec
+epoch 4, total_step 64380, total loss is 14.92 , inference loss is 5.21, weight deacy loss is 9.71, training accuracy is 0.562500, time 56.175 samples/sec
+epoch 4, total_step 64400, total loss is 17.71 , inference loss is 8.01, weight deacy loss is 9.71, training accuracy is 0.375000, time 55.907 samples/sec
+epoch 4, total_step 64420, total loss is 19.39 , inference loss is 9.69, weight deacy loss is 9.70, training accuracy is 0.343750, time 56.389 samples/sec
+epoch 4, total_step 64440, total loss is 21.89 , inference loss is 12.19, weight deacy loss is 9.70, training accuracy is 0.281250, time 56.505 samples/sec
+epoch 4, total_step 64460, total loss is 21.89 , inference loss is 12.19, weight deacy loss is 9.70, training accuracy is 0.281250, time 55.842 samples/sec
+epoch 4, total_step 64480, total loss is 24.67 , inference loss is 14.97, weight deacy loss is 9.70, training accuracy is 0.343750, time 55.643 samples/sec
+epoch 4, total_step 64500, total loss is 23.09 , inference loss is 13.39, weight deacy loss is 9.70, training accuracy is 0.312500, time 55.040 samples/sec
+epoch 4, total_step 64520, total loss is 22.02 , inference loss is 12.32, weight deacy loss is 9.70, training accuracy is 0.343750, time 56.179 samples/sec
+epoch 4, total_step 64540, total loss is 19.92 , inference loss is 10.22, weight deacy loss is 9.70, training accuracy is 0.312500, time 55.824 samples/sec
+epoch 4, total_step 64560, total loss is 19.95 , inference loss is 10.26, weight deacy loss is 9.70, training accuracy is 0.312500, time 55.371 samples/sec
+epoch 4, total_step 64580, total loss is 21.93 , inference loss is 12.23, weight deacy loss is 9.70, training accuracy is 0.343750, time 57.081 samples/sec
+epoch 4, total_step 64600, total loss is 18.31 , inference loss is 8.61, weight deacy loss is 9.69, training accuracy is 0.406250, time 55.866 samples/sec
+epoch 4, total_step 64620, total loss is 21.56 , inference loss is 11.86, weight deacy loss is 9.69, training accuracy is 0.156250, time 54.381 samples/sec
+epoch 4, total_step 64640, total loss is 23.09 , inference loss is 13.40, weight deacy loss is 9.69, training accuracy is 0.312500, time 56.560 samples/sec
+epoch 4, total_step 64660, total loss is 20.74 , inference loss is 11.05, weight deacy loss is 9.69, training accuracy is 0.437500, time 55.810 samples/sec
+epoch 4, total_step 64680, total loss is 20.07 , inference loss is 10.38, weight deacy loss is 9.69, training accuracy is 0.312500, time 57.110 samples/sec
+epoch 4, total_step 64700, total loss is 22.54 , inference loss is 12.85, weight deacy loss is 9.69, training accuracy is 0.312500, time 55.530 samples/sec
+epoch 4, total_step 64720, total loss is 24.65 , inference loss is 14.96, weight deacy loss is 9.69, training accuracy is 0.281250, time 56.526 samples/sec
+epoch 4, total_step 64740, total loss is 22.40 , inference loss is 12.71, weight deacy loss is 9.69, training accuracy is 0.156250, time 55.620 samples/sec
+epoch 4, total_step 64760, total loss is 19.78 , inference loss is 10.10, weight deacy loss is 9.69, training accuracy is 0.312500, time 55.522 samples/sec
+epoch 4, total_step 64780, total loss is 18.47 , inference loss is 8.79, weight deacy loss is 9.68, training accuracy is 0.375000, time 55.128 samples/sec
+epoch 4, total_step 64800, total loss is 22.25 , inference loss is 12.57, weight deacy loss is 9.68, training accuracy is 0.250000, time 56.516 samples/sec
+epoch 4, total_step 64820, total loss is 20.40 , inference loss is 10.72, weight deacy loss is 9.68, training accuracy is 0.250000, time 56.285 samples/sec
+epoch 4, total_step 64840, total loss is 22.05 , inference loss is 12.37, weight deacy loss is 9.68, training accuracy is 0.218750, time 55.629 samples/sec
+epoch 4, total_step 64860, total loss is 22.35 , inference loss is 12.67, weight deacy loss is 9.68, training accuracy is 0.281250, time 56.215 samples/sec
+epoch 4, total_step 64880, total loss is 21.32 , inference loss is 11.64, weight deacy loss is 9.68, training accuracy is 0.375000, time 56.024 samples/sec
+epoch 4, total_step 64900, total loss is 20.74 , inference loss is 11.06, weight deacy loss is 9.68, training accuracy is 0.218750, time 56.558 samples/sec
+epoch 4, total_step 64920, total loss is 22.28 , inference loss is 12.60, weight deacy loss is 9.68, training accuracy is 0.187500, time 56.324 samples/sec
+epoch 4, total_step 64940, total loss is 24.64 , inference loss is 14.96, weight deacy loss is 9.68, training accuracy is 0.125000, time 56.538 samples/sec
+epoch 4, total_step 64960, total loss is 23.67 , inference loss is 13.99, weight deacy loss is 9.67, training accuracy is 0.250000, time 56.231 samples/sec
+epoch 4, total_step 64980, total loss is 22.71 , inference loss is 13.04, weight deacy loss is 9.67, training accuracy is 0.281250, time 55.715 samples/sec
+epoch 4, total_step 65000, total loss is 22.24 , inference loss is 12.57, weight deacy loss is 9.67, training accuracy is 0.312500, time 56.265 samples/sec
+epoch 4, total_step 65020, total loss is 17.80 , inference loss is 8.13, weight deacy loss is 9.67, training accuracy is 0.281250, time 56.257 samples/sec
+epoch 4, total_step 65040, total loss is 25.21 , inference loss is 15.54, weight deacy loss is 9.67, training accuracy is 0.343750, time 55.549 samples/sec
+epoch 4, total_step 65060, total loss is 21.58 , inference loss is 11.91, weight deacy loss is 9.67, training accuracy is 0.187500, time 55.651 samples/sec
+epoch 4, total_step 65080, total loss is 24.91 , inference loss is 15.24, weight deacy loss is 9.67, training accuracy is 0.218750, time 56.433 samples/sec
+epoch 4, total_step 65100, total loss is 22.93 , inference loss is 13.26, weight deacy loss is 9.67, training accuracy is 0.312500, time 54.922 samples/sec
+epoch 4, total_step 65120, total loss is 21.70 , inference loss is 12.04, weight deacy loss is 9.67, training accuracy is 0.343750, time 55.960 samples/sec
+epoch 4, total_step 65140, total loss is 20.94 , inference loss is 11.27, weight deacy loss is 9.66, training accuracy is 0.250000, time 56.868 samples/sec
+epoch 4, total_step 65160, total loss is 21.84 , inference loss is 12.17, weight deacy loss is 9.66, training accuracy is 0.281250, time 56.484 samples/sec
+epoch 4, total_step 65180, total loss is 21.52 , inference loss is 11.85, weight deacy loss is 9.66, training accuracy is 0.281250, time 56.475 samples/sec
+epoch 4, total_step 65200, total loss is 19.54 , inference loss is 9.88, weight deacy loss is 9.66, training accuracy is 0.343750, time 56.160 samples/sec
+epoch 4, total_step 65220, total loss is 21.85 , inference loss is 12.19, weight deacy loss is 9.66, training accuracy is 0.281250, time 56.073 samples/sec
+epoch 4, total_step 65240, total loss is 20.84 , inference loss is 11.18, weight deacy loss is 9.66, training accuracy is 0.250000, time 55.906 samples/sec
+epoch 4, total_step 65260, total loss is 17.77 , inference loss is 8.11, weight deacy loss is 9.66, training accuracy is 0.406250, time 56.339 samples/sec
+epoch 4, total_step 65280, total loss is 22.50 , inference loss is 12.84, weight deacy loss is 9.66, training accuracy is 0.218750, time 55.928 samples/sec
+epoch 4, total_step 65300, total loss is 21.18 , inference loss is 11.52, weight deacy loss is 9.66, training accuracy is 0.250000, time 55.799 samples/sec
+epoch 4, total_step 65320, total loss is 21.16 , inference loss is 11.50, weight deacy loss is 9.65, training accuracy is 0.187500, time 56.247 samples/sec
+epoch 4, total_step 65340, total loss is 19.99 , inference loss is 10.34, weight deacy loss is 9.65, training accuracy is 0.281250, time 57.526 samples/sec
+epoch 4, total_step 65360, total loss is 20.72 , inference loss is 11.07, weight deacy loss is 9.65, training accuracy is 0.343750, time 55.687 samples/sec
+epoch 4, total_step 65380, total loss is 20.21 , inference loss is 10.56, weight deacy loss is 9.65, training accuracy is 0.312500, time 56.525 samples/sec
+epoch 4, total_step 65400, total loss is 19.46 , inference loss is 9.81, weight deacy loss is 9.65, training accuracy is 0.375000, time 56.452 samples/sec
+epoch 4, total_step 65420, total loss is 22.88 , inference loss is 13.23, weight deacy loss is 9.65, training accuracy is 0.250000, time 56.195 samples/sec
+epoch 4, total_step 65440, total loss is 23.24 , inference loss is 13.59, weight deacy loss is 9.65, training accuracy is 0.125000, time 56.482 samples/sec
+epoch 4, total_step 65460, total loss is 18.98 , inference loss is 9.33, weight deacy loss is 9.65, training accuracy is 0.281250, time 56.094 samples/sec
+epoch 4, total_step 65480, total loss is 18.32 , inference loss is 8.67, weight deacy loss is 9.65, training accuracy is 0.375000, time 55.666 samples/sec
+epoch 4, total_step 65500, total loss is 19.59 , inference loss is 9.95, weight deacy loss is 9.64, training accuracy is 0.312500, time 55.713 samples/sec
+epoch 4, total_step 65520, total loss is 23.53 , inference loss is 13.88, weight deacy loss is 9.64, training accuracy is 0.312500, time 56.392 samples/sec
+epoch 4, total_step 65540, total loss is 23.12 , inference loss is 13.48, weight deacy loss is 9.64, training accuracy is 0.156250, time 55.821 samples/sec
+epoch 4, total_step 65560, total loss is 22.94 , inference loss is 13.30, weight deacy loss is 9.64, training accuracy is 0.125000, time 57.005 samples/sec
+epoch 4, total_step 65580, total loss is 22.29 , inference loss is 12.65, weight deacy loss is 9.64, training accuracy is 0.312500, time 55.531 samples/sec
+epoch 4, total_step 65600, total loss is 17.97 , inference loss is 8.33, weight deacy loss is 9.64, training accuracy is 0.406250, time 56.217 samples/sec
+epoch 4, total_step 65620, total loss is 20.45 , inference loss is 10.81, weight deacy loss is 9.64, training accuracy is 0.281250, time 56.320 samples/sec
+epoch 4, total_step 65640, total loss is 19.82 , inference loss is 10.19, weight deacy loss is 9.64, training accuracy is 0.187500, time 56.394 samples/sec
+epoch 4, total_step 65660, total loss is 22.59 , inference loss is 12.96, weight deacy loss is 9.64, training accuracy is 0.187500, time 56.478 samples/sec
+epoch 4, total_step 65680, total loss is 23.45 , inference loss is 13.82, weight deacy loss is 9.63, training accuracy is 0.250000, time 56.317 samples/sec
+epoch 4, total_step 65700, total loss is 23.18 , inference loss is 13.54, weight deacy loss is 9.63, training accuracy is 0.218750, time 55.896 samples/sec
+epoch 4, total_step 65720, total loss is 20.23 , inference loss is 10.60, weight deacy loss is 9.63, training accuracy is 0.281250, time 54.546 samples/sec
+epoch 4, total_step 65740, total loss is 21.86 , inference loss is 12.23, weight deacy loss is 9.63, training accuracy is 0.187500, time 56.070 samples/sec
+epoch 4, total_step 65760, total loss is 19.59 , inference loss is 9.96, weight deacy loss is 9.63, training accuracy is 0.312500, time 55.603 samples/sec
+epoch 4, total_step 65780, total loss is 18.39 , inference loss is 8.76, weight deacy loss is 9.63, training accuracy is 0.312500, time 54.829 samples/sec
+epoch 4, total_step 65800, total loss is 21.20 , inference loss is 11.57, weight deacy loss is 9.63, training accuracy is 0.281250, time 56.465 samples/sec
+epoch 4, total_step 65820, total loss is 20.09 , inference loss is 10.46, weight deacy loss is 9.63, training accuracy is 0.343750, time 55.336 samples/sec
+epoch 4, total_step 65840, total loss is 20.59 , inference loss is 10.96, weight deacy loss is 9.63, training accuracy is 0.312500, time 56.571 samples/sec
+epoch 4, total_step 65860, total loss is 19.88 , inference loss is 10.25, weight deacy loss is 9.62, training accuracy is 0.187500, time 55.914 samples/sec
+epoch 4, total_step 65880, total loss is 20.41 , inference loss is 10.79, weight deacy loss is 9.62, training accuracy is 0.281250, time 54.813 samples/sec
+epoch 4, total_step 65900, total loss is 26.62 , inference loss is 17.00, weight deacy loss is 9.62, training accuracy is 0.156250, time 56.578 samples/sec
+epoch 4, total_step 65920, total loss is 23.97 , inference loss is 14.34, weight deacy loss is 9.62, training accuracy is 0.187500, time 57.215 samples/sec
+epoch 4, total_step 65940, total loss is 22.20 , inference loss is 12.58, weight deacy loss is 9.62, training accuracy is 0.187500, time 56.445 samples/sec
+epoch 4, total_step 65960, total loss is 20.60 , inference loss is 10.98, weight deacy loss is 9.62, training accuracy is 0.312500, time 55.444 samples/sec
+epoch 4, total_step 65980, total loss is 23.18 , inference loss is 13.56, weight deacy loss is 9.62, training accuracy is 0.125000, time 56.289 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.45404599999996
+best_threshold_index 114 0.9851851851851852
+best_threshold_index 114 0.9846296296296296
+best_threshold_index 114 0.9859259259259259
+best_threshold_index 114 0.9855555555555555
+best_threshold_index 114 0.9861111111111112
+best_threshold_index 114 0.9844444444444445
+best_threshold_index 114 0.9846296296296296
+best_threshold_index 114 0.9842592592592593
+best_threshold_index 114 0.9835185185185186
+best_threshold_index 114 0.9840740740740741
+[lfw][66000]XNorm: 22.270934
+[lfw][66000]Accuracy-Flip: 0.98483+-0.00717
+testing verification..
+(14000, 512)
+infer time 86.50112799999987
+best_threshold_index 116 0.9855555555555555
+best_threshold_index 116 0.984920634920635
+best_threshold_index 116 0.9844444444444445
+best_threshold_index 116 0.984920634920635
+best_threshold_index 116 0.984920634920635
+best_threshold_index 116 0.9839682539682539
+best_threshold_index 116 0.9857142857142858
+best_threshold_index 116 0.9850793650793651
+best_threshold_index 116 0.9844444444444445
+best_threshold_index 116 0.986031746031746
+[cfp_ff][66000]XNorm: 22.097975
+[cfp_ff][66000]Accuracy-Flip: 0.98500+-0.00539
+testing verification..
+(14000, 512)
+infer time 86.33369199999993
+best_threshold_index 139 0.8995238095238095
+best_threshold_index 139 0.8987301587301587
+best_threshold_index 139 0.8984126984126984
+best_threshold_index 134 0.900952380952381
+best_threshold_index 141 0.9012698412698412
+best_threshold_index 139 0.8968253968253969
+best_threshold_index 134 0.8977777777777778
+best_threshold_index 138 0.8990476190476191
+best_threshold_index 134 0.8998412698412699
+best_threshold_index 141 0.9003174603174603
+[cfp_fp][66000]XNorm: 22.116254
+[cfp_fp][66000]Accuracy-Flip: 0.89571+-0.01325
+testing verification..
+(12000, 512)
+infer time 73.86018900000005
+best_threshold_index 143 0.8737037037037036
+best_threshold_index 144 0.8709259259259259
+best_threshold_index 143 0.8737037037037036
+best_threshold_index 144 0.8744444444444445
+best_threshold_index 143 0.8777777777777778
+best_threshold_index 146 0.8731481481481481
+best_threshold_index 135 0.8725925925925926
+best_threshold_index 143 0.8798148148148148
+best_threshold_index 146 0.8787037037037037
+best_threshold_index 138 0.8761111111111111
+[agedb_30][66000]XNorm: 22.582456
+[agedb_30][66000]Accuracy-Flip: 0.86783+-0.02471
+test accuracy is: 0.9848333333333334
+epoch 4, total_step 66000, total loss is 22.24 , inference loss is 12.62, weight deacy loss is 9.62, training accuracy is 0.125000, time 54.409 samples/sec
+epoch 4, total_step 66020, total loss is 20.71 , inference loss is 11.10, weight deacy loss is 9.62, training accuracy is 0.312500, time 55.606 samples/sec
+epoch 4, total_step 66040, total loss is 20.83 , inference loss is 11.21, weight deacy loss is 9.62, training accuracy is 0.250000, time 56.555 samples/sec
+epoch 4, total_step 66060, total loss is 17.66 , inference loss is 8.05, weight deacy loss is 9.61, training accuracy is 0.312500, time 56.353 samples/sec
+epoch 4, total_step 66080, total loss is 22.83 , inference loss is 13.22, weight deacy loss is 9.61, training accuracy is 0.250000, time 55.659 samples/sec
+epoch 4, total_step 66100, total loss is 22.40 , inference loss is 12.79, weight deacy loss is 9.61, training accuracy is 0.375000, time 56.732 samples/sec
+epoch 4, total_step 66120, total loss is 18.69 , inference loss is 9.08, weight deacy loss is 9.61, training accuracy is 0.375000, time 56.392 samples/sec
+epoch 4, total_step 66140, total loss is 21.28 , inference loss is 11.67, weight deacy loss is 9.61, training accuracy is 0.187500, time 56.392 samples/sec
+epoch 4, total_step 66160, total loss is 18.01 , inference loss is 8.40, weight deacy loss is 9.61, training accuracy is 0.312500, time 56.565 samples/sec
+epoch 4, total_step 66180, total loss is 26.93 , inference loss is 17.32, weight deacy loss is 9.61, training accuracy is 0.125000, time 56.517 samples/sec
+epoch 4, total_step 66200, total loss is 20.89 , inference loss is 11.28, weight deacy loss is 9.61, training accuracy is 0.156250, time 55.937 samples/sec
+epoch 4, total_step 66220, total loss is 18.92 , inference loss is 9.31, weight deacy loss is 9.61, training accuracy is 0.343750, time 55.279 samples/sec
+epoch 4, total_step 66240, total loss is 24.31 , inference loss is 14.71, weight deacy loss is 9.60, training accuracy is 0.187500, time 56.017 samples/sec
+epoch 4, total_step 66260, total loss is 23.29 , inference loss is 13.69, weight deacy loss is 9.60, training accuracy is 0.281250, time 56.289 samples/sec
+epoch 4, total_step 66280, total loss is 22.04 , inference loss is 12.44, weight deacy loss is 9.60, training accuracy is 0.250000, time 56.389 samples/sec
+epoch 4, total_step 66300, total loss is 21.68 , inference loss is 12.08, weight deacy loss is 9.60, training accuracy is 0.187500, time 54.272 samples/sec
+epoch 4, total_step 66320, total loss is 23.40 , inference loss is 13.80, weight deacy loss is 9.60, training accuracy is 0.281250, time 57.671 samples/sec
+epoch 4, total_step 66340, total loss is 26.55 , inference loss is 16.96, weight deacy loss is 9.60, training accuracy is 0.218750, time 56.476 samples/sec
+epoch 4, total_step 66360, total loss is 20.11 , inference loss is 10.51, weight deacy loss is 9.60, training accuracy is 0.281250, time 56.758 samples/sec
+epoch 4, total_step 66380, total loss is 21.22 , inference loss is 11.63, weight deacy loss is 9.60, training accuracy is 0.250000, time 55.454 samples/sec
+epoch 4, total_step 66400, total loss is 21.55 , inference loss is 11.95, weight deacy loss is 9.60, training accuracy is 0.093750, time 55.702 samples/sec
+epoch 4, total_step 66420, total loss is 22.52 , inference loss is 12.93, weight deacy loss is 9.59, training accuracy is 0.218750, time 56.010 samples/sec
+epoch 4, total_step 66440, total loss is 18.19 , inference loss is 8.60, weight deacy loss is 9.59, training accuracy is 0.312500, time 57.425 samples/sec
+epoch 4, total_step 66460, total loss is 20.48 , inference loss is 10.88, weight deacy loss is 9.59, training accuracy is 0.250000, time 55.983 samples/sec
+epoch 4, total_step 66480, total loss is 20.35 , inference loss is 10.76, weight deacy loss is 9.59, training accuracy is 0.218750, time 56.240 samples/sec
+epoch 4, total_step 66500, total loss is 23.62 , inference loss is 14.03, weight deacy loss is 9.59, training accuracy is 0.125000, time 56.428 samples/sec
+epoch 4, total_step 66520, total loss is 19.21 , inference loss is 9.62, weight deacy loss is 9.59, training accuracy is 0.343750, time 56.035 samples/sec
+epoch 4, total_step 66540, total loss is 22.55 , inference loss is 12.96, weight deacy loss is 9.59, training accuracy is 0.375000, time 56.151 samples/sec
+epoch 4, total_step 66560, total loss is 24.24 , inference loss is 14.66, weight deacy loss is 9.59, training accuracy is 0.250000, time 56.654 samples/sec
+epoch 4, total_step 66580, total loss is 20.55 , inference loss is 10.96, weight deacy loss is 9.59, training accuracy is 0.281250, time 55.153 samples/sec
+epoch 4, total_step 66600, total loss is 24.23 , inference loss is 14.64, weight deacy loss is 9.58, training accuracy is 0.281250, time 56.470 samples/sec
+epoch 4, total_step 66620, total loss is 22.18 , inference loss is 12.59, weight deacy loss is 9.58, training accuracy is 0.250000, time 54.984 samples/sec
+epoch 4, total_step 66640, total loss is 19.73 , inference loss is 10.15, weight deacy loss is 9.58, training accuracy is 0.406250, time 55.967 samples/sec
+epoch 4, total_step 66660, total loss is 23.57 , inference loss is 13.98, weight deacy loss is 9.58, training accuracy is 0.250000, time 56.459 samples/sec
+epoch 4, total_step 66680, total loss is 21.84 , inference loss is 12.26, weight deacy loss is 9.58, training accuracy is 0.281250, time 55.595 samples/sec
+epoch 4, total_step 66700, total loss is 23.69 , inference loss is 14.12, weight deacy loss is 9.58, training accuracy is 0.218750, time 56.248 samples/sec
+epoch 4, total_step 66720, total loss is 18.38 , inference loss is 8.80, weight deacy loss is 9.58, training accuracy is 0.312500, time 55.997 samples/sec
+epoch 4, total_step 66740, total loss is 23.08 , inference loss is 13.51, weight deacy loss is 9.58, training accuracy is 0.281250, time 56.025 samples/sec
+epoch 4, total_step 66760, total loss is 20.72 , inference loss is 11.14, weight deacy loss is 9.58, training accuracy is 0.468750, time 54.513 samples/sec
+epoch 4, total_step 66780, total loss is 20.29 , inference loss is 10.72, weight deacy loss is 9.57, training accuracy is 0.281250, time 56.316 samples/sec
+epoch 4, total_step 66800, total loss is 22.91 , inference loss is 13.34, weight deacy loss is 9.57, training accuracy is 0.156250, time 56.408 samples/sec
+epoch 4, total_step 66820, total loss is 24.29 , inference loss is 14.72, weight deacy loss is 9.57, training accuracy is 0.250000, time 56.283 samples/sec
+epoch 4, total_step 66840, total loss is 20.35 , inference loss is 10.78, weight deacy loss is 9.57, training accuracy is 0.218750, time 56.623 samples/sec
+epoch 4, total_step 66860, total loss is 22.03 , inference loss is 12.46, weight deacy loss is 9.57, training accuracy is 0.187500, time 56.582 samples/sec
+epoch 4, total_step 66880, total loss is 23.08 , inference loss is 13.51, weight deacy loss is 9.57, training accuracy is 0.218750, time 56.834 samples/sec
+epoch 4, total_step 66900, total loss is 19.28 , inference loss is 9.71, weight deacy loss is 9.57, training accuracy is 0.281250, time 56.110 samples/sec
+epoch 4, total_step 66920, total loss is 25.84 , inference loss is 16.27, weight deacy loss is 9.57, training accuracy is 0.156250, time 55.752 samples/sec
+epoch 4, total_step 66940, total loss is 20.60 , inference loss is 11.04, weight deacy loss is 9.57, training accuracy is 0.187500, time 54.816 samples/sec
+epoch 4, total_step 66960, total loss is 21.96 , inference loss is 12.39, weight deacy loss is 9.56, training accuracy is 0.218750, time 56.164 samples/sec
+epoch 4, total_step 66980, total loss is 18.69 , inference loss is 9.12, weight deacy loss is 9.56, training accuracy is 0.375000, time 56.194 samples/sec
+epoch 4, total_step 67000, total loss is 23.61 , inference loss is 14.04, weight deacy loss is 9.56, training accuracy is 0.250000, time 56.197 samples/sec
+epoch 4, total_step 67020, total loss is 20.54 , inference loss is 10.97, weight deacy loss is 9.56, training accuracy is 0.375000, time 54.797 samples/sec
+epoch 4, total_step 67040, total loss is 21.07 , inference loss is 11.51, weight deacy loss is 9.56, training accuracy is 0.187500, time 54.525 samples/sec
+epoch 4, total_step 67060, total loss is 18.66 , inference loss is 9.11, weight deacy loss is 9.56, training accuracy is 0.437500, time 56.251 samples/sec
+epoch 4, total_step 67080, total loss is 19.16 , inference loss is 9.60, weight deacy loss is 9.56, training accuracy is 0.250000, time 57.081 samples/sec
+epoch 4, total_step 67100, total loss is 18.91 , inference loss is 9.35, weight deacy loss is 9.56, training accuracy is 0.312500, time 54.256 samples/sec
+epoch 4, total_step 67120, total loss is 24.30 , inference loss is 14.75, weight deacy loss is 9.56, training accuracy is 0.187500, time 56.338 samples/sec
+epoch 4, total_step 67140, total loss is 20.54 , inference loss is 10.98, weight deacy loss is 9.56, training accuracy is 0.281250, time 55.093 samples/sec
+epoch 4, total_step 67160, total loss is 22.39 , inference loss is 12.84, weight deacy loss is 9.55, training accuracy is 0.218750, time 56.088 samples/sec
+epoch 4, total_step 67180, total loss is 21.66 , inference loss is 12.11, weight deacy loss is 9.55, training accuracy is 0.250000, time 55.691 samples/sec
+epoch 4, total_step 67200, total loss is 19.33 , inference loss is 9.77, weight deacy loss is 9.55, training accuracy is 0.250000, time 54.413 samples/sec
+epoch 4, total_step 67220, total loss is 20.34 , inference loss is 10.79, weight deacy loss is 9.55, training accuracy is 0.312500, time 55.766 samples/sec
+epoch 4, total_step 67240, total loss is 22.75 , inference loss is 13.20, weight deacy loss is 9.55, training accuracy is 0.281250, time 52.856 samples/sec
+epoch 4, total_step 67260, total loss is 21.85 , inference loss is 12.30, weight deacy loss is 9.55, training accuracy is 0.343750, time 56.949 samples/sec
+epoch 4, total_step 67280, total loss is 21.33 , inference loss is 11.78, weight deacy loss is 9.55, training accuracy is 0.156250, time 56.514 samples/sec
+epoch 4, total_step 67300, total loss is 21.16 , inference loss is 11.62, weight deacy loss is 9.55, training accuracy is 0.312500, time 56.958 samples/sec
+epoch 4, total_step 67320, total loss is 20.62 , inference loss is 11.08, weight deacy loss is 9.55, training accuracy is 0.406250, time 56.405 samples/sec
+epoch 4, total_step 67340, total loss is 19.18 , inference loss is 9.63, weight deacy loss is 9.54, training accuracy is 0.312500, time 55.512 samples/sec
+epoch 4, total_step 67360, total loss is 20.18 , inference loss is 10.64, weight deacy loss is 9.54, training accuracy is 0.250000, time 56.284 samples/sec
+epoch 4, total_step 67380, total loss is 20.06 , inference loss is 10.52, weight deacy loss is 9.54, training accuracy is 0.281250, time 56.227 samples/sec
+epoch 4, total_step 67400, total loss is 22.32 , inference loss is 12.78, weight deacy loss is 9.54, training accuracy is 0.250000, time 56.049 samples/sec
+epoch 4, total_step 67420, total loss is 19.39 , inference loss is 9.85, weight deacy loss is 9.54, training accuracy is 0.281250, time 55.949 samples/sec
+epoch 4, total_step 67440, total loss is 20.34 , inference loss is 10.80, weight deacy loss is 9.54, training accuracy is 0.343750, time 56.409 samples/sec
+epoch 4, total_step 67460, total loss is 19.09 , inference loss is 9.55, weight deacy loss is 9.54, training accuracy is 0.218750, time 55.200 samples/sec
+epoch 4, total_step 67480, total loss is 23.15 , inference loss is 13.61, weight deacy loss is 9.54, training accuracy is 0.218750, time 56.573 samples/sec
+epoch 4, total_step 67500, total loss is 18.37 , inference loss is 8.83, weight deacy loss is 9.54, training accuracy is 0.281250, time 54.915 samples/sec
+epoch 4, total_step 67520, total loss is 18.59 , inference loss is 9.06, weight deacy loss is 9.53, training accuracy is 0.343750, time 56.309 samples/sec
+epoch 4, total_step 67540, total loss is 20.80 , inference loss is 11.27, weight deacy loss is 9.53, training accuracy is 0.250000, time 56.519 samples/sec
+epoch 4, total_step 67560, total loss is 20.44 , inference loss is 10.91, weight deacy loss is 9.53, training accuracy is 0.250000, time 56.083 samples/sec
+epoch 4, total_step 67580, total loss is 23.86 , inference loss is 14.33, weight deacy loss is 9.53, training accuracy is 0.250000, time 56.102 samples/sec
+epoch 4, total_step 67600, total loss is 20.85 , inference loss is 11.32, weight deacy loss is 9.53, training accuracy is 0.312500, time 56.450 samples/sec
+epoch 4, total_step 67620, total loss is 19.16 , inference loss is 9.63, weight deacy loss is 9.53, training accuracy is 0.343750, time 56.093 samples/sec
+epoch 4, total_step 67640, total loss is 22.43 , inference loss is 12.90, weight deacy loss is 9.53, training accuracy is 0.281250, time 56.270 samples/sec
+epoch 4, total_step 67660, total loss is 22.89 , inference loss is 13.37, weight deacy loss is 9.53, training accuracy is 0.125000, time 55.842 samples/sec
+epoch 4, total_step 67680, total loss is 19.86 , inference loss is 10.33, weight deacy loss is 9.53, training accuracy is 0.312500, time 56.507 samples/sec
+epoch 4, total_step 67700, total loss is 23.28 , inference loss is 13.76, weight deacy loss is 9.52, training accuracy is 0.187500, time 56.157 samples/sec
+epoch 4, total_step 67720, total loss is 20.46 , inference loss is 10.94, weight deacy loss is 9.52, training accuracy is 0.406250, time 56.445 samples/sec
+epoch 4, total_step 67740, total loss is 27.20 , inference loss is 17.68, weight deacy loss is 9.52, training accuracy is 0.156250, time 57.154 samples/sec
+epoch 4, total_step 67760, total loss is 19.99 , inference loss is 10.47, weight deacy loss is 9.52, training accuracy is 0.281250, time 56.042 samples/sec
+epoch 4, total_step 67780, total loss is 24.77 , inference loss is 15.25, weight deacy loss is 9.52, training accuracy is 0.218750, time 55.854 samples/sec
+epoch 4, total_step 67800, total loss is 22.56 , inference loss is 13.04, weight deacy loss is 9.52, training accuracy is 0.250000, time 56.213 samples/sec
+epoch 4, total_step 67820, total loss is 21.73 , inference loss is 12.21, weight deacy loss is 9.52, training accuracy is 0.187500, time 56.441 samples/sec
+epoch 4, total_step 67840, total loss is 19.48 , inference loss is 9.96, weight deacy loss is 9.52, training accuracy is 0.218750, time 57.001 samples/sec
+epoch 4, total_step 67860, total loss is 15.91 , inference loss is 6.40, weight deacy loss is 9.52, training accuracy is 0.281250, time 56.389 samples/sec
+epoch 4, total_step 67880, total loss is 26.28 , inference loss is 16.76, weight deacy loss is 9.51, training accuracy is 0.187500, time 56.677 samples/sec
+epoch 4, total_step 67900, total loss is 17.82 , inference loss is 8.31, weight deacy loss is 9.51, training accuracy is 0.343750, time 55.945 samples/sec
+epoch 4, total_step 67920, total loss is 22.48 , inference loss is 12.97, weight deacy loss is 9.51, training accuracy is 0.312500, time 57.241 samples/sec
+epoch 4, total_step 67940, total loss is 22.71 , inference loss is 13.20, weight deacy loss is 9.51, training accuracy is 0.156250, time 57.612 samples/sec
+epoch 4, total_step 67960, total loss is 22.70 , inference loss is 13.19, weight deacy loss is 9.51, training accuracy is 0.281250, time 56.933 samples/sec
+epoch 4, total_step 67980, total loss is 25.91 , inference loss is 16.40, weight deacy loss is 9.51, training accuracy is 0.156250, time 56.626 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.64168300000004
+best_threshold_index 108 0.9814814814814815
+best_threshold_index 110 0.9803703703703703
+best_threshold_index 108 0.9814814814814815
+best_threshold_index 110 0.9822222222222222
+best_threshold_index 110 0.9812962962962963
+best_threshold_index 108 0.9798148148148148
+best_threshold_index 108 0.9812962962962963
+best_threshold_index 110 0.9798148148148148
+best_threshold_index 110 0.9801851851851852
+best_threshold_index 110 0.9803703703703703
+[lfw][68000]XNorm: 23.756261
+[lfw][68000]Accuracy-Flip: 0.98017+-0.00724
+testing verification..
+(14000, 512)
+infer time 86.90441200000001
+best_threshold_index 116 0.983015873015873
+best_threshold_index 113 0.981904761904762
+best_threshold_index 116 0.9825396825396825
+best_threshold_index 116 0.9820634920634921
+best_threshold_index 113 0.9825396825396825
+best_threshold_index 108 0.9815873015873016
+best_threshold_index 116 0.9822222222222222
+best_threshold_index 116 0.9815873015873016
+best_threshold_index 113 0.9820634920634921
+best_threshold_index 116 0.9822222222222222
+[cfp_ff][68000]XNorm: 22.728200
+[cfp_ff][68000]Accuracy-Flip: 0.98086+-0.00333
+testing verification..
+(14000, 512)
+infer time 86.91835500000003
+best_threshold_index 128 0.893015873015873
+best_threshold_index 128 0.8915873015873016
+best_threshold_index 130 0.8903174603174603
+best_threshold_index 136 0.8926984126984127
+best_threshold_index 128 0.8961904761904762
+best_threshold_index 136 0.8888888888888888
+best_threshold_index 128 0.8890476190476191
+best_threshold_index 128 0.8893650793650794
+best_threshold_index 131 0.8917460317460317
+best_threshold_index 136 0.8923809523809524
+[cfp_fp][68000]XNorm: 22.306235
+[cfp_fp][68000]Accuracy-Flip: 0.88829+-0.01910
+testing verification..
+(12000, 512)
+infer time 74.61222700000005
+best_threshold_index 133 0.8877777777777778
+best_threshold_index 133 0.8851851851851852
+best_threshold_index 133 0.887962962962963
+best_threshold_index 133 0.8848148148148148
+best_threshold_index 132 0.8903703703703704
+best_threshold_index 133 0.8838888888888888
+best_threshold_index 133 0.8870370370370371
+best_threshold_index 133 0.8909259259259259
+best_threshold_index 133 0.8894444444444445
+best_threshold_index 133 0.8892592592592593
+[agedb_30][68000]XNorm: 23.833160
+[agedb_30][68000]Accuracy-Flip: 0.88717+-0.02129
+test accuracy is: 0.9801666666666666
+epoch 4, total_step 68000, total loss is 22.73 , inference loss is 13.22, weight deacy loss is 9.51, training accuracy is 0.281250, time 53.275 samples/sec
+epoch 4, total_step 68020, total loss is 16.20 , inference loss is 6.69, weight deacy loss is 9.51, training accuracy is 0.343750, time 56.637 samples/sec
+epoch 4, total_step 68040, total loss is 21.93 , inference loss is 12.43, weight deacy loss is 9.51, training accuracy is 0.312500, time 54.101 samples/sec
+epoch 4, total_step 68060, total loss is 25.16 , inference loss is 15.65, weight deacy loss is 9.51, training accuracy is 0.250000, time 54.950 samples/sec
+epoch 4, total_step 68080, total loss is 23.23 , inference loss is 13.72, weight deacy loss is 9.50, training accuracy is 0.187500, time 55.836 samples/sec
+epoch 4, total_step 68100, total loss is 22.07 , inference loss is 12.57, weight deacy loss is 9.50, training accuracy is 0.250000, time 56.109 samples/sec
+epoch 4, total_step 68120, total loss is 16.98 , inference loss is 7.48, weight deacy loss is 9.50, training accuracy is 0.312500, time 56.058 samples/sec
+epoch 4, total_step 68140, total loss is 25.14 , inference loss is 15.64, weight deacy loss is 9.50, training accuracy is 0.218750, time 56.339 samples/sec
+epoch 4, total_step 68160, total loss is 20.13 , inference loss is 10.63, weight deacy loss is 9.50, training accuracy is 0.156250, time 55.813 samples/sec
+epoch 4, total_step 68180, total loss is 17.93 , inference loss is 8.43, weight deacy loss is 9.50, training accuracy is 0.437500, time 56.180 samples/sec
+epoch 4, total_step 68200, total loss is 21.45 , inference loss is 11.95, weight deacy loss is 9.50, training accuracy is 0.250000, time 57.139 samples/sec
+epoch 4, total_step 68220, total loss is 21.03 , inference loss is 11.53, weight deacy loss is 9.50, training accuracy is 0.187500, time 56.306 samples/sec
+epoch 4, total_step 68240, total loss is 23.61 , inference loss is 14.11, weight deacy loss is 9.50, training accuracy is 0.218750, time 56.012 samples/sec
+epoch 4, total_step 68260, total loss is 22.86 , inference loss is 13.37, weight deacy loss is 9.49, training accuracy is 0.312500, time 55.518 samples/sec
+epoch 4, total_step 68280, total loss is 23.47 , inference loss is 13.97, weight deacy loss is 9.49, training accuracy is 0.062500, time 56.432 samples/sec
+epoch 4, total_step 68300, total loss is 19.98 , inference loss is 10.49, weight deacy loss is 9.49, training accuracy is 0.187500, time 55.959 samples/sec
+epoch 4, total_step 68320, total loss is 25.92 , inference loss is 16.43, weight deacy loss is 9.49, training accuracy is 0.281250, time 56.172 samples/sec
+epoch 4, total_step 68340, total loss is 22.94 , inference loss is 13.45, weight deacy loss is 9.49, training accuracy is 0.250000, time 55.784 samples/sec
+epoch 4, total_step 68360, total loss is 22.63 , inference loss is 13.14, weight deacy loss is 9.49, training accuracy is 0.156250, time 56.397 samples/sec
+epoch 4, total_step 68380, total loss is 23.59 , inference loss is 14.11, weight deacy loss is 9.49, training accuracy is 0.187500, time 56.424 samples/sec
+epoch 4, total_step 68400, total loss is 24.65 , inference loss is 15.17, weight deacy loss is 9.49, training accuracy is 0.093750, time 55.257 samples/sec
+epoch 4, total_step 68420, total loss is 22.05 , inference loss is 12.56, weight deacy loss is 9.49, training accuracy is 0.156250, time 56.388 samples/sec
+epoch 4, total_step 68440, total loss is 20.64 , inference loss is 11.15, weight deacy loss is 9.48, training accuracy is 0.156250, time 55.635 samples/sec
+epoch 4, total_step 68460, total loss is 18.23 , inference loss is 8.74, weight deacy loss is 9.48, training accuracy is 0.343750, time 54.317 samples/sec
+epoch 4, total_step 68480, total loss is 20.99 , inference loss is 11.51, weight deacy loss is 9.48, training accuracy is 0.312500, time 54.783 samples/sec
+epoch 4, total_step 68500, total loss is 17.76 , inference loss is 8.28, weight deacy loss is 9.48, training accuracy is 0.312500, time 56.028 samples/sec
+epoch 4, total_step 68520, total loss is 20.58 , inference loss is 11.10, weight deacy loss is 9.48, training accuracy is 0.250000, time 56.265 samples/sec
+epoch 4, total_step 68540, total loss is 21.98 , inference loss is 12.50, weight deacy loss is 9.48, training accuracy is 0.125000, time 56.239 samples/sec
+epoch 4, total_step 68560, total loss is 22.94 , inference loss is 13.46, weight deacy loss is 9.48, training accuracy is 0.250000, time 56.582 samples/sec
+epoch 4, total_step 68580, total loss is 19.54 , inference loss is 10.06, weight deacy loss is 9.48, training accuracy is 0.218750, time 55.635 samples/sec
+epoch 4, total_step 68600, total loss is 27.44 , inference loss is 17.96, weight deacy loss is 9.48, training accuracy is 0.125000, time 56.442 samples/sec
+epoch 4, total_step 68620, total loss is 19.65 , inference loss is 10.18, weight deacy loss is 9.48, training accuracy is 0.156250, time 55.924 samples/sec
+epoch 4, total_step 68640, total loss is 21.03 , inference loss is 11.56, weight deacy loss is 9.47, training accuracy is 0.312500, time 56.661 samples/sec
+epoch 4, total_step 68660, total loss is 18.04 , inference loss is 8.56, weight deacy loss is 9.47, training accuracy is 0.312500, time 56.290 samples/sec
+epoch 4, total_step 68680, total loss is 18.08 , inference loss is 8.61, weight deacy loss is 9.47, training accuracy is 0.281250, time 56.528 samples/sec
+epoch 4, total_step 68700, total loss is 19.14 , inference loss is 9.67, weight deacy loss is 9.47, training accuracy is 0.281250, time 55.007 samples/sec
+epoch 4, total_step 68720, total loss is 21.19 , inference loss is 11.72, weight deacy loss is 9.47, training accuracy is 0.312500, time 55.662 samples/sec
+epoch 4, total_step 68740, total loss is 24.91 , inference loss is 15.44, weight deacy loss is 9.47, training accuracy is 0.187500, time 55.726 samples/sec
+epoch 4, total_step 68760, total loss is 17.97 , inference loss is 8.50, weight deacy loss is 9.47, training accuracy is 0.343750, time 56.262 samples/sec
+epoch 4, total_step 68780, total loss is 18.93 , inference loss is 9.47, weight deacy loss is 9.47, training accuracy is 0.187500, time 54.362 samples/sec
+epoch 4, total_step 68800, total loss is 22.54 , inference loss is 13.08, weight deacy loss is 9.47, training accuracy is 0.187500, time 58.332 samples/sec
+epoch 4, total_step 68820, total loss is 21.22 , inference loss is 11.75, weight deacy loss is 9.46, training accuracy is 0.250000, time 56.501 samples/sec
+epoch 4, total_step 68840, total loss is 23.55 , inference loss is 14.09, weight deacy loss is 9.46, training accuracy is 0.250000, time 55.497 samples/sec
+epoch 4, total_step 68860, total loss is 24.64 , inference loss is 15.18, weight deacy loss is 9.46, training accuracy is 0.156250, time 56.670 samples/sec
+epoch 4, total_step 68880, total loss is 21.85 , inference loss is 12.39, weight deacy loss is 9.46, training accuracy is 0.250000, time 55.406 samples/sec
+epoch 4, total_step 68900, total loss is 23.61 , inference loss is 14.15, weight deacy loss is 9.46, training accuracy is 0.187500, time 56.064 samples/sec
+epoch 4, total_step 68920, total loss is 17.37 , inference loss is 7.91, weight deacy loss is 9.46, training accuracy is 0.531250, time 55.975 samples/sec
+epoch 4, total_step 68940, total loss is 20.91 , inference loss is 11.45, weight deacy loss is 9.46, training accuracy is 0.281250, time 56.504 samples/sec
+epoch 4, total_step 68960, total loss is 19.11 , inference loss is 9.65, weight deacy loss is 9.46, training accuracy is 0.312500, time 55.637 samples/sec
+epoch 4, total_step 68980, total loss is 21.27 , inference loss is 11.81, weight deacy loss is 9.46, training accuracy is 0.312500, time 56.458 samples/sec
+epoch 4, total_step 69000, total loss is 20.09 , inference loss is 10.64, weight deacy loss is 9.45, training accuracy is 0.375000, time 55.543 samples/sec
+epoch 4, total_step 69020, total loss is 20.84 , inference loss is 11.39, weight deacy loss is 9.45, training accuracy is 0.343750, time 55.944 samples/sec
+epoch 4, total_step 69040, total loss is 17.86 , inference loss is 8.41, weight deacy loss is 9.45, training accuracy is 0.312500, time 56.448 samples/sec
+epoch 4, total_step 69060, total loss is 18.76 , inference loss is 9.31, weight deacy loss is 9.45, training accuracy is 0.375000, time 56.087 samples/sec
+epoch 4, total_step 69080, total loss is 17.97 , inference loss is 8.52, weight deacy loss is 9.45, training accuracy is 0.406250, time 56.125 samples/sec
+epoch 4, total_step 69100, total loss is 21.73 , inference loss is 12.28, weight deacy loss is 9.45, training accuracy is 0.343750, time 55.782 samples/sec
+epoch 4, total_step 69120, total loss is 18.42 , inference loss is 8.97, weight deacy loss is 9.45, training accuracy is 0.250000, time 56.309 samples/sec
+epoch 4, total_step 69140, total loss is 19.15 , inference loss is 9.71, weight deacy loss is 9.45, training accuracy is 0.343750, time 56.512 samples/sec
+epoch 4, total_step 69160, total loss is 20.69 , inference loss is 11.24, weight deacy loss is 9.45, training accuracy is 0.250000, time 56.151 samples/sec
+epoch 4, total_step 69180, total loss is 24.05 , inference loss is 14.60, weight deacy loss is 9.44, training accuracy is 0.281250, time 56.578 samples/sec
+epoch 4, total_step 69200, total loss is 25.22 , inference loss is 15.78, weight deacy loss is 9.44, training accuracy is 0.187500, time 55.694 samples/sec
+epoch 4, total_step 69220, total loss is 19.58 , inference loss is 10.14, weight deacy loss is 9.44, training accuracy is 0.187500, time 56.405 samples/sec
+epoch 4, total_step 69240, total loss is 22.08 , inference loss is 12.63, weight deacy loss is 9.44, training accuracy is 0.343750, time 56.511 samples/sec
+epoch 4, total_step 69260, total loss is 19.32 , inference loss is 9.88, weight deacy loss is 9.44, training accuracy is 0.312500, time 56.340 samples/sec
+epoch 4, total_step 69280, total loss is 24.13 , inference loss is 14.69, weight deacy loss is 9.44, training accuracy is 0.281250, time 56.289 samples/sec
+epoch 4, total_step 69300, total loss is 19.05 , inference loss is 9.61, weight deacy loss is 9.44, training accuracy is 0.250000, time 55.531 samples/sec
+epoch 4, total_step 69320, total loss is 18.83 , inference loss is 9.39, weight deacy loss is 9.44, training accuracy is 0.218750, time 56.400 samples/sec
+epoch 4, total_step 69340, total loss is 22.98 , inference loss is 13.54, weight deacy loss is 9.44, training accuracy is 0.312500, time 55.578 samples/sec
+epoch 4, total_step 69360, total loss is 18.15 , inference loss is 8.72, weight deacy loss is 9.43, training accuracy is 0.312500, time 55.936 samples/sec
+epoch 4, total_step 69380, total loss is 19.73 , inference loss is 10.29, weight deacy loss is 9.43, training accuracy is 0.375000, time 56.439 samples/sec
+epoch 4, total_step 69400, total loss is 23.45 , inference loss is 14.02, weight deacy loss is 9.43, training accuracy is 0.125000, time 55.923 samples/sec
+epoch 4, total_step 69420, total loss is 22.10 , inference loss is 12.67, weight deacy loss is 9.43, training accuracy is 0.281250, time 54.565 samples/sec
+epoch 4, total_step 69440, total loss is 18.57 , inference loss is 9.14, weight deacy loss is 9.43, training accuracy is 0.375000, time 55.985 samples/sec
+epoch 4, total_step 69460, total loss is 20.14 , inference loss is 10.71, weight deacy loss is 9.43, training accuracy is 0.375000, time 56.375 samples/sec
+epoch 4, total_step 69480, total loss is 20.22 , inference loss is 10.79, weight deacy loss is 9.43, training accuracy is 0.312500, time 55.911 samples/sec
+epoch 4, total_step 69500, total loss is 20.22 , inference loss is 10.79, weight deacy loss is 9.43, training accuracy is 0.281250, time 56.714 samples/sec
+epoch 4, total_step 69520, total loss is 22.53 , inference loss is 13.10, weight deacy loss is 9.43, training accuracy is 0.281250, time 57.216 samples/sec
+epoch 4, total_step 69540, total loss is 24.92 , inference loss is 15.49, weight deacy loss is 9.43, training accuracy is 0.125000, time 56.237 samples/sec
+epoch 4, total_step 69560, total loss is 19.68 , inference loss is 10.25, weight deacy loss is 9.42, training accuracy is 0.281250, time 56.622 samples/sec
+epoch 4, total_step 69580, total loss is 20.45 , inference loss is 11.03, weight deacy loss is 9.42, training accuracy is 0.250000, time 56.094 samples/sec
+epoch 4, total_step 69600, total loss is 22.76 , inference loss is 13.33, weight deacy loss is 9.42, training accuracy is 0.125000, time 56.095 samples/sec
+epoch 4, total_step 69620, total loss is 25.79 , inference loss is 16.37, weight deacy loss is 9.42, training accuracy is 0.187500, time 56.190 samples/sec
+epoch 4, total_step 69640, total loss is 20.80 , inference loss is 11.38, weight deacy loss is 9.42, training accuracy is 0.218750, time 56.226 samples/sec
+epoch 4, total_step 69660, total loss is 19.60 , inference loss is 10.18, weight deacy loss is 9.42, training accuracy is 0.281250, time 56.672 samples/sec
+epoch 4, total_step 69680, total loss is 20.99 , inference loss is 11.57, weight deacy loss is 9.42, training accuracy is 0.187500, time 56.528 samples/sec
+epoch 4, total_step 69700, total loss is 20.72 , inference loss is 11.31, weight deacy loss is 9.42, training accuracy is 0.343750, time 56.038 samples/sec
+epoch 4, total_step 69720, total loss is 19.74 , inference loss is 10.32, weight deacy loss is 9.42, training accuracy is 0.218750, time 56.159 samples/sec
+epoch 4, total_step 69740, total loss is 18.32 , inference loss is 8.91, weight deacy loss is 9.41, training accuracy is 0.250000, time 56.796 samples/sec
+epoch 4, total_step 69760, total loss is 25.98 , inference loss is 16.57, weight deacy loss is 9.41, training accuracy is 0.156250, time 56.304 samples/sec
+epoch 4, total_step 69780, total loss is 22.16 , inference loss is 12.74, weight deacy loss is 9.41, training accuracy is 0.187500, time 56.542 samples/sec
+epoch 4, total_step 69800, total loss is 19.50 , inference loss is 10.09, weight deacy loss is 9.41, training accuracy is 0.312500, time 55.579 samples/sec
+epoch 4, total_step 69820, total loss is 18.13 , inference loss is 8.72, weight deacy loss is 9.41, training accuracy is 0.375000, time 56.964 samples/sec
+epoch 4, total_step 69840, total loss is 22.32 , inference loss is 12.91, weight deacy loss is 9.41, training accuracy is 0.218750, time 56.611 samples/sec
+epoch 4, total_step 69860, total loss is 18.07 , inference loss is 8.66, weight deacy loss is 9.41, training accuracy is 0.343750, time 55.105 samples/sec
+epoch 4, total_step 69880, total loss is 19.27 , inference loss is 9.86, weight deacy loss is 9.41, training accuracy is 0.406250, time 56.404 samples/sec
+epoch 4, total_step 69900, total loss is 21.54 , inference loss is 12.14, weight deacy loss is 9.41, training accuracy is 0.218750, time 54.011 samples/sec
+epoch 4, total_step 69920, total loss is 17.05 , inference loss is 7.65, weight deacy loss is 9.41, training accuracy is 0.312500, time 56.121 samples/sec
+epoch 4, total_step 69940, total loss is 20.23 , inference loss is 10.83, weight deacy loss is 9.40, training accuracy is 0.218750, time 57.287 samples/sec
+epoch 4, total_step 69960, total loss is 19.49 , inference loss is 10.09, weight deacy loss is 9.40, training accuracy is 0.281250, time 56.175 samples/sec
+epoch 4, total_step 69980, total loss is 23.86 , inference loss is 14.45, weight deacy loss is 9.40, training accuracy is 0.187500, time 55.569 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.32174199999999
+best_threshold_index 110 0.987037037037037
+best_threshold_index 110 0.9855555555555555
+best_threshold_index 110 0.9872222222222222
+best_threshold_index 110 0.9868518518518519
+best_threshold_index 110 0.9866666666666667
+best_threshold_index 110 0.9857407407407407
+best_threshold_index 110 0.9868518518518519
+best_threshold_index 110 0.9855555555555555
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 110 0.9848148148148148
+[lfw][70000]XNorm: 21.619157
+[lfw][70000]Accuracy-Flip: 0.98617+-0.00723
+testing verification..
+(14000, 512)
+infer time 85.832773
+best_threshold_index 114 0.9855555555555555
+best_threshold_index 114 0.9836507936507937
+best_threshold_index 114 0.9842857142857143
+best_threshold_index 114 0.9839682539682539
+best_threshold_index 114 0.9844444444444445
+best_threshold_index 114 0.9838095238095238
+best_threshold_index 114 0.9842857142857143
+best_threshold_index 114 0.9841269841269841
+best_threshold_index 114 0.9841269841269841
+best_threshold_index 114 0.9846031746031746
+[cfp_ff][70000]XNorm: 20.797693
+[cfp_ff][70000]Accuracy-Flip: 0.98429+-0.00452
+testing verification..
+(14000, 512)
+infer time 86.36746599999985
+best_threshold_index 138 0.9103174603174603
+best_threshold_index 138 0.909047619047619
+best_threshold_index 138 0.9085714285714286
+best_threshold_index 136 0.9088888888888889
+best_threshold_index 138 0.9104761904761904
+best_threshold_index 138 0.9068253968253969
+best_threshold_index 138 0.906984126984127
+best_threshold_index 138 0.9074603174603174
+best_threshold_index 138 0.9080952380952381
+best_threshold_index 138 0.9092063492063492
+[cfp_fp][70000]XNorm: 20.659394
+[cfp_fp][70000]Accuracy-Flip: 0.90800+-0.01100
+testing verification..
+(12000, 512)
+infer time 73.43159300000003
+best_threshold_index 137 0.8838888888888888
+best_threshold_index 137 0.8861111111111111
+best_threshold_index 135 0.8881481481481481
+best_threshold_index 137 0.8875925925925926
+best_threshold_index 136 0.8912962962962963
+best_threshold_index 137 0.8853703703703704
+best_threshold_index 137 0.8877777777777778
+best_threshold_index 137 0.892962962962963
+best_threshold_index 137 0.8914814814814814
+best_threshold_index 137 0.8888888888888888
+[agedb_30][70000]XNorm: 22.150193
+[agedb_30][70000]Accuracy-Flip: 0.88717+-0.02501
+test accuracy is: 0.9861666666666666
+epoch 4, total_step 70000, total loss is 20.97 , inference loss is 11.56, weight deacy loss is 9.40, training accuracy is 0.406250, time 54.430 samples/sec
+epoch 4, total_step 70020, total loss is 23.15 , inference loss is 13.75, weight deacy loss is 9.40, training accuracy is 0.187500, time 55.282 samples/sec
+epoch 4, total_step 70040, total loss is 26.29 , inference loss is 16.89, weight deacy loss is 9.40, training accuracy is 0.187500, time 56.167 samples/sec
+epoch 4, total_step 70060, total loss is 23.90 , inference loss is 14.51, weight deacy loss is 9.40, training accuracy is 0.187500, time 56.114 samples/sec
+epoch 4, total_step 70080, total loss is 22.27 , inference loss is 12.87, weight deacy loss is 9.40, training accuracy is 0.281250, time 56.515 samples/sec
+epoch 4, total_step 70100, total loss is 23.26 , inference loss is 13.87, weight deacy loss is 9.40, training accuracy is 0.187500, time 56.314 samples/sec
+epoch 4, total_step 70120, total loss is 20.76 , inference loss is 11.36, weight deacy loss is 9.40, training accuracy is 0.218750, time 55.969 samples/sec
+epoch 4, total_step 70140, total loss is 19.45 , inference loss is 10.06, weight deacy loss is 9.39, training accuracy is 0.250000, time 54.031 samples/sec
+epoch 4, total_step 70160, total loss is 18.58 , inference loss is 9.18, weight deacy loss is 9.39, training accuracy is 0.312500, time 56.123 samples/sec
+epoch 4, total_step 70180, total loss is 20.69 , inference loss is 11.30, weight deacy loss is 9.39, training accuracy is 0.125000, time 56.045 samples/sec
+epoch 4, total_step 70200, total loss is 20.79 , inference loss is 11.40, weight deacy loss is 9.39, training accuracy is 0.281250, time 56.812 samples/sec
+epoch 4, total_step 70220, total loss is 22.28 , inference loss is 12.89, weight deacy loss is 9.39, training accuracy is 0.312500, time 56.452 samples/sec
+epoch 4, total_step 70240, total loss is 20.93 , inference loss is 11.54, weight deacy loss is 9.39, training accuracy is 0.343750, time 54.250 samples/sec
+epoch 4, total_step 70260, total loss is 27.27 , inference loss is 17.89, weight deacy loss is 9.39, training accuracy is 0.187500, time 56.277 samples/sec
+epoch 4, total_step 70280, total loss is 17.20 , inference loss is 7.81, weight deacy loss is 9.39, training accuracy is 0.281250, time 56.590 samples/sec
+epoch 4, total_step 70300, total loss is 21.47 , inference loss is 12.08, weight deacy loss is 9.39, training accuracy is 0.281250, time 55.818 samples/sec
+epoch 4, total_step 70320, total loss is 25.50 , inference loss is 16.12, weight deacy loss is 9.38, training accuracy is 0.187500, time 55.888 samples/sec
+epoch 4, total_step 70340, total loss is 21.46 , inference loss is 12.07, weight deacy loss is 9.38, training accuracy is 0.437500, time 56.115 samples/sec
+epoch 4, total_step 70360, total loss is 22.30 , inference loss is 12.92, weight deacy loss is 9.38, training accuracy is 0.312500, time 56.082 samples/sec
+epoch 4, total_step 70380, total loss is 20.46 , inference loss is 11.07, weight deacy loss is 9.38, training accuracy is 0.250000, time 56.169 samples/sec
+epoch 4, total_step 70400, total loss is 19.90 , inference loss is 10.51, weight deacy loss is 9.38, training accuracy is 0.281250, time 56.098 samples/sec
+epoch 4, total_step 70420, total loss is 18.99 , inference loss is 9.61, weight deacy loss is 9.38, training accuracy is 0.375000, time 56.640 samples/sec
+epoch 4, total_step 70440, total loss is 19.84 , inference loss is 10.46, weight deacy loss is 9.38, training accuracy is 0.312500, time 56.232 samples/sec
+epoch 4, total_step 70460, total loss is 19.93 , inference loss is 10.56, weight deacy loss is 9.38, training accuracy is 0.250000, time 56.277 samples/sec
+epoch 4, total_step 70480, total loss is 22.25 , inference loss is 12.87, weight deacy loss is 9.38, training accuracy is 0.187500, time 56.111 samples/sec
+epoch 4, total_step 70500, total loss is 18.12 , inference loss is 8.75, weight deacy loss is 9.38, training accuracy is 0.375000, time 55.222 samples/sec
+epoch 4, total_step 70520, total loss is 19.44 , inference loss is 10.06, weight deacy loss is 9.37, training accuracy is 0.312500, time 55.985 samples/sec
+epoch 4, total_step 70540, total loss is 20.03 , inference loss is 10.66, weight deacy loss is 9.37, training accuracy is 0.312500, time 56.573 samples/sec
+epoch 4, total_step 70560, total loss is 20.54 , inference loss is 11.17, weight deacy loss is 9.37, training accuracy is 0.250000, time 55.325 samples/sec
+epoch 4, total_step 70580, total loss is 23.83 , inference loss is 14.46, weight deacy loss is 9.37, training accuracy is 0.156250, time 55.396 samples/sec
+epoch 4, total_step 70600, total loss is 23.55 , inference loss is 14.17, weight deacy loss is 9.37, training accuracy is 0.156250, time 55.173 samples/sec
+epoch 4, total_step 70620, total loss is 19.37 , inference loss is 10.00, weight deacy loss is 9.37, training accuracy is 0.406250, time 56.513 samples/sec
+epoch 4, total_step 70640, total loss is 19.14 , inference loss is 9.78, weight deacy loss is 9.37, training accuracy is 0.281250, time 55.998 samples/sec
+epoch 4, total_step 70660, total loss is 22.40 , inference loss is 13.03, weight deacy loss is 9.37, training accuracy is 0.156250, time 56.615 samples/sec
+epoch 4, total_step 70680, total loss is 26.63 , inference loss is 17.26, weight deacy loss is 9.37, training accuracy is 0.187500, time 56.266 samples/sec
+epoch 4, total_step 70700, total loss is 21.83 , inference loss is 12.46, weight deacy loss is 9.36, training accuracy is 0.218750, time 56.526 samples/sec
+epoch 4, total_step 70720, total loss is 22.45 , inference loss is 13.08, weight deacy loss is 9.36, training accuracy is 0.218750, time 55.578 samples/sec
+epoch 4, total_step 70740, total loss is 21.48 , inference loss is 12.11, weight deacy loss is 9.36, training accuracy is 0.312500, time 56.865 samples/sec
+epoch 4, total_step 70760, total loss is 22.25 , inference loss is 12.89, weight deacy loss is 9.36, training accuracy is 0.187500, time 56.072 samples/sec
+epoch 4, total_step 70780, total loss is 21.47 , inference loss is 12.11, weight deacy loss is 9.36, training accuracy is 0.250000, time 56.914 samples/sec
+epoch 4, total_step 70800, total loss is 28.15 , inference loss is 18.79, weight deacy loss is 9.36, training accuracy is 0.093750, time 56.096 samples/sec
+epoch 4, total_step 70820, total loss is 24.78 , inference loss is 15.42, weight deacy loss is 9.36, training accuracy is 0.156250, time 55.194 samples/sec
+epoch 4, total_step 70840, total loss is 22.35 , inference loss is 12.99, weight deacy loss is 9.36, training accuracy is 0.281250, time 57.666 samples/sec
+epoch 4, total_step 70860, total loss is 23.78 , inference loss is 14.43, weight deacy loss is 9.36, training accuracy is 0.281250, time 56.149 samples/sec
+epoch 4, total_step 70880, total loss is 23.67 , inference loss is 14.31, weight deacy loss is 9.36, training accuracy is 0.281250, time 55.750 samples/sec
+epoch 4, total_step 70900, total loss is 20.47 , inference loss is 11.11, weight deacy loss is 9.35, training accuracy is 0.250000, time 56.145 samples/sec
+epoch 4, total_step 70920, total loss is 18.03 , inference loss is 8.68, weight deacy loss is 9.35, training accuracy is 0.406250, time 56.879 samples/sec
+epoch 4, total_step 70940, total loss is 25.03 , inference loss is 15.68, weight deacy loss is 9.35, training accuracy is 0.156250, time 56.834 samples/sec
+epoch 4, total_step 70960, total loss is 22.65 , inference loss is 13.30, weight deacy loss is 9.35, training accuracy is 0.312500, time 56.310 samples/sec
+epoch 4, total_step 70980, total loss is 19.38 , inference loss is 10.03, weight deacy loss is 9.35, training accuracy is 0.218750, time 56.246 samples/sec
+epoch 4, total_step 71000, total loss is 20.70 , inference loss is 11.35, weight deacy loss is 9.35, training accuracy is 0.218750, time 56.346 samples/sec
+epoch 4, total_step 71020, total loss is 24.26 , inference loss is 14.91, weight deacy loss is 9.35, training accuracy is 0.250000, time 56.329 samples/sec
+epoch 4, total_step 71040, total loss is 20.65 , inference loss is 11.30, weight deacy loss is 9.35, training accuracy is 0.375000, time 56.020 samples/sec
+epoch 4, total_step 71060, total loss is 26.55 , inference loss is 17.21, weight deacy loss is 9.35, training accuracy is 0.218750, time 56.464 samples/sec
+epoch 4, total_step 71080, total loss is 17.19 , inference loss is 7.85, weight deacy loss is 9.35, training accuracy is 0.343750, time 56.333 samples/sec
+epoch 4, total_step 71100, total loss is 23.59 , inference loss is 14.25, weight deacy loss is 9.34, training accuracy is 0.250000, time 55.832 samples/sec
+epoch 4, total_step 71120, total loss is 17.50 , inference loss is 8.15, weight deacy loss is 9.34, training accuracy is 0.375000, time 55.798 samples/sec
+epoch 4, total_step 71140, total loss is 23.31 , inference loss is 13.96, weight deacy loss is 9.34, training accuracy is 0.218750, time 56.827 samples/sec
+epoch 4, total_step 71160, total loss is 20.41 , inference loss is 11.07, weight deacy loss is 9.34, training accuracy is 0.281250, time 56.512 samples/sec
+epoch 4, total_step 71180, total loss is 23.16 , inference loss is 13.82, weight deacy loss is 9.34, training accuracy is 0.125000, time 55.961 samples/sec
+epoch 4, total_step 71200, total loss is 23.41 , inference loss is 14.07, weight deacy loss is 9.34, training accuracy is 0.218750, time 55.869 samples/sec
+epoch 4, total_step 71220, total loss is 19.34 , inference loss is 10.00, weight deacy loss is 9.34, training accuracy is 0.187500, time 56.451 samples/sec
+epoch 4, total_step 71240, total loss is 20.49 , inference loss is 11.15, weight deacy loss is 9.34, training accuracy is 0.250000, time 55.530 samples/sec
+epoch 4, total_step 71260, total loss is 22.49 , inference loss is 13.15, weight deacy loss is 9.34, training accuracy is 0.343750, time 53.130 samples/sec
+epoch 4, total_step 71280, total loss is 22.27 , inference loss is 12.94, weight deacy loss is 9.33, training accuracy is 0.187500, time 56.194 samples/sec
+epoch 4, total_step 71300, total loss is 21.31 , inference loss is 11.98, weight deacy loss is 9.33, training accuracy is 0.312500, time 55.761 samples/sec
+epoch 4, total_step 71320, total loss is 21.46 , inference loss is 12.12, weight deacy loss is 9.33, training accuracy is 0.281250, time 56.176 samples/sec
+epoch 4, total_step 71340, total loss is 27.98 , inference loss is 18.65, weight deacy loss is 9.33, training accuracy is 0.062500, time 56.573 samples/sec
+epoch 4, total_step 71360, total loss is 22.38 , inference loss is 13.05, weight deacy loss is 9.33, training accuracy is 0.218750, time 56.602 samples/sec
+epoch 4, total_step 71380, total loss is 22.25 , inference loss is 12.92, weight deacy loss is 9.33, training accuracy is 0.218750, time 56.188 samples/sec
+epoch 4, total_step 71400, total loss is 20.80 , inference loss is 11.48, weight deacy loss is 9.33, training accuracy is 0.250000, time 56.389 samples/sec
+epoch 4, total_step 71420, total loss is 23.77 , inference loss is 14.44, weight deacy loss is 9.33, training accuracy is 0.218750, time 55.662 samples/sec
+epoch 4, total_step 71440, total loss is 19.71 , inference loss is 10.38, weight deacy loss is 9.33, training accuracy is 0.312500, time 56.025 samples/sec
+epoch 4, total_step 71460, total loss is 15.85 , inference loss is 6.52, weight deacy loss is 9.33, training accuracy is 0.375000, time 56.197 samples/sec
+epoch 4, total_step 71480, total loss is 21.49 , inference loss is 12.17, weight deacy loss is 9.32, training accuracy is 0.218750, time 56.209 samples/sec
+epoch 4, total_step 71500, total loss is 19.51 , inference loss is 10.19, weight deacy loss is 9.32, training accuracy is 0.250000, time 56.694 samples/sec
+epoch 4, total_step 71520, total loss is 23.06 , inference loss is 13.73, weight deacy loss is 9.32, training accuracy is 0.218750, time 56.245 samples/sec
+epoch 4, total_step 71540, total loss is 19.47 , inference loss is 10.15, weight deacy loss is 9.32, training accuracy is 0.406250, time 55.772 samples/sec
+epoch 4, total_step 71560, total loss is 22.50 , inference loss is 13.18, weight deacy loss is 9.32, training accuracy is 0.187500, time 56.121 samples/sec
+epoch 4, total_step 71580, total loss is 20.50 , inference loss is 11.18, weight deacy loss is 9.32, training accuracy is 0.250000, time 56.083 samples/sec
+epoch 4, total_step 71600, total loss is 20.17 , inference loss is 10.86, weight deacy loss is 9.32, training accuracy is 0.312500, time 56.560 samples/sec
+epoch 4, total_step 71620, total loss is 20.97 , inference loss is 11.65, weight deacy loss is 9.32, training accuracy is 0.218750, time 53.270 samples/sec
+epoch 4, total_step 71640, total loss is 16.39 , inference loss is 7.07, weight deacy loss is 9.32, training accuracy is 0.375000, time 55.679 samples/sec
+epoch 4, total_step 71660, total loss is 18.65 , inference loss is 9.34, weight deacy loss is 9.31, training accuracy is 0.218750, time 54.658 samples/sec
+epoch 4, total_step 71680, total loss is 16.04 , inference loss is 6.73, weight deacy loss is 9.31, training accuracy is 0.375000, time 56.294 samples/sec
+epoch 4, total_step 71700, total loss is 19.56 , inference loss is 10.24, weight deacy loss is 9.31, training accuracy is 0.250000, time 55.965 samples/sec
+epoch 4, total_step 71720, total loss is 19.18 , inference loss is 9.87, weight deacy loss is 9.31, training accuracy is 0.375000, time 56.829 samples/sec
+epoch 4, total_step 71740, total loss is 20.67 , inference loss is 11.36, weight deacy loss is 9.31, training accuracy is 0.343750, time 55.883 samples/sec
+epoch 4, total_step 71760, total loss is 19.11 , inference loss is 9.80, weight deacy loss is 9.31, training accuracy is 0.250000, time 55.472 samples/sec
+epoch 4, total_step 71780, total loss is 18.01 , inference loss is 8.71, weight deacy loss is 9.31, training accuracy is 0.281250, time 55.682 samples/sec
+epoch 4, total_step 71800, total loss is 18.72 , inference loss is 9.41, weight deacy loss is 9.31, training accuracy is 0.343750, time 56.787 samples/sec
+epoch 4, total_step 71820, total loss is 20.79 , inference loss is 11.48, weight deacy loss is 9.31, training accuracy is 0.406250, time 57.118 samples/sec
+epoch 4, total_step 71840, total loss is 21.74 , inference loss is 12.43, weight deacy loss is 9.31, training accuracy is 0.312500, time 55.869 samples/sec
+epoch 4, total_step 71860, total loss is 18.68 , inference loss is 9.38, weight deacy loss is 9.30, training accuracy is 0.343750, time 56.662 samples/sec
+epoch 4, total_step 71880, total loss is 25.11 , inference loss is 15.80, weight deacy loss is 9.30, training accuracy is 0.281250, time 56.381 samples/sec
+epoch 4, total_step 71900, total loss is 21.90 , inference loss is 12.60, weight deacy loss is 9.30, training accuracy is 0.281250, time 57.168 samples/sec
+epoch 4, total_step 71920, total loss is 19.47 , inference loss is 10.17, weight deacy loss is 9.30, training accuracy is 0.312500, time 55.216 samples/sec
+epoch 4, total_step 71940, total loss is 23.38 , inference loss is 14.08, weight deacy loss is 9.30, training accuracy is 0.218750, time 56.348 samples/sec
+epoch 4, total_step 71960, total loss is 20.62 , inference loss is 11.32, weight deacy loss is 9.30, training accuracy is 0.156250, time 53.703 samples/sec
+epoch 4, total_step 71980, total loss is 20.43 , inference loss is 11.14, weight deacy loss is 9.30, training accuracy is 0.343750, time 56.398 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.78750100000006
+best_threshold_index 101 0.9835185185185186
+best_threshold_index 101 0.982037037037037
+best_threshold_index 101 0.9837037037037037
+best_threshold_index 101 0.9840740740740741
+best_threshold_index 101 0.9825925925925926
+best_threshold_index 110 0.9818518518518519
+best_threshold_index 103 0.9829629629629629
+best_threshold_index 103 0.9812962962962963
+best_threshold_index 101 0.9807407407407407
+best_threshold_index 101 0.9814814814814815
+[lfw][72000]XNorm: 22.262635
+[lfw][72000]Accuracy-Flip: 0.98083+-0.00944
+testing verification..
+(14000, 512)
+infer time 86.29105800000008
+best_threshold_index 114 0.9839682539682539
+best_threshold_index 120 0.9831746031746031
+best_threshold_index 114 0.9822222222222222
+best_threshold_index 118 0.9828571428571429
+best_threshold_index 117 0.9836507936507937
+best_threshold_index 114 0.9828571428571429
+best_threshold_index 116 0.9833333333333333
+best_threshold_index 117 0.9823809523809524
+best_threshold_index 118 0.9831746031746031
+best_threshold_index 120 0.983015873015873
+[cfp_ff][72000]XNorm: 22.245158
+[cfp_ff][72000]Accuracy-Flip: 0.98157+-0.00463
+testing verification..
+(14000, 512)
+infer time 84.99238900000009
+best_threshold_index 137 0.9087301587301587
+best_threshold_index 136 0.9066666666666666
+best_threshold_index 135 0.9057142857142857
+best_threshold_index 136 0.9098412698412699
+best_threshold_index 137 0.9096825396825396
+best_threshold_index 135 0.9047619047619048
+best_threshold_index 133 0.9047619047619048
+best_threshold_index 136 0.9061904761904762
+best_threshold_index 135 0.9071428571428571
+best_threshold_index 135 0.9076190476190477
+[cfp_fp][72000]XNorm: 21.991281
+[cfp_fp][72000]Accuracy-Flip: 0.90471+-0.01539
+testing verification..
+(12000, 512)
+infer time 73.50211399999995
+best_threshold_index 141 0.887962962962963
+best_threshold_index 141 0.8861111111111111
+best_threshold_index 141 0.8877777777777778
+best_threshold_index 141 0.887962962962963
+best_threshold_index 138 0.8911111111111111
+best_threshold_index 141 0.8868518518518519
+best_threshold_index 141 0.8874074074074074
+best_threshold_index 141 0.8937037037037037
+best_threshold_index 141 0.8901851851851852
+best_threshold_index 141 0.8911111111111111
+[agedb_30][72000]XNorm: 22.256264
+[agedb_30][72000]Accuracy-Flip: 0.88817+-0.02114
+test accuracy is: 0.9808333333333333
+epoch 4, total_step 72000, total loss is 19.53 , inference loss is 10.24, weight deacy loss is 9.30, training accuracy is 0.343750, time 54.233 samples/sec
+epoch 4, total_step 72020, total loss is 21.03 , inference loss is 11.74, weight deacy loss is 9.30, training accuracy is 0.250000, time 54.581 samples/sec
+epoch 4, total_step 72040, total loss is 21.96 , inference loss is 12.66, weight deacy loss is 9.29, training accuracy is 0.250000, time 55.834 samples/sec
+epoch 4, total_step 72060, total loss is 17.18 , inference loss is 7.89, weight deacy loss is 9.29, training accuracy is 0.281250, time 55.748 samples/sec
+epoch 4, total_step 72080, total loss is 23.24 , inference loss is 13.95, weight deacy loss is 9.29, training accuracy is 0.312500, time 56.736 samples/sec
+epoch 4, total_step 72100, total loss is 19.94 , inference loss is 10.65, weight deacy loss is 9.29, training accuracy is 0.312500, time 54.673 samples/sec
+epoch 4, total_step 72120, total loss is 25.71 , inference loss is 16.42, weight deacy loss is 9.29, training accuracy is 0.187500, time 56.386 samples/sec
+epoch 4, total_step 72140, total loss is 21.45 , inference loss is 12.16, weight deacy loss is 9.29, training accuracy is 0.250000, time 56.136 samples/sec
+epoch 4, total_step 72160, total loss is 21.33 , inference loss is 12.04, weight deacy loss is 9.29, training accuracy is 0.281250, time 57.021 samples/sec
+epoch 4, total_step 72180, total loss is 20.90 , inference loss is 11.62, weight deacy loss is 9.29, training accuracy is 0.375000, time 56.673 samples/sec
+epoch 4, total_step 72200, total loss is 22.10 , inference loss is 12.81, weight deacy loss is 9.29, training accuracy is 0.343750, time 56.380 samples/sec
+epoch 4, total_step 72220, total loss is 22.93 , inference loss is 13.64, weight deacy loss is 9.29, training accuracy is 0.218750, time 55.638 samples/sec
+epoch 4, total_step 72240, total loss is 22.56 , inference loss is 13.28, weight deacy loss is 9.28, training accuracy is 0.312500, time 56.249 samples/sec
+epoch 4, total_step 72260, total loss is 17.40 , inference loss is 8.11, weight deacy loss is 9.28, training accuracy is 0.375000, time 55.983 samples/sec
+epoch 4, total_step 72280, total loss is 20.08 , inference loss is 10.79, weight deacy loss is 9.28, training accuracy is 0.281250, time 55.996 samples/sec
+epoch 4, total_step 72300, total loss is 20.03 , inference loss is 10.75, weight deacy loss is 9.28, training accuracy is 0.218750, time 55.247 samples/sec
+epoch 4, total_step 72320, total loss is 19.34 , inference loss is 10.06, weight deacy loss is 9.28, training accuracy is 0.343750, time 56.058 samples/sec
+epoch 4, total_step 72340, total loss is 19.27 , inference loss is 9.99, weight deacy loss is 9.28, training accuracy is 0.343750, time 55.259 samples/sec
+epoch 4, total_step 72360, total loss is 19.19 , inference loss is 9.91, weight deacy loss is 9.28, training accuracy is 0.375000, time 55.515 samples/sec
+epoch 4, total_step 72380, total loss is 16.63 , inference loss is 7.36, weight deacy loss is 9.28, training accuracy is 0.281250, time 57.081 samples/sec
+epoch 4, total_step 72400, total loss is 20.04 , inference loss is 10.76, weight deacy loss is 9.28, training accuracy is 0.281250, time 56.225 samples/sec
+epoch 4, total_step 72420, total loss is 17.61 , inference loss is 8.33, weight deacy loss is 9.28, training accuracy is 0.375000, time 56.652 samples/sec
+epoch 4, total_step 72440, total loss is 22.31 , inference loss is 13.03, weight deacy loss is 9.27, training accuracy is 0.125000, time 56.182 samples/sec
+epoch 4, total_step 72460, total loss is 16.64 , inference loss is 7.37, weight deacy loss is 9.27, training accuracy is 0.406250, time 56.169 samples/sec
+epoch 4, total_step 72480, total loss is 24.87 , inference loss is 15.59, weight deacy loss is 9.27, training accuracy is 0.218750, time 55.795 samples/sec
+epoch 4, total_step 72500, total loss is 20.75 , inference loss is 11.48, weight deacy loss is 9.27, training accuracy is 0.312500, time 56.263 samples/sec
+epoch 4, total_step 72520, total loss is 20.80 , inference loss is 11.53, weight deacy loss is 9.27, training accuracy is 0.156250, time 54.595 samples/sec
+epoch 4, total_step 72540, total loss is 21.37 , inference loss is 12.10, weight deacy loss is 9.27, training accuracy is 0.250000, time 55.997 samples/sec
+epoch 4, total_step 72560, total loss is 22.87 , inference loss is 13.60, weight deacy loss is 9.27, training accuracy is 0.156250, time 55.926 samples/sec
+epoch 4, total_step 72580, total loss is 22.06 , inference loss is 12.80, weight deacy loss is 9.27, training accuracy is 0.250000, time 56.429 samples/sec
+epoch 4, total_step 72600, total loss is 19.08 , inference loss is 9.81, weight deacy loss is 9.27, training accuracy is 0.281250, time 55.874 samples/sec
+epoch 4, total_step 72620, total loss is 22.71 , inference loss is 13.44, weight deacy loss is 9.27, training accuracy is 0.250000, time 56.652 samples/sec
+epoch 4, total_step 72640, total loss is 20.19 , inference loss is 10.92, weight deacy loss is 9.26, training accuracy is 0.250000, time 55.458 samples/sec
+epoch 4, total_step 72660, total loss is 17.61 , inference loss is 8.34, weight deacy loss is 9.26, training accuracy is 0.343750, time 55.856 samples/sec
+epoch 4, total_step 72680, total loss is 21.12 , inference loss is 11.86, weight deacy loss is 9.26, training accuracy is 0.156250, time 55.879 samples/sec
+epoch 4, total_step 72700, total loss is 18.02 , inference loss is 8.76, weight deacy loss is 9.26, training accuracy is 0.250000, time 55.873 samples/sec
+epoch 4, total_step 72720, total loss is 23.29 , inference loss is 14.03, weight deacy loss is 9.26, training accuracy is 0.156250, time 56.363 samples/sec
+epoch 4, total_step 72740, total loss is 16.25 , inference loss is 6.99, weight deacy loss is 9.26, training accuracy is 0.468750, time 56.996 samples/sec
+epoch 4, total_step 72760, total loss is 23.33 , inference loss is 14.08, weight deacy loss is 9.26, training accuracy is 0.187500, time 55.969 samples/sec
+epoch 4, total_step 72780, total loss is 21.78 , inference loss is 12.52, weight deacy loss is 9.26, training accuracy is 0.281250, time 56.250 samples/sec
+epoch 4, total_step 72800, total loss is 24.13 , inference loss is 14.88, weight deacy loss is 9.26, training accuracy is 0.125000, time 56.573 samples/sec
+epoch 4, total_step 72820, total loss is 18.36 , inference loss is 9.10, weight deacy loss is 9.26, training accuracy is 0.250000, time 56.177 samples/sec
+epoch 4, total_step 72840, total loss is 28.24 , inference loss is 18.99, weight deacy loss is 9.25, training accuracy is 0.156250, time 56.301 samples/sec
+epoch 4, total_step 72860, total loss is 19.60 , inference loss is 10.35, weight deacy loss is 9.25, training accuracy is 0.187500, time 56.257 samples/sec
+epoch 4, total_step 72880, total loss is 23.03 , inference loss is 13.78, weight deacy loss is 9.25, training accuracy is 0.218750, time 56.663 samples/sec
+epoch 4, total_step 72900, total loss is 26.04 , inference loss is 16.79, weight deacy loss is 9.25, training accuracy is 0.031250, time 56.313 samples/sec
+epoch 4, total_step 72920, total loss is 15.11 , inference loss is 5.86, weight deacy loss is 9.25, training accuracy is 0.375000, time 56.768 samples/sec
+epoch 4, total_step 72940, total loss is 23.14 , inference loss is 13.89, weight deacy loss is 9.25, training accuracy is 0.156250, time 56.055 samples/sec
+epoch 4, total_step 72960, total loss is 21.57 , inference loss is 12.32, weight deacy loss is 9.25, training accuracy is 0.156250, time 56.042 samples/sec
+epoch 4, total_step 72980, total loss is 21.08 , inference loss is 11.84, weight deacy loss is 9.25, training accuracy is 0.125000, time 56.520 samples/sec
+epoch 4, total_step 73000, total loss is 21.55 , inference loss is 12.31, weight deacy loss is 9.25, training accuracy is 0.375000, time 55.726 samples/sec
+epoch 4, total_step 73020, total loss is 23.87 , inference loss is 14.63, weight deacy loss is 9.24, training accuracy is 0.125000, time 56.428 samples/sec
+epoch 4, total_step 73040, total loss is 15.22 , inference loss is 5.97, weight deacy loss is 9.24, training accuracy is 0.437500, time 56.123 samples/sec
+epoch 4, total_step 73060, total loss is 22.08 , inference loss is 12.84, weight deacy loss is 9.24, training accuracy is 0.156250, time 56.909 samples/sec
+epoch 4, total_step 73080, total loss is 20.62 , inference loss is 11.38, weight deacy loss is 9.24, training accuracy is 0.125000, time 54.958 samples/sec
+epoch 4, total_step 73100, total loss is 25.18 , inference loss is 15.94, weight deacy loss is 9.24, training accuracy is 0.093750, time 56.274 samples/sec
+epoch 4, total_step 73120, total loss is 21.62 , inference loss is 12.38, weight deacy loss is 9.24, training accuracy is 0.250000, time 55.504 samples/sec
+epoch 4, total_step 73140, total loss is 23.73 , inference loss is 14.49, weight deacy loss is 9.24, training accuracy is 0.156250, time 55.933 samples/sec
+epoch 4, total_step 73160, total loss is 21.10 , inference loss is 11.86, weight deacy loss is 9.24, training accuracy is 0.312500, time 55.413 samples/sec
+epoch 4, total_step 73180, total loss is 20.78 , inference loss is 11.54, weight deacy loss is 9.24, training accuracy is 0.125000, time 56.006 samples/sec
+epoch 4, total_step 73200, total loss is 20.28 , inference loss is 11.04, weight deacy loss is 9.24, training accuracy is 0.343750, time 56.216 samples/sec
+epoch 4, total_step 73220, total loss is 23.28 , inference loss is 14.05, weight deacy loss is 9.23, training accuracy is 0.218750, time 55.359 samples/sec
+epoch 4, total_step 73240, total loss is 23.08 , inference loss is 13.85, weight deacy loss is 9.23, training accuracy is 0.312500, time 55.787 samples/sec
+epoch 4, total_step 73260, total loss is 17.55 , inference loss is 8.32, weight deacy loss is 9.23, training accuracy is 0.187500, time 55.584 samples/sec
+epoch 4, total_step 73280, total loss is 26.38 , inference loss is 17.15, weight deacy loss is 9.23, training accuracy is 0.156250, time 56.682 samples/sec
+epoch 4, total_step 73300, total loss is 22.83 , inference loss is 13.60, weight deacy loss is 9.23, training accuracy is 0.250000, time 56.891 samples/sec
+epoch 4, total_step 73320, total loss is 27.00 , inference loss is 17.77, weight deacy loss is 9.23, training accuracy is 0.125000, time 56.534 samples/sec
+epoch 4, total_step 73340, total loss is 22.56 , inference loss is 13.33, weight deacy loss is 9.23, training accuracy is 0.093750, time 56.694 samples/sec
+epoch 4, total_step 73360, total loss is 18.15 , inference loss is 8.93, weight deacy loss is 9.23, training accuracy is 0.343750, time 56.541 samples/sec
+epoch 4, total_step 73380, total loss is 21.64 , inference loss is 12.41, weight deacy loss is 9.23, training accuracy is 0.281250, time 55.568 samples/sec
+epoch 4, total_step 73400, total loss is 22.15 , inference loss is 12.92, weight deacy loss is 9.23, training accuracy is 0.250000, time 56.388 samples/sec
+epoch 4, total_step 73420, total loss is 24.04 , inference loss is 14.81, weight deacy loss is 9.22, training accuracy is 0.156250, time 56.213 samples/sec
+epoch 4, total_step 73440, total loss is 25.11 , inference loss is 15.89, weight deacy loss is 9.22, training accuracy is 0.250000, time 55.352 samples/sec
+epoch 4, total_step 73460, total loss is 21.39 , inference loss is 12.17, weight deacy loss is 9.22, training accuracy is 0.125000, time 56.976 samples/sec
+epoch 4, total_step 73480, total loss is 20.88 , inference loss is 11.66, weight deacy loss is 9.22, training accuracy is 0.281250, time 56.863 samples/sec
+epoch 4, total_step 73500, total loss is 22.23 , inference loss is 13.01, weight deacy loss is 9.22, training accuracy is 0.281250, time 55.905 samples/sec
+epoch 4, total_step 73520, total loss is 22.29 , inference loss is 13.07, weight deacy loss is 9.22, training accuracy is 0.218750, time 56.469 samples/sec
+epoch 4, total_step 73540, total loss is 16.98 , inference loss is 7.76, weight deacy loss is 9.22, training accuracy is 0.281250, time 55.647 samples/sec
+epoch 4, total_step 73560, total loss is 24.85 , inference loss is 15.64, weight deacy loss is 9.22, training accuracy is 0.156250, time 56.814 samples/sec
+epoch 4, total_step 73580, total loss is 23.37 , inference loss is 14.15, weight deacy loss is 9.22, training accuracy is 0.093750, time 56.115 samples/sec
+epoch 4, total_step 73600, total loss is 21.26 , inference loss is 12.04, weight deacy loss is 9.22, training accuracy is 0.062500, time 56.683 samples/sec
+epoch 4, total_step 73620, total loss is 23.56 , inference loss is 14.34, weight deacy loss is 9.21, training accuracy is 0.187500, time 56.035 samples/sec
+epoch 4, total_step 73640, total loss is 25.83 , inference loss is 16.61, weight deacy loss is 9.21, training accuracy is 0.187500, time 52.879 samples/sec
+epoch 4, total_step 73660, total loss is 23.06 , inference loss is 13.85, weight deacy loss is 9.21, training accuracy is 0.156250, time 56.609 samples/sec
+epoch 4, total_step 73680, total loss is 18.63 , inference loss is 9.42, weight deacy loss is 9.21, training accuracy is 0.281250, time 52.836 samples/sec
+epoch 4, total_step 73700, total loss is 26.63 , inference loss is 17.42, weight deacy loss is 9.21, training accuracy is 0.125000, time 55.822 samples/sec
+epoch 4, total_step 73720, total loss is 19.43 , inference loss is 10.22, weight deacy loss is 9.21, training accuracy is 0.250000, time 54.384 samples/sec
+epoch 4, total_step 73740, total loss is 21.10 , inference loss is 11.89, weight deacy loss is 9.21, training accuracy is 0.250000, time 56.125 samples/sec
+epoch 4, total_step 73760, total loss is 24.61 , inference loss is 15.41, weight deacy loss is 9.21, training accuracy is 0.156250, time 56.429 samples/sec
+epoch 4, total_step 73780, total loss is 25.34 , inference loss is 16.13, weight deacy loss is 9.21, training accuracy is 0.218750, time 55.461 samples/sec
+epoch 4, total_step 73800, total loss is 22.96 , inference loss is 13.75, weight deacy loss is 9.21, training accuracy is 0.187500, time 55.970 samples/sec
+epoch 4, total_step 73820, total loss is 21.07 , inference loss is 11.87, weight deacy loss is 9.20, training accuracy is 0.187500, time 56.264 samples/sec
+epoch 4, total_step 73840, total loss is 20.75 , inference loss is 11.55, weight deacy loss is 9.20, training accuracy is 0.187500, time 56.170 samples/sec
+epoch 4, total_step 73860, total loss is 19.08 , inference loss is 9.88, weight deacy loss is 9.20, training accuracy is 0.281250, time 56.034 samples/sec
+epoch 4, total_step 73880, total loss is 20.47 , inference loss is 11.26, weight deacy loss is 9.20, training accuracy is 0.156250, time 55.831 samples/sec
+epoch 4, total_step 73900, total loss is 22.69 , inference loss is 13.49, weight deacy loss is 9.20, training accuracy is 0.187500, time 55.751 samples/sec
+epoch 4, total_step 73920, total loss is 18.74 , inference loss is 9.55, weight deacy loss is 9.20, training accuracy is 0.218750, time 56.646 samples/sec
+epoch 4, total_step 73940, total loss is 20.21 , inference loss is 11.01, weight deacy loss is 9.20, training accuracy is 0.187500, time 54.788 samples/sec
+epoch 4, total_step 73960, total loss is 20.38 , inference loss is 11.19, weight deacy loss is 9.20, training accuracy is 0.218750, time 54.701 samples/sec
+epoch 4, total_step 73980, total loss is 21.32 , inference loss is 12.13, weight deacy loss is 9.20, training accuracy is 0.218750, time 56.239 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.63401699999994
+best_threshold_index 111 0.9842592592592593
+best_threshold_index 111 0.9844444444444445
+best_threshold_index 111 0.9848148148148148
+best_threshold_index 111 0.985
+best_threshold_index 111 0.9840740740740741
+best_threshold_index 111 0.9833333333333333
+best_threshold_index 111 0.9833333333333333
+best_threshold_index 111 0.9831481481481481
+best_threshold_index 111 0.9831481481481481
+best_threshold_index 111 0.9827777777777778
+[lfw][74000]XNorm: 23.199039
+[lfw][74000]Accuracy-Flip: 0.98383+-0.00667
+testing verification..
+(14000, 512)
+infer time 85.79491999999995
+best_threshold_index 116 0.9858730158730159
+best_threshold_index 120 0.9841269841269841
+best_threshold_index 120 0.9844444444444445
+best_threshold_index 116 0.984920634920635
+best_threshold_index 120 0.9853968253968254
+best_threshold_index 120 0.9844444444444445
+best_threshold_index 115 0.9847619047619047
+best_threshold_index 116 0.9842857142857143
+best_threshold_index 116 0.9839682539682539
+best_threshold_index 120 0.984920634920635
+[cfp_ff][74000]XNorm: 22.581232
+[cfp_ff][74000]Accuracy-Flip: 0.98314+-0.00506
+testing verification..
+(14000, 512)
+infer time 85.43090499999995
+best_threshold_index 137 0.9049206349206349
+best_threshold_index 137 0.9055555555555556
+best_threshold_index 137 0.9041269841269841
+best_threshold_index 137 0.9066666666666666
+best_threshold_index 137 0.9084126984126984
+best_threshold_index 137 0.9046031746031746
+best_threshold_index 137 0.9014285714285715
+best_threshold_index 137 0.903015873015873
+best_threshold_index 137 0.9060317460317461
+best_threshold_index 137 0.9066666666666666
+[cfp_fp][74000]XNorm: 21.729610
+[cfp_fp][74000]Accuracy-Flip: 0.90514+-0.01707
+testing verification..
+(12000, 512)
+infer time 73.80099200000004
+best_threshold_index 135 0.8933333333333333
+best_threshold_index 135 0.8896296296296297
+best_threshold_index 135 0.8911111111111111
+best_threshold_index 135 0.8905555555555555
+best_threshold_index 135 0.8985185185185185
+best_threshold_index 135 0.8896296296296297
+best_threshold_index 135 0.8916666666666667
+best_threshold_index 135 0.8992592592592593
+best_threshold_index 135 0.8942592592592593
+best_threshold_index 135 0.8920370370370371
+[agedb_30][74000]XNorm: 22.551504
+[agedb_30][74000]Accuracy-Flip: 0.89300+-0.02934
+test accuracy is: 0.9838333333333334
+epoch 4, total_step 74000, total loss is 23.71 , inference loss is 14.51, weight deacy loss is 9.20, training accuracy is 0.187500, time 53.365 samples/sec
+epoch 4, total_step 74020, total loss is 20.81 , inference loss is 11.61, weight deacy loss is 9.19, training accuracy is 0.218750, time 56.045 samples/sec
+epoch 4, total_step 74040, total loss is 21.40 , inference loss is 12.20, weight deacy loss is 9.19, training accuracy is 0.031250, time 56.280 samples/sec
+epoch 4, total_step 74060, total loss is 19.91 , inference loss is 10.72, weight deacy loss is 9.19, training accuracy is 0.312500, time 56.703 samples/sec
+epoch 4, total_step 74080, total loss is 20.98 , inference loss is 11.78, weight deacy loss is 9.19, training accuracy is 0.281250, time 56.092 samples/sec
+epoch 4, total_step 74100, total loss is 24.15 , inference loss is 14.96, weight deacy loss is 9.19, training accuracy is 0.125000, time 56.131 samples/sec
+epoch 4, total_step 74120, total loss is 19.78 , inference loss is 10.59, weight deacy loss is 9.19, training accuracy is 0.250000, time 57.185 samples/sec
+epoch 4, total_step 74140, total loss is 24.56 , inference loss is 15.37, weight deacy loss is 9.19, training accuracy is 0.218750, time 55.367 samples/sec
+epoch 4, total_step 74160, total loss is 22.93 , inference loss is 13.74, weight deacy loss is 9.19, training accuracy is 0.250000, time 56.203 samples/sec
+epoch 4, total_step 74180, total loss is 24.56 , inference loss is 15.37, weight deacy loss is 9.19, training accuracy is 0.156250, time 56.018 samples/sec
+epoch 4, total_step 74200, total loss is 20.45 , inference loss is 11.27, weight deacy loss is 9.19, training accuracy is 0.218750, time 56.274 samples/sec
+epoch 4, total_step 74220, total loss is 20.25 , inference loss is 11.06, weight deacy loss is 9.18, training accuracy is 0.187500, time 55.805 samples/sec
+epoch 4, total_step 74240, total loss is 17.70 , inference loss is 8.52, weight deacy loss is 9.18, training accuracy is 0.187500, time 55.880 samples/sec
+epoch 4, total_step 74260, total loss is 24.96 , inference loss is 15.78, weight deacy loss is 9.18, training accuracy is 0.093750, time 54.627 samples/sec
+epoch 4, total_step 74280, total loss is 23.00 , inference loss is 13.82, weight deacy loss is 9.18, training accuracy is 0.156250, time 56.216 samples/sec
+epoch 4, total_step 74300, total loss is 24.01 , inference loss is 14.83, weight deacy loss is 9.18, training accuracy is 0.156250, time 56.076 samples/sec
+epoch 4, total_step 74320, total loss is 23.95 , inference loss is 14.78, weight deacy loss is 9.18, training accuracy is 0.125000, time 55.552 samples/sec
+epoch 4, total_step 74340, total loss is 22.56 , inference loss is 13.38, weight deacy loss is 9.18, training accuracy is 0.125000, time 56.634 samples/sec
+epoch 4, total_step 74360, total loss is 22.84 , inference loss is 13.66, weight deacy loss is 9.18, training accuracy is 0.250000, time 56.298 samples/sec
+epoch 4, total_step 74380, total loss is 28.03 , inference loss is 18.86, weight deacy loss is 9.18, training accuracy is 0.062500, time 56.604 samples/sec
+epoch 4, total_step 74400, total loss is 25.12 , inference loss is 15.94, weight deacy loss is 9.18, training accuracy is 0.218750, time 54.678 samples/sec
+epoch 4, total_step 74420, total loss is 23.92 , inference loss is 14.75, weight deacy loss is 9.17, training accuracy is 0.250000, time 55.629 samples/sec
+epoch 4, total_step 74440, total loss is 22.59 , inference loss is 13.41, weight deacy loss is 9.17, training accuracy is 0.156250, time 55.714 samples/sec
+epoch 4, total_step 74460, total loss is 19.72 , inference loss is 10.54, weight deacy loss is 9.17, training accuracy is 0.062500, time 55.884 samples/sec
+epoch 4, total_step 74480, total loss is 21.28 , inference loss is 12.10, weight deacy loss is 9.17, training accuracy is 0.250000, time 57.251 samples/sec
+epoch 4, total_step 74500, total loss is 21.24 , inference loss is 12.07, weight deacy loss is 9.17, training accuracy is 0.125000, time 56.038 samples/sec
+epoch 4, total_step 74520, total loss is 18.16 , inference loss is 8.99, weight deacy loss is 9.17, training accuracy is 0.375000, time 55.470 samples/sec
+epoch 4, total_step 74540, total loss is 23.36 , inference loss is 14.19, weight deacy loss is 9.17, training accuracy is 0.218750, time 55.856 samples/sec
+epoch 4, total_step 74560, total loss is 20.17 , inference loss is 11.00, weight deacy loss is 9.17, training accuracy is 0.125000, time 56.519 samples/sec
+epoch 4, total_step 74580, total loss is 20.74 , inference loss is 11.57, weight deacy loss is 9.17, training accuracy is 0.281250, time 53.473 samples/sec
+epoch 4, total_step 74600, total loss is 18.82 , inference loss is 9.65, weight deacy loss is 9.17, training accuracy is 0.281250, time 55.904 samples/sec
+epoch 4, total_step 74620, total loss is 19.14 , inference loss is 9.98, weight deacy loss is 9.16, training accuracy is 0.187500, time 56.140 samples/sec
+epoch 4, total_step 74640, total loss is 24.91 , inference loss is 15.75, weight deacy loss is 9.16, training accuracy is 0.218750, time 56.183 samples/sec
+epoch 4, total_step 74660, total loss is 20.83 , inference loss is 11.67, weight deacy loss is 9.16, training accuracy is 0.281250, time 56.075 samples/sec
+epoch 4, total_step 74680, total loss is 21.24 , inference loss is 12.08, weight deacy loss is 9.16, training accuracy is 0.218750, time 56.562 samples/sec
+epoch 4, total_step 74700, total loss is 24.13 , inference loss is 14.97, weight deacy loss is 9.16, training accuracy is 0.218750, time 54.590 samples/sec
+epoch 4, total_step 74720, total loss is 27.04 , inference loss is 17.88, weight deacy loss is 9.16, training accuracy is 0.125000, time 55.850 samples/sec
+epoch 4, total_step 74740, total loss is 23.47 , inference loss is 14.31, weight deacy loss is 9.16, training accuracy is 0.156250, time 56.095 samples/sec
+epoch 4, total_step 74760, total loss is 19.36 , inference loss is 10.20, weight deacy loss is 9.16, training accuracy is 0.187500, time 55.947 samples/sec
+epoch 4, total_step 74780, total loss is 20.26 , inference loss is 11.11, weight deacy loss is 9.16, training accuracy is 0.250000, time 56.948 samples/sec
+epoch 4, total_step 74800, total loss is 23.63 , inference loss is 14.48, weight deacy loss is 9.16, training accuracy is 0.187500, time 56.349 samples/sec
+epoch 4, total_step 74820, total loss is 24.50 , inference loss is 15.35, weight deacy loss is 9.15, training accuracy is 0.187500, time 57.042 samples/sec
+epoch 4, total_step 74840, total loss is 20.43 , inference loss is 11.27, weight deacy loss is 9.15, training accuracy is 0.187500, time 56.624 samples/sec
+epoch 4, total_step 74860, total loss is 18.19 , inference loss is 9.03, weight deacy loss is 9.15, training accuracy is 0.218750, time 56.707 samples/sec
+epoch 4, total_step 74880, total loss is 22.96 , inference loss is 13.81, weight deacy loss is 9.15, training accuracy is 0.187500, time 56.856 samples/sec
+epoch 4, total_step 74900, total loss is 20.77 , inference loss is 11.62, weight deacy loss is 9.15, training accuracy is 0.250000, time 56.234 samples/sec
+epoch 4, total_step 74920, total loss is 19.01 , inference loss is 9.87, weight deacy loss is 9.15, training accuracy is 0.156250, time 55.719 samples/sec
+epoch 4, total_step 74940, total loss is 28.26 , inference loss is 19.12, weight deacy loss is 9.15, training accuracy is 0.125000, time 56.457 samples/sec
+epoch 4, total_step 74960, total loss is 26.94 , inference loss is 17.79, weight deacy loss is 9.15, training accuracy is 0.062500, time 57.042 samples/sec
+epoch 4, total_step 74980, total loss is 29.27 , inference loss is 20.12, weight deacy loss is 9.15, training accuracy is 0.125000, time 55.872 samples/sec
+epoch 4, total_step 75000, total loss is 22.90 , inference loss is 13.76, weight deacy loss is 9.15, training accuracy is 0.187500, time 53.613 samples/sec
+epoch 4, total_step 75020, total loss is 22.37 , inference loss is 13.22, weight deacy loss is 9.14, training accuracy is 0.281250, time 54.914 samples/sec
+epoch 4, total_step 75040, total loss is 21.52 , inference loss is 12.38, weight deacy loss is 9.14, training accuracy is 0.156250, time 57.526 samples/sec
+epoch 4, total_step 75060, total loss is 20.66 , inference loss is 11.51, weight deacy loss is 9.14, training accuracy is 0.250000, time 55.627 samples/sec
+epoch 4, total_step 75080, total loss is 21.65 , inference loss is 12.51, weight deacy loss is 9.14, training accuracy is 0.156250, time 56.585 samples/sec
+epoch 4, total_step 75100, total loss is 20.97 , inference loss is 11.83, weight deacy loss is 9.14, training accuracy is 0.156250, time 56.286 samples/sec
+epoch 4, total_step 75120, total loss is 24.84 , inference loss is 15.71, weight deacy loss is 9.14, training accuracy is 0.218750, time 56.504 samples/sec
+epoch 4, total_step 75140, total loss is 23.96 , inference loss is 14.82, weight deacy loss is 9.14, training accuracy is 0.093750, time 55.898 samples/sec
+epoch 4, total_step 75160, total loss is 22.47 , inference loss is 13.33, weight deacy loss is 9.14, training accuracy is 0.281250, time 54.518 samples/sec
+epoch 4, total_step 75180, total loss is 20.41 , inference loss is 11.28, weight deacy loss is 9.14, training accuracy is 0.218750, time 57.343 samples/sec
+epoch 4, total_step 75200, total loss is 22.59 , inference loss is 13.45, weight deacy loss is 9.14, training accuracy is 0.125000, time 56.303 samples/sec
+epoch 4, total_step 75220, total loss is 22.82 , inference loss is 13.68, weight deacy loss is 9.13, training accuracy is 0.187500, time 56.396 samples/sec
+epoch 4, total_step 75240, total loss is 25.57 , inference loss is 16.44, weight deacy loss is 9.13, training accuracy is 0.125000, time 56.574 samples/sec
+epoch 4, total_step 75260, total loss is 20.33 , inference loss is 11.20, weight deacy loss is 9.13, training accuracy is 0.250000, time 56.629 samples/sec
+epoch 4, total_step 75280, total loss is 21.52 , inference loss is 12.39, weight deacy loss is 9.13, training accuracy is 0.187500, time 55.910 samples/sec
+epoch 4, total_step 75300, total loss is 23.24 , inference loss is 14.11, weight deacy loss is 9.13, training accuracy is 0.156250, time 56.425 samples/sec
+epoch 4, total_step 75320, total loss is 22.76 , inference loss is 13.63, weight deacy loss is 9.13, training accuracy is 0.093750, time 55.451 samples/sec
+epoch 4, total_step 75340, total loss is 23.11 , inference loss is 13.99, weight deacy loss is 9.13, training accuracy is 0.187500, time 56.513 samples/sec
+epoch 4, total_step 75360, total loss is 22.16 , inference loss is 13.03, weight deacy loss is 9.13, training accuracy is 0.125000, time 56.323 samples/sec
+epoch 4, total_step 75380, total loss is 20.82 , inference loss is 11.69, weight deacy loss is 9.13, training accuracy is 0.125000, time 55.961 samples/sec
+epoch 4, total_step 75400, total loss is 20.42 , inference loss is 11.30, weight deacy loss is 9.13, training accuracy is 0.187500, time 56.535 samples/sec
+epoch 4, total_step 75420, total loss is 25.05 , inference loss is 15.93, weight deacy loss is 9.12, training accuracy is 0.218750, time 56.286 samples/sec
+epoch 4, total_step 75440, total loss is 21.93 , inference loss is 12.80, weight deacy loss is 9.12, training accuracy is 0.062500, time 55.565 samples/sec
+epoch 4, total_step 75460, total loss is 23.88 , inference loss is 14.75, weight deacy loss is 9.12, training accuracy is 0.187500, time 57.405 samples/sec
+epoch 4, total_step 75480, total loss is 26.25 , inference loss is 17.13, weight deacy loss is 9.12, training accuracy is 0.156250, time 56.808 samples/sec
+epoch 4, total_step 75500, total loss is 24.01 , inference loss is 14.89, weight deacy loss is 9.12, training accuracy is 0.093750, time 56.912 samples/sec
+epoch 4, total_step 75520, total loss is 25.02 , inference loss is 15.90, weight deacy loss is 9.12, training accuracy is 0.156250, time 56.516 samples/sec
+epoch 4, total_step 75540, total loss is 20.25 , inference loss is 11.13, weight deacy loss is 9.12, training accuracy is 0.156250, time 56.194 samples/sec
+epoch 4, total_step 75560, total loss is 23.02 , inference loss is 13.90, weight deacy loss is 9.12, training accuracy is 0.125000, time 55.571 samples/sec
+epoch 4, total_step 75580, total loss is 19.58 , inference loss is 10.47, weight deacy loss is 9.12, training accuracy is 0.312500, time 56.400 samples/sec
+epoch 4, total_step 75600, total loss is 23.76 , inference loss is 14.65, weight deacy loss is 9.12, training accuracy is 0.187500, time 55.854 samples/sec
+epoch 4, total_step 75620, total loss is 21.60 , inference loss is 12.49, weight deacy loss is 9.11, training accuracy is 0.187500, time 57.233 samples/sec
+epoch 4, total_step 75640, total loss is 22.63 , inference loss is 13.52, weight deacy loss is 9.11, training accuracy is 0.156250, time 56.524 samples/sec
+epoch 4, total_step 75660, total loss is 19.63 , inference loss is 10.52, weight deacy loss is 9.11, training accuracy is 0.187500, time 56.070 samples/sec
+epoch 4, total_step 75680, total loss is 21.75 , inference loss is 12.63, weight deacy loss is 9.11, training accuracy is 0.250000, time 56.736 samples/sec
+epoch 4, total_step 75700, total loss is 21.65 , inference loss is 12.54, weight deacy loss is 9.11, training accuracy is 0.093750, time 57.127 samples/sec
+epoch 4, total_step 75720, total loss is 21.57 , inference loss is 12.46, weight deacy loss is 9.11, training accuracy is 0.125000, time 56.265 samples/sec
+epoch 4, total_step 75740, total loss is 22.61 , inference loss is 13.50, weight deacy loss is 9.11, training accuracy is 0.093750, time 55.954 samples/sec
+epoch 4, total_step 75760, total loss is 27.12 , inference loss is 18.02, weight deacy loss is 9.11, training accuracy is 0.125000, time 56.961 samples/sec
+epoch 4, total_step 75780, total loss is 18.88 , inference loss is 9.77, weight deacy loss is 9.11, training accuracy is 0.312500, time 56.800 samples/sec
+epoch 4, total_step 75800, total loss is 18.37 , inference loss is 9.27, weight deacy loss is 9.11, training accuracy is 0.375000, time 55.910 samples/sec
+epoch 4, total_step 75820, total loss is 22.71 , inference loss is 13.60, weight deacy loss is 9.10, training accuracy is 0.187500, time 56.352 samples/sec
+epoch 4, total_step 75840, total loss is 19.28 , inference loss is 10.17, weight deacy loss is 9.10, training accuracy is 0.250000, time 55.827 samples/sec
+epoch 4, total_step 75860, total loss is 27.09 , inference loss is 17.98, weight deacy loss is 9.10, training accuracy is 0.125000, time 56.966 samples/sec
+epoch 4, total_step 75880, total loss is 20.54 , inference loss is 11.44, weight deacy loss is 9.10, training accuracy is 0.250000, time 56.265 samples/sec
+epoch 4, total_step 75900, total loss is 21.67 , inference loss is 12.57, weight deacy loss is 9.10, training accuracy is 0.156250, time 56.406 samples/sec
+epoch 4, total_step 75920, total loss is 22.61 , inference loss is 13.51, weight deacy loss is 9.10, training accuracy is 0.156250, time 57.075 samples/sec
+epoch 4, total_step 75940, total loss is 23.61 , inference loss is 14.51, weight deacy loss is 9.10, training accuracy is 0.125000, time 56.596 samples/sec
+epoch 4, total_step 75960, total loss is 19.18 , inference loss is 10.09, weight deacy loss is 9.10, training accuracy is 0.187500, time 55.812 samples/sec
+epoch 4, total_step 75980, total loss is 20.68 , inference loss is 11.58, weight deacy loss is 9.10, training accuracy is 0.156250, time 55.383 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.562921
+best_threshold_index 105 0.9835185185185186
+best_threshold_index 106 0.982037037037037
+best_threshold_index 106 0.9827777777777778
+best_threshold_index 105 0.9848148148148148
+best_threshold_index 103 0.9829629629629629
+best_threshold_index 105 0.9812962962962963
+best_threshold_index 108 0.9825925925925926
+best_threshold_index 105 0.9814814814814815
+best_threshold_index 105 0.980925925925926
+best_threshold_index 105 0.9814814814814815
+[lfw][76000]XNorm: 23.064518
+[lfw][76000]Accuracy-Flip: 0.98133+-0.01043
+testing verification..
+(14000, 512)
+infer time 85.82724200000011
+best_threshold_index 115 0.9838095238095238
+best_threshold_index 113 0.9836507936507937
+best_threshold_index 118 0.9834920634920635
+best_threshold_index 113 0.9828571428571429
+best_threshold_index 117 0.9834920634920635
+best_threshold_index 118 0.983015873015873
+best_threshold_index 117 0.9841269841269841
+best_threshold_index 117 0.9826984126984127
+best_threshold_index 111 0.983015873015873
+best_threshold_index 117 0.9838095238095238
+[cfp_ff][76000]XNorm: 22.718067
+[cfp_ff][76000]Accuracy-Flip: 0.98129+-0.00396
+testing verification..
+(14000, 512)
+infer time 85.61406200000002
+best_threshold_index 132 0.8990476190476191
+best_threshold_index 132 0.8992063492063492
+best_threshold_index 136 0.8979365079365079
+best_threshold_index 132 0.9019047619047619
+best_threshold_index 132 0.903015873015873
+best_threshold_index 132 0.8957142857142857
+best_threshold_index 132 0.8966666666666666
+best_threshold_index 132 0.8985714285714286
+best_threshold_index 133 0.8987301587301587
+best_threshold_index 136 0.9001587301587302
+[cfp_fp][76000]XNorm: 21.879688
+[cfp_fp][76000]Accuracy-Flip: 0.89743+-0.01877
+testing verification..
+(12000, 512)
+infer time 73.73037099999993
+best_threshold_index 135 0.8955555555555555
+best_threshold_index 133 0.8927777777777778
+best_threshold_index 134 0.8968518518518519
+best_threshold_index 135 0.8961111111111111
+best_threshold_index 135 0.9011111111111111
+best_threshold_index 135 0.8964814814814814
+best_threshold_index 134 0.8957407407407407
+best_threshold_index 135 0.9022222222222223
+best_threshold_index 135 0.8996296296296297
+best_threshold_index 133 0.8974074074074074
+[agedb_30][76000]XNorm: 23.105808
+[agedb_30][76000]Accuracy-Flip: 0.89483+-0.02282
+test accuracy is: 0.9813333333333333
+epoch 4, total_step 76000, total loss is 21.66 , inference loss is 12.57, weight deacy loss is 9.10, training accuracy is 0.156250, time 53.067 samples/sec
+epoch 4, total_step 76020, total loss is 31.04 , inference loss is 21.95, weight deacy loss is 9.09, training accuracy is 0.031250, time 56.104 samples/sec
+epoch 4, total_step 76040, total loss is 20.47 , inference loss is 11.38, weight deacy loss is 9.09, training accuracy is 0.218750, time 56.102 samples/sec
+epoch 4, total_step 76060, total loss is 20.50 , inference loss is 11.41, weight deacy loss is 9.09, training accuracy is 0.218750, time 56.314 samples/sec
+epoch 4, total_step 76080, total loss is 24.56 , inference loss is 15.47, weight deacy loss is 9.09, training accuracy is 0.125000, time 55.562 samples/sec2022-12-01 14:18:50.602643: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:111] Filling up shuffle buffer (this may take a while): 8478 of 10000
+2022-12-01 14:18:52.267078: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:162] Shuffle buffer filled.
+
+epoch 4, total_step 76100, total loss is 23.76 , inference loss is 14.67, weight deacy loss is 9.09, training accuracy is 0.156250, time 56.413 samples/sec
+epoch 4, total_step 76120, total loss is 24.79 , inference loss is 15.70, weight deacy loss is 9.09, training accuracy is 0.031250, time 56.732 samples/sec
+epoch 4, total_step 76140, total loss is 25.41 , inference loss is 16.32, weight deacy loss is 9.09, training accuracy is 0.187500, time 56.629 samples/sec
+epoch 4, total_step 76160, total loss is 23.09 , inference loss is 14.00, weight deacy loss is 9.09, training accuracy is 0.250000, time 56.178 samples/sec
+epoch 4, total_step 76180, total loss is 19.71 , inference loss is 10.62, weight deacy loss is 9.09, training accuracy is 0.156250, time 56.561 samples/sec
+epoch 4, total_step 76200, total loss is 18.31 , inference loss is 9.22, weight deacy loss is 9.09, training accuracy is 0.218750, time 56.225 samples/sec
+epoch 4, total_step 76220, total loss is 22.83 , inference loss is 13.74, weight deacy loss is 9.09, training accuracy is 0.187500, time 56.550 samples/sec
+epoch 4, total_step 76240, total loss is 23.28 , inference loss is 14.19, weight deacy loss is 9.08, training accuracy is 0.125000, time 54.909 samples/sec
+epoch 4, total_step 76260, total loss is 23.23 , inference loss is 14.15, weight deacy loss is 9.08, training accuracy is 0.218750, time 56.112 samples/sec
+epoch 4, total_step 76280, total loss is 23.09 , inference loss is 14.01, weight deacy loss is 9.08, training accuracy is 0.125000, time 56.867 samples/sec
+epoch 4, total_step 76300, total loss is 20.29 , inference loss is 11.21, weight deacy loss is 9.08, training accuracy is 0.156250, time 56.867 samples/sec
+epoch 4, total_step 76320, total loss is 19.48 , inference loss is 10.40, weight deacy loss is 9.08, training accuracy is 0.187500, time 56.033 samples/sec
+epoch 4, total_step 76340, total loss is 24.26 , inference loss is 15.18, weight deacy loss is 9.08, training accuracy is 0.156250, time 56.507 samples/sec
+epoch 4, total_step 76360, total loss is 25.47 , inference loss is 16.39, weight deacy loss is 9.08, training accuracy is 0.093750, time 57.175 samples/sec
+epoch 4, total_step 76380, total loss is 23.52 , inference loss is 14.44, weight deacy loss is 9.08, training accuracy is 0.125000, time 55.505 samples/sec
+epoch 4, total_step 76400, total loss is 20.75 , inference loss is 11.67, weight deacy loss is 9.08, training accuracy is 0.125000, time 55.705 samples/sec
+epoch 4, total_step 76420, total loss is 23.60 , inference loss is 14.52, weight deacy loss is 9.08, training accuracy is 0.218750, time 56.003 samples/sec
+epoch 4, total_step 76440, total loss is 24.83 , inference loss is 15.75, weight deacy loss is 9.07, training accuracy is 0.187500, time 54.719 samples/sec
+epoch 4, total_step 76460, total loss is 22.38 , inference loss is 13.30, weight deacy loss is 9.07, training accuracy is 0.218750, time 55.892 samples/sec
+epoch 4, total_step 76480, total loss is 23.44 , inference loss is 14.37, weight deacy loss is 9.07, training accuracy is 0.187500, time 56.609 samples/sec
+epoch 4, total_step 76500, total loss is 21.06 , inference loss is 11.99, weight deacy loss is 9.07, training accuracy is 0.093750, time 55.708 samples/sec
+epoch 4, total_step 76520, total loss is 23.05 , inference loss is 13.98, weight deacy loss is 9.07, training accuracy is 0.062500, time 56.532 samples/sec
+epoch 4, total_step 76540, total loss is 20.12 , inference loss is 11.05, weight deacy loss is 9.07, training accuracy is 0.187500, time 56.281 samples/sec
+epoch 4, total_step 76560, total loss is 23.74 , inference loss is 14.67, weight deacy loss is 9.07, training accuracy is 0.156250, time 54.261 samples/sec
+epoch 4, total_step 76580, total loss is 16.19 , inference loss is 7.12, weight deacy loss is 9.07, training accuracy is 0.343750, time 55.249 samples/sec
+epoch 4, total_step 76600, total loss is 22.04 , inference loss is 12.98, weight deacy loss is 9.07, training accuracy is 0.125000, time 55.620 samples/sec
+epoch 4, total_step 76620, total loss is 19.12 , inference loss is 10.05, weight deacy loss is 9.07, training accuracy is 0.312500, time 56.406 samples/sec
+epoch 4, total_step 76640, total loss is 20.43 , inference loss is 11.36, weight deacy loss is 9.07, training accuracy is 0.156250, time 56.783 samples/sec
+
+
+
+
+End of epoch 4
+epoch 5, total_step 76660, total loss is 34.39 , inference loss is 25.32, weight deacy loss is 9.06, training accuracy is 0.000000, time 53.949 samples/sec
+epoch 5, total_step 76680, total loss is 20.82 , inference loss is 11.76, weight deacy loss is 9.06, training accuracy is 0.281250, time 55.698 samples/sec
+epoch 5, total_step 76700, total loss is 15.88 , inference loss is 6.82, weight deacy loss is 9.06, training accuracy is 0.468750, time 55.302 samples/sec
+epoch 5, total_step 76720, total loss is 20.03 , inference loss is 10.97, weight deacy loss is 9.06, training accuracy is 0.375000, time 56.257 samples/sec
+epoch 5, total_step 76740, total loss is 18.03 , inference loss is 8.97, weight deacy loss is 9.06, training accuracy is 0.437500, time 55.692 samples/sec
+epoch 5, total_step 76760, total loss is 20.25 , inference loss is 11.19, weight deacy loss is 9.06, training accuracy is 0.343750, time 56.207 samples/sec
+epoch 5, total_step 76780, total loss is 17.26 , inference loss is 8.20, weight deacy loss is 9.06, training accuracy is 0.468750, time 55.861 samples/sec
+epoch 5, total_step 76800, total loss is 21.29 , inference loss is 12.23, weight deacy loss is 9.06, training accuracy is 0.437500, time 56.390 samples/sec
+epoch 5, total_step 76820, total loss is 22.21 , inference loss is 13.15, weight deacy loss is 9.06, training accuracy is 0.343750, time 56.225 samples/sec
+epoch 5, total_step 76840, total loss is 16.84 , inference loss is 7.79, weight deacy loss is 9.06, training accuracy is 0.531250, time 56.173 samples/sec
+epoch 5, total_step 76860, total loss is 17.10 , inference loss is 8.05, weight deacy loss is 9.05, training accuracy is 0.468750, time 56.921 samples/sec
+epoch 5, total_step 76880, total loss is 15.83 , inference loss is 6.78, weight deacy loss is 9.05, training accuracy is 0.531250, time 56.080 samples/sec
+epoch 5, total_step 76900, total loss is 18.68 , inference loss is 9.63, weight deacy loss is 9.05, training accuracy is 0.500000, time 56.786 samples/sec
+epoch 5, total_step 76920, total loss is 16.83 , inference loss is 7.78, weight deacy loss is 9.05, training accuracy is 0.437500, time 55.994 samples/sec
+epoch 5, total_step 76940, total loss is 16.84 , inference loss is 7.79, weight deacy loss is 9.05, training accuracy is 0.500000, time 56.765 samples/sec
+epoch 5, total_step 76960, total loss is 15.94 , inference loss is 6.89, weight deacy loss is 9.05, training accuracy is 0.531250, time 56.357 samples/sec
+epoch 5, total_step 76980, total loss is 15.40 , inference loss is 6.35, weight deacy loss is 9.05, training accuracy is 0.593750, time 52.496 samples/sec
+epoch 5, total_step 77000, total loss is 17.33 , inference loss is 8.29, weight deacy loss is 9.05, training accuracy is 0.562500, time 56.453 samples/sec
+epoch 5, total_step 77020, total loss is 18.80 , inference loss is 9.76, weight deacy loss is 9.05, training accuracy is 0.500000, time 56.019 samples/sec
+epoch 5, total_step 77040, total loss is 21.64 , inference loss is 12.60, weight deacy loss is 9.04, training accuracy is 0.343750, time 53.929 samples/sec
+epoch 5, total_step 77060, total loss is 18.85 , inference loss is 9.80, weight deacy loss is 9.04, training accuracy is 0.281250, time 56.171 samples/sec
+epoch 5, total_step 77080, total loss is 17.28 , inference loss is 8.24, weight deacy loss is 9.04, training accuracy is 0.500000, time 55.916 samples/sec
+epoch 5, total_step 77100, total loss is 18.93 , inference loss is 9.89, weight deacy loss is 9.04, training accuracy is 0.343750, time 55.895 samples/sec
+epoch 5, total_step 77120, total loss is 16.88 , inference loss is 7.84, weight deacy loss is 9.04, training accuracy is 0.312500, time 55.692 samples/sec
+epoch 5, total_step 77140, total loss is 14.97 , inference loss is 5.93, weight deacy loss is 9.04, training accuracy is 0.500000, time 56.396 samples/sec
+epoch 5, total_step 77160, total loss is 13.56 , inference loss is 4.52, weight deacy loss is 9.04, training accuracy is 0.687500, time 55.382 samples/sec
+epoch 5, total_step 77180, total loss is 15.87 , inference loss is 6.83, weight deacy loss is 9.04, training accuracy is 0.562500, time 56.557 samples/sec
+epoch 5, total_step 77200, total loss is 17.21 , inference loss is 8.17, weight deacy loss is 9.04, training accuracy is 0.437500, time 56.365 samples/sec
+epoch 5, total_step 77220, total loss is 23.26 , inference loss is 14.22, weight deacy loss is 9.04, training accuracy is 0.218750, time 55.541 samples/sec
+epoch 5, total_step 77240, total loss is 21.80 , inference loss is 12.76, weight deacy loss is 9.04, training accuracy is 0.312500, time 56.363 samples/sec
+epoch 5, total_step 77260, total loss is 20.26 , inference loss is 11.22, weight deacy loss is 9.03, training accuracy is 0.375000, time 56.617 samples/sec
+epoch 5, total_step 77280, total loss is 19.95 , inference loss is 10.92, weight deacy loss is 9.03, training accuracy is 0.406250, time 53.326 samples/sec
+epoch 5, total_step 77300, total loss is 18.34 , inference loss is 9.30, weight deacy loss is 9.03, training accuracy is 0.312500, time 56.916 samples/sec
+epoch 5, total_step 77320, total loss is 22.95 , inference loss is 13.92, weight deacy loss is 9.03, training accuracy is 0.375000, time 55.497 samples/sec
+epoch 5, total_step 77340, total loss is 21.65 , inference loss is 12.62, weight deacy loss is 9.03, training accuracy is 0.281250, time 56.719 samples/sec
+epoch 5, total_step 77360, total loss is 20.65 , inference loss is 11.62, weight deacy loss is 9.03, training accuracy is 0.375000, time 55.769 samples/sec
+epoch 5, total_step 77380, total loss is 16.79 , inference loss is 7.76, weight deacy loss is 9.03, training accuracy is 0.437500, time 56.403 samples/sec
+epoch 5, total_step 77400, total loss is 21.48 , inference loss is 12.45, weight deacy loss is 9.03, training accuracy is 0.218750, time 56.500 samples/sec
+epoch 5, total_step 77420, total loss is 17.75 , inference loss is 8.72, weight deacy loss is 9.03, training accuracy is 0.437500, time 56.286 samples/sec
+epoch 5, total_step 77440, total loss is 15.04 , inference loss is 6.01, weight deacy loss is 9.03, training accuracy is 0.468750, time 55.676 samples/sec
+epoch 5, total_step 77460, total loss is 19.63 , inference loss is 10.60, weight deacy loss is 9.03, training accuracy is 0.343750, time 55.602 samples/sec
+epoch 5, total_step 77480, total loss is 20.40 , inference loss is 11.37, weight deacy loss is 9.02, training accuracy is 0.187500, time 55.617 samples/sec
+epoch 5, total_step 77500, total loss is 16.69 , inference loss is 7.66, weight deacy loss is 9.02, training accuracy is 0.375000, time 56.004 samples/sec
+epoch 5, total_step 77520, total loss is 22.08 , inference loss is 13.06, weight deacy loss is 9.02, training accuracy is 0.218750, time 51.941 samples/sec
+epoch 5, total_step 77540, total loss is 21.05 , inference loss is 12.03, weight deacy loss is 9.02, training accuracy is 0.250000, time 53.652 samples/sec
+epoch 5, total_step 77560, total loss is 22.45 , inference loss is 13.42, weight deacy loss is 9.02, training accuracy is 0.312500, time 55.469 samples/sec
+epoch 5, total_step 77580, total loss is 16.24 , inference loss is 7.22, weight deacy loss is 9.02, training accuracy is 0.500000, time 54.851 samples/sec
+epoch 5, total_step 77600, total loss is 14.66 , inference loss is 5.64, weight deacy loss is 9.02, training accuracy is 0.625000, time 54.861 samples/sec
+epoch 5, total_step 77620, total loss is 16.80 , inference loss is 7.78, weight deacy loss is 9.02, training accuracy is 0.312500, time 54.995 samples/sec
+epoch 5, total_step 77640, total loss is 19.26 , inference loss is 10.24, weight deacy loss is 9.02, training accuracy is 0.406250, time 54.269 samples/sec
+epoch 5, total_step 77660, total loss is 17.88 , inference loss is 8.86, weight deacy loss is 9.02, training accuracy is 0.437500, time 54.691 samples/sec
+epoch 5, total_step 77680, total loss is 19.36 , inference loss is 10.35, weight deacy loss is 9.02, training accuracy is 0.312500, time 54.575 samples/sec
+epoch 5, total_step 77700, total loss is 17.70 , inference loss is 8.69, weight deacy loss is 9.02, training accuracy is 0.312500, time 56.371 samples/sec
+epoch 5, total_step 77720, total loss is 18.92 , inference loss is 9.90, weight deacy loss is 9.01, training accuracy is 0.375000, time 56.374 samples/sec
+epoch 5, total_step 77740, total loss is 21.87 , inference loss is 12.85, weight deacy loss is 9.01, training accuracy is 0.312500, time 54.327 samples/sec
+epoch 5, total_step 77760, total loss is 20.14 , inference loss is 11.13, weight deacy loss is 9.01, training accuracy is 0.406250, time 56.925 samples/sec
+epoch 5, total_step 77780, total loss is 19.12 , inference loss is 10.10, weight deacy loss is 9.01, training accuracy is 0.437500, time 57.157 samples/sec
+epoch 5, total_step 77800, total loss is 18.07 , inference loss is 9.06, weight deacy loss is 9.01, training accuracy is 0.437500, time 56.478 samples/sec
+epoch 5, total_step 77820, total loss is 20.11 , inference loss is 11.10, weight deacy loss is 9.01, training accuracy is 0.312500, time 55.237 samples/sec
+epoch 5, total_step 77840, total loss is 17.16 , inference loss is 8.16, weight deacy loss is 9.01, training accuracy is 0.531250, time 56.519 samples/sec
+epoch 5, total_step 77860, total loss is 24.44 , inference loss is 15.44, weight deacy loss is 9.01, training accuracy is 0.281250, time 55.085 samples/sec
+epoch 5, total_step 77880, total loss is 17.45 , inference loss is 8.45, weight deacy loss is 9.01, training accuracy is 0.250000, time 55.761 samples/sec
+epoch 5, total_step 77900, total loss is 19.60 , inference loss is 10.59, weight deacy loss is 9.01, training accuracy is 0.312500, time 55.751 samples/sec
+epoch 5, total_step 77920, total loss is 16.27 , inference loss is 7.26, weight deacy loss is 9.01, training accuracy is 0.406250, time 55.008 samples/sec
+epoch 5, total_step 77940, total loss is 20.10 , inference loss is 11.09, weight deacy loss is 9.00, training accuracy is 0.375000, time 55.219 samples/sec
+epoch 5, total_step 77960, total loss is 19.59 , inference loss is 10.59, weight deacy loss is 9.00, training accuracy is 0.406250, time 56.235 samples/sec
+epoch 5, total_step 77980, total loss is 19.18 , inference loss is 10.18, weight deacy loss is 9.00, training accuracy is 0.375000, time 51.644 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.00990300000007
+best_threshold_index 120 0.9861111111111112
+best_threshold_index 118 0.9853703703703703
+best_threshold_index 120 0.9855555555555555
+best_threshold_index 120 0.9857407407407407
+best_threshold_index 120 0.9859259259259259
+best_threshold_index 120 0.9855555555555555
+best_threshold_index 120 0.9857407407407407
+best_threshold_index 120 0.985
+best_threshold_index 120 0.9840740740740741
+best_threshold_index 120 0.9844444444444445
+[lfw][78000]XNorm: 22.036339
+[lfw][78000]Accuracy-Flip: 0.98483+-0.00570
+testing verification..
+(14000, 512)
+infer time 88.11309700000005
+best_threshold_index 117 0.9806349206349206
+best_threshold_index 117 0.9807936507936508
+best_threshold_index 117 0.9809523809523809
+best_threshold_index 117 0.98
+best_threshold_index 117 0.9801587301587301
+best_threshold_index 117 0.98
+best_threshold_index 115 0.9806349206349206
+best_threshold_index 117 0.98
+best_threshold_index 115 0.9798412698412698
+best_threshold_index 112 0.9815873015873016
+[cfp_ff][78000]XNorm: 21.544887
+[cfp_ff][78000]Accuracy-Flip: 0.97914+-0.00636
+testing verification..
+(14000, 512)
+infer time 85.56519999999995
+best_threshold_index 140 0.8998412698412699
+best_threshold_index 140 0.8980952380952381
+best_threshold_index 140 0.8966666666666666
+best_threshold_index 133 0.9
+best_threshold_index 138 0.9007936507936508
+best_threshold_index 138 0.896984126984127
+best_threshold_index 139 0.8973015873015873
+best_threshold_index 140 0.8984126984126984
+best_threshold_index 140 0.8982539682539683
+best_threshold_index 139 0.8990476190476191
+[cfp_fp][78000]XNorm: 21.668011
+[cfp_fp][78000]Accuracy-Flip: 0.89529+-0.01366
+testing verification..
+(12000, 512)
+infer time 73.11135199999987
+best_threshold_index 142 0.8938888888888888
+best_threshold_index 142 0.8944444444444445
+best_threshold_index 142 0.8962962962962963
+best_threshold_index 142 0.8966666666666666
+best_threshold_index 142 0.9018518518518519
+best_threshold_index 142 0.8964814814814814
+best_threshold_index 142 0.8957407407407407
+best_threshold_index 142 0.9011111111111111
+best_threshold_index 143 0.8992592592592593
+best_threshold_index 142 0.8994444444444445
+[agedb_30][78000]XNorm: 22.007006
+[agedb_30][78000]Accuracy-Flip: 0.89683+-0.02376
+test accuracy is: 0.9848333333333334
+epoch 5, total_step 78000, total loss is 15.08 , inference loss is 6.08, weight deacy loss is 9.00, training accuracy is 0.500000, time 54.318 samples/sec
+epoch 5, total_step 78020, total loss is 14.73 , inference loss is 5.73, weight deacy loss is 9.00, training accuracy is 0.562500, time 56.432 samples/sec
+epoch 5, total_step 78040, total loss is 17.80 , inference loss is 8.80, weight deacy loss is 9.00, training accuracy is 0.500000, time 56.658 samples/sec
+epoch 5, total_step 78060, total loss is 18.89 , inference loss is 9.89, weight deacy loss is 9.00, training accuracy is 0.531250, time 56.544 samples/sec
+epoch 5, total_step 78080, total loss is 22.58 , inference loss is 13.58, weight deacy loss is 9.00, training accuracy is 0.250000, time 56.426 samples/sec
+epoch 5, total_step 78100, total loss is 17.01 , inference loss is 8.01, weight deacy loss is 9.00, training accuracy is 0.312500, time 56.501 samples/sec
+epoch 5, total_step 78120, total loss is 21.93 , inference loss is 12.94, weight deacy loss is 9.00, training accuracy is 0.250000, time 57.042 samples/sec
+epoch 5, total_step 78140, total loss is 20.67 , inference loss is 11.67, weight deacy loss is 9.00, training accuracy is 0.343750, time 57.125 samples/sec
+epoch 5, total_step 78160, total loss is 18.16 , inference loss is 9.17, weight deacy loss is 8.99, training accuracy is 0.343750, time 56.496 samples/sec
+epoch 5, total_step 78180, total loss is 17.29 , inference loss is 8.30, weight deacy loss is 8.99, training accuracy is 0.312500, time 57.433 samples/sec
+epoch 5, total_step 78200, total loss is 19.90 , inference loss is 10.90, weight deacy loss is 8.99, training accuracy is 0.437500, time 56.746 samples/sec
+epoch 5, total_step 78220, total loss is 20.03 , inference loss is 11.04, weight deacy loss is 8.99, training accuracy is 0.375000, time 56.246 samples/sec
+epoch 5, total_step 78240, total loss is 21.84 , inference loss is 12.85, weight deacy loss is 8.99, training accuracy is 0.125000, time 56.764 samples/sec
+epoch 5, total_step 78260, total loss is 20.61 , inference loss is 11.62, weight deacy loss is 8.99, training accuracy is 0.312500, time 56.856 samples/sec
+epoch 5, total_step 78280, total loss is 20.52 , inference loss is 11.53, weight deacy loss is 8.99, training accuracy is 0.218750, time 56.561 samples/sec
+epoch 5, total_step 78300, total loss is 17.69 , inference loss is 8.70, weight deacy loss is 8.99, training accuracy is 0.437500, time 57.101 samples/sec
+epoch 5, total_step 78320, total loss is 21.41 , inference loss is 12.42, weight deacy loss is 8.99, training accuracy is 0.218750, time 56.919 samples/sec
+epoch 5, total_step 78340, total loss is 20.34 , inference loss is 11.36, weight deacy loss is 8.99, training accuracy is 0.187500, time 56.576 samples/sec
+epoch 5, total_step 78360, total loss is 22.68 , inference loss is 13.69, weight deacy loss is 8.99, training accuracy is 0.250000, time 56.392 samples/sec
+epoch 5, total_step 78380, total loss is 16.16 , inference loss is 7.18, weight deacy loss is 8.98, training accuracy is 0.375000, time 56.644 samples/sec
+epoch 5, total_step 78400, total loss is 18.31 , inference loss is 9.33, weight deacy loss is 8.98, training accuracy is 0.406250, time 56.757 samples/sec
+epoch 5, total_step 78420, total loss is 19.46 , inference loss is 10.47, weight deacy loss is 8.98, training accuracy is 0.250000, time 56.673 samples/sec
+epoch 5, total_step 78440, total loss is 17.76 , inference loss is 8.78, weight deacy loss is 8.98, training accuracy is 0.343750, time 56.363 samples/sec
+epoch 5, total_step 78460, total loss is 22.05 , inference loss is 13.07, weight deacy loss is 8.98, training accuracy is 0.281250, time 57.168 samples/sec
+epoch 5, total_step 78480, total loss is 18.02 , inference loss is 9.04, weight deacy loss is 8.98, training accuracy is 0.312500, time 56.273 samples/sec
+epoch 5, total_step 78500, total loss is 18.18 , inference loss is 9.20, weight deacy loss is 8.98, training accuracy is 0.281250, time 56.188 samples/sec
+epoch 5, total_step 78520, total loss is 16.06 , inference loss is 7.08, weight deacy loss is 8.98, training accuracy is 0.343750, time 56.259 samples/sec
+epoch 5, total_step 78540, total loss is 23.64 , inference loss is 14.66, weight deacy loss is 8.98, training accuracy is 0.218750, time 56.771 samples/sec
+epoch 5, total_step 78560, total loss is 17.83 , inference loss is 8.85, weight deacy loss is 8.98, training accuracy is 0.406250, time 56.991 samples/sec
+epoch 5, total_step 78580, total loss is 18.45 , inference loss is 9.47, weight deacy loss is 8.98, training accuracy is 0.375000, time 57.165 samples/sec
+epoch 5, total_step 78600, total loss is 20.26 , inference loss is 11.28, weight deacy loss is 8.97, training accuracy is 0.375000, time 55.981 samples/sec
+epoch 5, total_step 78620, total loss is 18.06 , inference loss is 9.09, weight deacy loss is 8.97, training accuracy is 0.375000, time 56.252 samples/sec
+epoch 5, total_step 78640, total loss is 19.19 , inference loss is 10.22, weight deacy loss is 8.97, training accuracy is 0.406250, time 55.694 samples/sec
+epoch 5, total_step 78660, total loss is 16.96 , inference loss is 7.99, weight deacy loss is 8.97, training accuracy is 0.468750, time 56.076 samples/sec
+epoch 5, total_step 78680, total loss is 22.84 , inference loss is 13.87, weight deacy loss is 8.97, training accuracy is 0.281250, time 56.849 samples/sec
+epoch 5, total_step 78700, total loss is 19.74 , inference loss is 10.77, weight deacy loss is 8.97, training accuracy is 0.375000, time 55.889 samples/sec
+epoch 5, total_step 78720, total loss is 12.73 , inference loss is 3.77, weight deacy loss is 8.97, training accuracy is 0.531250, time 57.112 samples/sec
+epoch 5, total_step 78740, total loss is 17.97 , inference loss is 9.01, weight deacy loss is 8.97, training accuracy is 0.406250, time 56.439 samples/sec
+epoch 5, total_step 78760, total loss is 17.19 , inference loss is 8.22, weight deacy loss is 8.97, training accuracy is 0.437500, time 57.010 samples/sec
+epoch 5, total_step 78780, total loss is 17.92 , inference loss is 8.95, weight deacy loss is 8.97, training accuracy is 0.437500, time 56.634 samples/sec
+epoch 5, total_step 78800, total loss is 13.60 , inference loss is 4.64, weight deacy loss is 8.96, training accuracy is 0.500000, time 55.951 samples/sec
+epoch 5, total_step 78820, total loss is 18.24 , inference loss is 9.28, weight deacy loss is 8.96, training accuracy is 0.375000, time 56.646 samples/sec
+epoch 5, total_step 78840, total loss is 17.70 , inference loss is 8.73, weight deacy loss is 8.96, training accuracy is 0.375000, time 56.417 samples/sec
+epoch 5, total_step 78860, total loss is 26.23 , inference loss is 17.27, weight deacy loss is 8.96, training accuracy is 0.281250, time 57.039 samples/sec
+epoch 5, total_step 78880, total loss is 16.91 , inference loss is 7.95, weight deacy loss is 8.96, training accuracy is 0.468750, time 56.149 samples/sec
+epoch 5, total_step 78900, total loss is 22.36 , inference loss is 13.40, weight deacy loss is 8.96, training accuracy is 0.343750, time 56.841 samples/sec
+epoch 5, total_step 78920, total loss is 14.76 , inference loss is 5.80, weight deacy loss is 8.96, training accuracy is 0.375000, time 56.471 samples/sec
+epoch 5, total_step 78940, total loss is 22.00 , inference loss is 13.04, weight deacy loss is 8.96, training accuracy is 0.375000, time 56.665 samples/sec
+epoch 5, total_step 78960, total loss is 17.51 , inference loss is 8.55, weight deacy loss is 8.96, training accuracy is 0.500000, time 57.443 samples/sec
+epoch 5, total_step 78980, total loss is 15.93 , inference loss is 6.97, weight deacy loss is 8.96, training accuracy is 0.343750, time 55.556 samples/sec
+epoch 5, total_step 79000, total loss is 14.85 , inference loss is 5.89, weight deacy loss is 8.95, training accuracy is 0.468750, time 55.449 samples/sec
+epoch 5, total_step 79020, total loss is 18.54 , inference loss is 9.58, weight deacy loss is 8.95, training accuracy is 0.437500, time 57.004 samples/sec
+epoch 5, total_step 79040, total loss is 15.28 , inference loss is 6.32, weight deacy loss is 8.95, training accuracy is 0.468750, time 56.880 samples/sec
+epoch 5, total_step 79060, total loss is 15.11 , inference loss is 6.16, weight deacy loss is 8.95, training accuracy is 0.531250, time 56.853 samples/sec
+epoch 5, total_step 79080, total loss is 15.63 , inference loss is 6.68, weight deacy loss is 8.95, training accuracy is 0.531250, time 56.798 samples/sec
+epoch 5, total_step 79100, total loss is 20.29 , inference loss is 11.34, weight deacy loss is 8.95, training accuracy is 0.250000, time 56.721 samples/sec
+epoch 5, total_step 79120, total loss is 19.23 , inference loss is 10.28, weight deacy loss is 8.95, training accuracy is 0.312500, time 56.885 samples/sec
+epoch 5, total_step 79140, total loss is 15.08 , inference loss is 6.14, weight deacy loss is 8.95, training accuracy is 0.406250, time 56.693 samples/sec
+epoch 5, total_step 79160, total loss is 21.50 , inference loss is 12.55, weight deacy loss is 8.95, training accuracy is 0.437500, time 55.852 samples/sec
+epoch 5, total_step 79180, total loss is 19.33 , inference loss is 10.38, weight deacy loss is 8.95, training accuracy is 0.406250, time 56.980 samples/sec
+epoch 5, total_step 79200, total loss is 19.40 , inference loss is 10.45, weight deacy loss is 8.95, training accuracy is 0.437500, time 56.211 samples/sec
+epoch 5, total_step 79220, total loss is 16.87 , inference loss is 7.92, weight deacy loss is 8.94, training accuracy is 0.343750, time 54.929 samples/sec
+epoch 5, total_step 79240, total loss is 16.06 , inference loss is 7.12, weight deacy loss is 8.94, training accuracy is 0.281250, time 56.270 samples/sec
+epoch 5, total_step 79260, total loss is 19.60 , inference loss is 10.65, weight deacy loss is 8.94, training accuracy is 0.312500, time 55.887 samples/sec
+epoch 5, total_step 79280, total loss is 17.56 , inference loss is 8.62, weight deacy loss is 8.94, training accuracy is 0.437500, time 55.819 samples/sec
+epoch 5, total_step 79300, total loss is 16.23 , inference loss is 7.29, weight deacy loss is 8.94, training accuracy is 0.406250, time 56.185 samples/sec
+epoch 5, total_step 79320, total loss is 18.44 , inference loss is 9.50, weight deacy loss is 8.94, training accuracy is 0.343750, time 56.465 samples/sec
+epoch 5, total_step 79340, total loss is 20.00 , inference loss is 11.06, weight deacy loss is 8.94, training accuracy is 0.312500, time 57.745 samples/sec
+epoch 5, total_step 79360, total loss is 20.19 , inference loss is 11.25, weight deacy loss is 8.94, training accuracy is 0.312500, time 56.021 samples/sec
+epoch 5, total_step 79380, total loss is 16.68 , inference loss is 7.74, weight deacy loss is 8.94, training accuracy is 0.406250, time 56.826 samples/sec
+epoch 5, total_step 79400, total loss is 17.47 , inference loss is 8.54, weight deacy loss is 8.94, training accuracy is 0.468750, time 57.202 samples/sec
+epoch 5, total_step 79420, total loss is 15.34 , inference loss is 6.40, weight deacy loss is 8.93, training accuracy is 0.468750, time 56.479 samples/sec
+epoch 5, total_step 79440, total loss is 19.82 , inference loss is 10.89, weight deacy loss is 8.93, training accuracy is 0.375000, time 57.380 samples/sec
+epoch 5, total_step 79460, total loss is 17.56 , inference loss is 8.62, weight deacy loss is 8.93, training accuracy is 0.343750, time 55.834 samples/sec
+epoch 5, total_step 79480, total loss is 14.92 , inference loss is 5.99, weight deacy loss is 8.93, training accuracy is 0.593750, time 56.681 samples/sec
+epoch 5, total_step 79500, total loss is 18.82 , inference loss is 9.89, weight deacy loss is 8.93, training accuracy is 0.406250, time 56.572 samples/sec
+epoch 5, total_step 79520, total loss is 16.91 , inference loss is 7.98, weight deacy loss is 8.93, training accuracy is 0.406250, time 55.825 samples/sec
+epoch 5, total_step 79540, total loss is 17.04 , inference loss is 8.12, weight deacy loss is 8.93, training accuracy is 0.406250, time 56.499 samples/sec
+epoch 5, total_step 79560, total loss is 14.45 , inference loss is 5.52, weight deacy loss is 8.93, training accuracy is 0.562500, time 57.076 samples/sec
+epoch 5, total_step 79580, total loss is 12.82 , inference loss is 3.89, weight deacy loss is 8.93, training accuracy is 0.593750, time 56.746 samples/sec
+epoch 5, total_step 79600, total loss is 16.79 , inference loss is 7.86, weight deacy loss is 8.93, training accuracy is 0.468750, time 56.814 samples/sec
+epoch 5, total_step 79620, total loss is 16.46 , inference loss is 7.53, weight deacy loss is 8.92, training accuracy is 0.437500, time 56.819 samples/sec
+epoch 5, total_step 79640, total loss is 19.34 , inference loss is 10.41, weight deacy loss is 8.92, training accuracy is 0.281250, time 56.595 samples/sec
+epoch 5, total_step 79660, total loss is 19.87 , inference loss is 10.95, weight deacy loss is 8.92, training accuracy is 0.375000, time 56.457 samples/sec
+epoch 5, total_step 79680, total loss is 18.09 , inference loss is 9.16, weight deacy loss is 8.92, training accuracy is 0.406250, time 56.358 samples/sec
+epoch 5, total_step 79700, total loss is 14.35 , inference loss is 5.43, weight deacy loss is 8.92, training accuracy is 0.406250, time 56.423 samples/sec
+epoch 5, total_step 79720, total loss is 20.62 , inference loss is 11.70, weight deacy loss is 8.92, training accuracy is 0.312500, time 57.577 samples/sec
+epoch 5, total_step 79740, total loss is 21.15 , inference loss is 12.23, weight deacy loss is 8.92, training accuracy is 0.343750, time 55.800 samples/sec
+epoch 5, total_step 79760, total loss is 17.54 , inference loss is 8.63, weight deacy loss is 8.92, training accuracy is 0.343750, time 57.338 samples/sec
+epoch 5, total_step 79780, total loss is 19.33 , inference loss is 10.41, weight deacy loss is 8.92, training accuracy is 0.312500, time 56.616 samples/sec
+epoch 5, total_step 79800, total loss is 17.17 , inference loss is 8.25, weight deacy loss is 8.92, training accuracy is 0.437500, time 56.899 samples/sec
+epoch 5, total_step 79820, total loss is 19.64 , inference loss is 10.73, weight deacy loss is 8.91, training accuracy is 0.312500, time 56.552 samples/sec
+epoch 5, total_step 79840, total loss is 21.02 , inference loss is 12.10, weight deacy loss is 8.91, training accuracy is 0.281250, time 56.322 samples/sec
+epoch 5, total_step 79860, total loss is 21.83 , inference loss is 12.92, weight deacy loss is 8.91, training accuracy is 0.312500, time 56.554 samples/sec
+epoch 5, total_step 79880, total loss is 16.94 , inference loss is 8.03, weight deacy loss is 8.91, training accuracy is 0.437500, time 56.518 samples/sec
+epoch 5, total_step 79900, total loss is 17.00 , inference loss is 8.09, weight deacy loss is 8.91, training accuracy is 0.375000, time 54.996 samples/sec
+epoch 5, total_step 79920, total loss is 17.32 , inference loss is 8.41, weight deacy loss is 8.91, training accuracy is 0.375000, time 56.679 samples/sec
+epoch 5, total_step 79940, total loss is 21.88 , inference loss is 12.97, weight deacy loss is 8.91, training accuracy is 0.281250, time 56.656 samples/sec
+epoch 5, total_step 79960, total loss is 21.30 , inference loss is 12.39, weight deacy loss is 8.91, training accuracy is 0.281250, time 56.816 samples/sec
+epoch 5, total_step 79980, total loss is 18.43 , inference loss is 9.52, weight deacy loss is 8.91, training accuracy is 0.343750, time 56.430 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.57001799999995
+best_threshold_index 118 0.9862962962962963
+best_threshold_index 118 0.9846296296296296
+best_threshold_index 121 0.9862962962962963
+best_threshold_index 117 0.9859259259259259
+best_threshold_index 119 0.9862962962962963
+best_threshold_index 119 0.985
+best_threshold_index 118 0.9861111111111112
+best_threshold_index 118 0.985
+best_threshold_index 121 0.985
+best_threshold_index 118 0.9855555555555555
+[lfw][80000]XNorm: 22.361229
+[lfw][80000]Accuracy-Flip: 0.98433+-0.00569
+testing verification..
+(14000, 512)
+infer time 85.66269899999999
+best_threshold_index 119 0.9855555555555555
+best_threshold_index 117 0.9838095238095238
+best_threshold_index 120 0.9844444444444445
+best_threshold_index 122 0.9839682539682539
+best_threshold_index 120 0.9847619047619047
+best_threshold_index 117 0.9842857142857143
+best_threshold_index 120 0.9852380952380952
+best_threshold_index 120 0.9839682539682539
+best_threshold_index 117 0.9836507936507937
+best_threshold_index 120 0.9853968253968254
+[cfp_ff][80000]XNorm: 21.803976
+[cfp_ff][80000]Accuracy-Flip: 0.98300+-0.00569
+testing verification..
+(14000, 512)
+infer time 85.13096900000002
+best_threshold_index 136 0.9042857142857142
+best_threshold_index 138 0.9033333333333333
+best_threshold_index 141 0.9028571428571428
+best_threshold_index 138 0.9057142857142857
+best_threshold_index 138 0.9057142857142857
+best_threshold_index 139 0.9028571428571428
+best_threshold_index 139 0.903015873015873
+best_threshold_index 139 0.9034920634920635
+best_threshold_index 136 0.9022222222222223
+best_threshold_index 139 0.9044444444444445
+[cfp_fp][80000]XNorm: 21.586986
+[cfp_fp][80000]Accuracy-Flip: 0.90129+-0.01015
+testing verification..
+(12000, 512)
+infer time 73.40570999999987
+best_threshold_index 141 0.892962962962963
+best_threshold_index 140 0.8912962962962963
+best_threshold_index 140 0.8931481481481481
+best_threshold_index 140 0.894074074074074
+best_threshold_index 140 0.8988888888888888
+best_threshold_index 141 0.8942592592592593
+best_threshold_index 140 0.8951851851851852
+best_threshold_index 142 0.9012962962962963
+best_threshold_index 142 0.8996296296296297
+best_threshold_index 141 0.8974074074074074
+[agedb_30][80000]XNorm: 22.622171
+[agedb_30][80000]Accuracy-Flip: 0.89267+-0.02819
+test accuracy is: 0.9843333333333334
+epoch 5, total_step 80000, total loss is 18.09 , inference loss is 9.19, weight deacy loss is 8.91, training accuracy is 0.343750, time 56.191 samples/sec
+epoch 5, total_step 80020, total loss is 22.07 , inference loss is 13.16, weight deacy loss is 8.91, training accuracy is 0.250000, time 55.904 samples/sec
+epoch 5, total_step 80040, total loss is 19.83 , inference loss is 10.92, weight deacy loss is 8.90, training accuracy is 0.312500, time 56.222 samples/sec
+epoch 5, total_step 80060, total loss is 19.11 , inference loss is 10.20, weight deacy loss is 8.90, training accuracy is 0.375000, time 56.258 samples/sec
+epoch 5, total_step 80080, total loss is 22.34 , inference loss is 13.43, weight deacy loss is 8.90, training accuracy is 0.250000, time 57.160 samples/sec
+epoch 5, total_step 80100, total loss is 19.13 , inference loss is 10.22, weight deacy loss is 8.90, training accuracy is 0.312500, time 55.896 samples/sec
+epoch 5, total_step 80120, total loss is 18.48 , inference loss is 9.58, weight deacy loss is 8.90, training accuracy is 0.250000, time 56.696 samples/sec
+epoch 5, total_step 80140, total loss is 17.03 , inference loss is 8.13, weight deacy loss is 8.90, training accuracy is 0.375000, time 56.395 samples/sec
+epoch 5, total_step 80160, total loss is 20.02 , inference loss is 11.12, weight deacy loss is 8.90, training accuracy is 0.312500, time 56.856 samples/sec
+epoch 5, total_step 80180, total loss is 21.28 , inference loss is 12.39, weight deacy loss is 8.90, training accuracy is 0.250000, time 56.676 samples/sec
+epoch 5, total_step 80200, total loss is 19.90 , inference loss is 11.00, weight deacy loss is 8.90, training accuracy is 0.281250, time 55.722 samples/sec
+epoch 5, total_step 80220, total loss is 14.60 , inference loss is 5.71, weight deacy loss is 8.90, training accuracy is 0.500000, time 57.045 samples/sec
+epoch 5, total_step 80240, total loss is 16.41 , inference loss is 7.52, weight deacy loss is 8.90, training accuracy is 0.500000, time 56.835 samples/sec
+epoch 5, total_step 80260, total loss is 20.67 , inference loss is 11.78, weight deacy loss is 8.89, training accuracy is 0.500000, time 54.943 samples/sec
+epoch 5, total_step 80280, total loss is 20.40 , inference loss is 11.51, weight deacy loss is 8.89, training accuracy is 0.187500, time 57.141 samples/sec
+epoch 5, total_step 80300, total loss is 17.01 , inference loss is 8.11, weight deacy loss is 8.89, training accuracy is 0.343750, time 55.777 samples/sec
+epoch 5, total_step 80320, total loss is 19.52 , inference loss is 10.62, weight deacy loss is 8.89, training accuracy is 0.312500, time 55.922 samples/sec
+epoch 5, total_step 80340, total loss is 22.02 , inference loss is 13.13, weight deacy loss is 8.89, training accuracy is 0.312500, time 56.859 samples/sec
+epoch 5, total_step 80360, total loss is 22.41 , inference loss is 13.52, weight deacy loss is 8.89, training accuracy is 0.187500, time 56.098 samples/sec
+epoch 5, total_step 80380, total loss is 20.51 , inference loss is 11.62, weight deacy loss is 8.89, training accuracy is 0.312500, time 56.106 samples/sec
+epoch 5, total_step 80400, total loss is 19.89 , inference loss is 11.00, weight deacy loss is 8.89, training accuracy is 0.375000, time 56.221 samples/sec
+epoch 5, total_step 80420, total loss is 23.23 , inference loss is 14.34, weight deacy loss is 8.89, training accuracy is 0.281250, time 56.794 samples/sec
+epoch 5, total_step 80440, total loss is 16.75 , inference loss is 7.86, weight deacy loss is 8.89, training accuracy is 0.281250, time 57.203 samples/sec
+epoch 5, total_step 80460, total loss is 16.75 , inference loss is 7.86, weight deacy loss is 8.89, training accuracy is 0.437500, time 57.686 samples/sec
+epoch 5, total_step 80480, total loss is 17.32 , inference loss is 8.44, weight deacy loss is 8.88, training accuracy is 0.406250, time 55.774 samples/sec
+epoch 5, total_step 80500, total loss is 21.56 , inference loss is 12.67, weight deacy loss is 8.88, training accuracy is 0.281250, time 55.778 samples/sec
+epoch 5, total_step 80520, total loss is 18.54 , inference loss is 9.66, weight deacy loss is 8.88, training accuracy is 0.375000, time 55.704 samples/sec
+epoch 5, total_step 80540, total loss is 20.12 , inference loss is 11.24, weight deacy loss is 8.88, training accuracy is 0.375000, time 56.131 samples/sec
+epoch 5, total_step 80560, total loss is 18.74 , inference loss is 9.86, weight deacy loss is 8.88, training accuracy is 0.406250, time 55.567 samples/sec
+epoch 5, total_step 80580, total loss is 17.50 , inference loss is 8.62, weight deacy loss is 8.88, training accuracy is 0.437500, time 56.364 samples/sec
+epoch 5, total_step 80600, total loss is 18.65 , inference loss is 9.77, weight deacy loss is 8.88, training accuracy is 0.406250, time 56.294 samples/sec
+epoch 5, total_step 80620, total loss is 19.60 , inference loss is 10.72, weight deacy loss is 8.88, training accuracy is 0.343750, time 56.742 samples/sec
+epoch 5, total_step 80640, total loss is 17.69 , inference loss is 8.81, weight deacy loss is 8.88, training accuracy is 0.250000, time 57.154 samples/sec
+epoch 5, total_step 80660, total loss is 22.48 , inference loss is 13.60, weight deacy loss is 8.88, training accuracy is 0.250000, time 56.177 samples/sec
+epoch 5, total_step 80680, total loss is 22.59 , inference loss is 13.71, weight deacy loss is 8.88, training accuracy is 0.281250, time 56.403 samples/sec
+epoch 5, total_step 80700, total loss is 20.19 , inference loss is 11.31, weight deacy loss is 8.87, training accuracy is 0.281250, time 53.868 samples/sec
+epoch 5, total_step 80720, total loss is 19.70 , inference loss is 10.82, weight deacy loss is 8.87, training accuracy is 0.281250, time 56.449 samples/sec
+epoch 5, total_step 80740, total loss is 22.15 , inference loss is 13.27, weight deacy loss is 8.87, training accuracy is 0.281250, time 57.025 samples/sec
+epoch 5, total_step 80760, total loss is 20.42 , inference loss is 11.55, weight deacy loss is 8.87, training accuracy is 0.343750, time 56.363 samples/sec
+epoch 5, total_step 80780, total loss is 18.52 , inference loss is 9.65, weight deacy loss is 8.87, training accuracy is 0.312500, time 56.545 samples/sec
+epoch 5, total_step 80800, total loss is 18.97 , inference loss is 10.09, weight deacy loss is 8.87, training accuracy is 0.406250, time 56.754 samples/sec
+epoch 5, total_step 80820, total loss is 22.49 , inference loss is 13.62, weight deacy loss is 8.87, training accuracy is 0.343750, time 56.573 samples/sec
+epoch 5, total_step 80840, total loss is 18.04 , inference loss is 9.17, weight deacy loss is 8.87, training accuracy is 0.343750, time 56.183 samples/sec
+epoch 5, total_step 80860, total loss is 18.69 , inference loss is 9.82, weight deacy loss is 8.87, training accuracy is 0.343750, time 56.225 samples/sec
+epoch 5, total_step 80880, total loss is 24.02 , inference loss is 15.15, weight deacy loss is 8.87, training accuracy is 0.250000, time 56.976 samples/sec
+epoch 5, total_step 80900, total loss is 17.15 , inference loss is 8.28, weight deacy loss is 8.87, training accuracy is 0.375000, time 56.474 samples/sec
+epoch 5, total_step 80920, total loss is 16.72 , inference loss is 7.86, weight deacy loss is 8.87, training accuracy is 0.375000, time 56.312 samples/sec
+epoch 5, total_step 80940, total loss is 24.45 , inference loss is 15.59, weight deacy loss is 8.86, training accuracy is 0.187500, time 56.849 samples/sec
+epoch 5, total_step 80960, total loss is 20.95 , inference loss is 12.09, weight deacy loss is 8.86, training accuracy is 0.218750, time 56.646 samples/sec
+epoch 5, total_step 80980, total loss is 15.44 , inference loss is 6.58, weight deacy loss is 8.86, training accuracy is 0.406250, time 56.517 samples/sec
+epoch 5, total_step 81000, total loss is 16.27 , inference loss is 7.41, weight deacy loss is 8.86, training accuracy is 0.468750, time 56.583 samples/sec
+epoch 5, total_step 81020, total loss is 19.22 , inference loss is 10.36, weight deacy loss is 8.86, training accuracy is 0.375000, time 57.208 samples/sec
+epoch 5, total_step 81040, total loss is 18.07 , inference loss is 9.21, weight deacy loss is 8.86, training accuracy is 0.343750, time 57.045 samples/sec
+epoch 5, total_step 81060, total loss is 23.32 , inference loss is 14.46, weight deacy loss is 8.86, training accuracy is 0.250000, time 56.021 samples/sec
+epoch 5, total_step 81080, total loss is 20.21 , inference loss is 11.36, weight deacy loss is 8.86, training accuracy is 0.250000, time 56.653 samples/sec
+epoch 5, total_step 81100, total loss is 20.87 , inference loss is 12.01, weight deacy loss is 8.86, training accuracy is 0.281250, time 57.138 samples/sec
+epoch 5, total_step 81120, total loss is 19.11 , inference loss is 10.25, weight deacy loss is 8.86, training accuracy is 0.281250, time 57.177 samples/sec
+epoch 5, total_step 81140, total loss is 17.43 , inference loss is 8.57, weight deacy loss is 8.86, training accuracy is 0.218750, time 55.250 samples/sec
+epoch 5, total_step 81160, total loss is 19.73 , inference loss is 10.88, weight deacy loss is 8.85, training accuracy is 0.406250, time 56.012 samples/sec
+epoch 5, total_step 81180, total loss is 19.14 , inference loss is 10.29, weight deacy loss is 8.85, training accuracy is 0.468750, time 56.894 samples/sec
+epoch 5, total_step 81200, total loss is 22.96 , inference loss is 14.11, weight deacy loss is 8.85, training accuracy is 0.218750, time 57.129 samples/sec
+epoch 5, total_step 81220, total loss is 23.09 , inference loss is 14.24, weight deacy loss is 8.85, training accuracy is 0.250000, time 56.371 samples/sec
+epoch 5, total_step 81240, total loss is 22.79 , inference loss is 13.94, weight deacy loss is 8.85, training accuracy is 0.250000, time 57.419 samples/sec
+epoch 5, total_step 81260, total loss is 20.19 , inference loss is 11.34, weight deacy loss is 8.85, training accuracy is 0.343750, time 56.656 samples/sec
+epoch 5, total_step 81280, total loss is 19.54 , inference loss is 10.69, weight deacy loss is 8.85, training accuracy is 0.343750, time 57.227 samples/sec
+epoch 5, total_step 81300, total loss is 18.82 , inference loss is 9.98, weight deacy loss is 8.85, training accuracy is 0.281250, time 56.515 samples/sec
+epoch 5, total_step 81320, total loss is 22.51 , inference loss is 13.67, weight deacy loss is 8.85, training accuracy is 0.281250, time 56.754 samples/sec
+epoch 5, total_step 81340, total loss is 23.25 , inference loss is 14.40, weight deacy loss is 8.85, training accuracy is 0.218750, time 56.712 samples/sec
+epoch 5, total_step 81360, total loss is 19.66 , inference loss is 10.82, weight deacy loss is 8.85, training accuracy is 0.250000, time 57.097 samples/sec
+epoch 5, total_step 81380, total loss is 18.45 , inference loss is 9.60, weight deacy loss is 8.84, training accuracy is 0.375000, time 56.466 samples/sec
+epoch 5, total_step 81400, total loss is 16.34 , inference loss is 7.50, weight deacy loss is 8.84, training accuracy is 0.250000, time 57.123 samples/sec
+epoch 5, total_step 81420, total loss is 15.02 , inference loss is 6.18, weight deacy loss is 8.84, training accuracy is 0.406250, time 56.996 samples/sec
+epoch 5, total_step 81440, total loss is 20.62 , inference loss is 11.78, weight deacy loss is 8.84, training accuracy is 0.281250, time 56.714 samples/sec
+epoch 5, total_step 81460, total loss is 18.16 , inference loss is 9.32, weight deacy loss is 8.84, training accuracy is 0.437500, time 56.653 samples/sec
+epoch 5, total_step 81480, total loss is 19.93 , inference loss is 11.09, weight deacy loss is 8.84, training accuracy is 0.343750, time 56.581 samples/sec
+epoch 5, total_step 81500, total loss is 19.92 , inference loss is 11.08, weight deacy loss is 8.84, training accuracy is 0.312500, time 56.664 samples/sec
+epoch 5, total_step 81520, total loss is 16.48 , inference loss is 7.64, weight deacy loss is 8.84, training accuracy is 0.343750, time 56.650 samples/sec
+epoch 5, total_step 81540, total loss is 23.56 , inference loss is 14.72, weight deacy loss is 8.84, training accuracy is 0.312500, time 54.955 samples/sec
+epoch 5, total_step 81560, total loss is 16.61 , inference loss is 7.77, weight deacy loss is 8.84, training accuracy is 0.281250, time 56.910 samples/sec
+epoch 5, total_step 81580, total loss is 17.88 , inference loss is 9.05, weight deacy loss is 8.84, training accuracy is 0.343750, time 56.625 samples/sec
+epoch 5, total_step 81600, total loss is 25.09 , inference loss is 16.26, weight deacy loss is 8.83, training accuracy is 0.312500, time 56.760 samples/sec
+epoch 5, total_step 81620, total loss is 17.90 , inference loss is 9.07, weight deacy loss is 8.83, training accuracy is 0.375000, time 54.766 samples/sec
+epoch 5, total_step 81640, total loss is 17.41 , inference loss is 8.58, weight deacy loss is 8.83, training accuracy is 0.312500, time 57.223 samples/sec
+epoch 5, total_step 81660, total loss is 18.62 , inference loss is 9.78, weight deacy loss is 8.83, training accuracy is 0.312500, time 56.036 samples/sec
+epoch 5, total_step 81680, total loss is 20.59 , inference loss is 11.76, weight deacy loss is 8.83, training accuracy is 0.250000, time 55.326 samples/sec
+epoch 5, total_step 81700, total loss is 20.76 , inference loss is 11.93, weight deacy loss is 8.83, training accuracy is 0.312500, time 55.923 samples/sec
+epoch 5, total_step 81720, total loss is 17.13 , inference loss is 8.30, weight deacy loss is 8.83, training accuracy is 0.406250, time 56.035 samples/sec
+epoch 5, total_step 81740, total loss is 19.28 , inference loss is 10.45, weight deacy loss is 8.83, training accuracy is 0.343750, time 56.054 samples/sec
+epoch 5, total_step 81760, total loss is 22.98 , inference loss is 14.15, weight deacy loss is 8.83, training accuracy is 0.250000, time 55.287 samples/sec
+epoch 5, total_step 81780, total loss is 20.35 , inference loss is 11.52, weight deacy loss is 8.83, training accuracy is 0.406250, time 55.076 samples/sec
+epoch 5, total_step 81800, total loss is 26.23 , inference loss is 17.41, weight deacy loss is 8.83, training accuracy is 0.343750, time 57.309 samples/sec
+epoch 5, total_step 81820, total loss is 19.83 , inference loss is 11.01, weight deacy loss is 8.82, training accuracy is 0.312500, time 55.904 samples/sec
+epoch 5, total_step 81840, total loss is 20.65 , inference loss is 11.83, weight deacy loss is 8.82, training accuracy is 0.250000, time 56.703 samples/sec
+epoch 5, total_step 81860, total loss is 17.65 , inference loss is 8.82, weight deacy loss is 8.82, training accuracy is 0.375000, time 55.791 samples/sec
+epoch 5, total_step 81880, total loss is 15.90 , inference loss is 7.08, weight deacy loss is 8.82, training accuracy is 0.343750, time 56.320 samples/sec
+epoch 5, total_step 81900, total loss is 19.47 , inference loss is 10.65, weight deacy loss is 8.82, training accuracy is 0.312500, time 55.330 samples/sec
+epoch 5, total_step 81920, total loss is 20.37 , inference loss is 11.55, weight deacy loss is 8.82, training accuracy is 0.406250, time 56.584 samples/sec
+epoch 5, total_step 81940, total loss is 23.07 , inference loss is 14.25, weight deacy loss is 8.82, training accuracy is 0.312500, time 56.708 samples/sec
+epoch 5, total_step 81960, total loss is 20.87 , inference loss is 12.05, weight deacy loss is 8.82, training accuracy is 0.250000, time 56.104 samples/sec
+epoch 5, total_step 81980, total loss is 21.48 , inference loss is 12.66, weight deacy loss is 8.82, training accuracy is 0.406250, time 57.060 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.01123300000005
+best_threshold_index 116 0.9846296296296296
+best_threshold_index 116 0.9831481481481481
+best_threshold_index 115 0.9838888888888889
+best_threshold_index 116 0.9837037037037037
+best_threshold_index 116 0.9837037037037037
+best_threshold_index 123 0.9835185185185186
+best_threshold_index 116 0.9835185185185186
+best_threshold_index 120 0.9833333333333333
+best_threshold_index 116 0.9822222222222222
+best_threshold_index 116 0.9822222222222222
+[lfw][82000]XNorm: 21.737266
+[lfw][82000]Accuracy-Flip: 0.98183+-0.00669
+testing verification..
+(14000, 512)
+infer time 84.71736899999993
+best_threshold_index 118 0.9852380952380952
+best_threshold_index 118 0.9836507936507937
+best_threshold_index 118 0.9834920634920635
+best_threshold_index 118 0.9826984126984127
+best_threshold_index 118 0.9836507936507937
+best_threshold_index 118 0.983015873015873
+best_threshold_index 118 0.9839682539682539
+best_threshold_index 118 0.9828571428571429
+best_threshold_index 118 0.9828571428571429
+best_threshold_index 118 0.9842857142857143
+[cfp_ff][82000]XNorm: 21.390029
+[cfp_ff][82000]Accuracy-Flip: 0.98357+-0.00671
+testing verification..
+(14000, 512)
+infer time 85.81360100000003
+best_threshold_index 136 0.9087301587301587
+best_threshold_index 135 0.9073015873015873
+best_threshold_index 136 0.9061904761904762
+best_threshold_index 136 0.9095238095238095
+best_threshold_index 135 0.9087301587301587
+best_threshold_index 140 0.9052380952380953
+best_threshold_index 140 0.9049206349206349
+best_threshold_index 136 0.9063492063492063
+best_threshold_index 135 0.906984126984127
+best_threshold_index 136 0.9068253968253969
+[cfp_fp][82000]XNorm: 21.242795
+[cfp_fp][82000]Accuracy-Flip: 0.90557+-0.01282
+testing verification..
+(12000, 512)
+infer time 73.15508699999988
+best_threshold_index 139 0.8944444444444445
+best_threshold_index 135 0.8927777777777778
+best_threshold_index 138 0.8948148148148148
+best_threshold_index 135 0.8938888888888888
+best_threshold_index 139 0.8992592592592593
+best_threshold_index 135 0.8937037037037037
+best_threshold_index 135 0.895
+best_threshold_index 139 0.9027777777777778
+best_threshold_index 138 0.9007407407407407
+best_threshold_index 138 0.8972222222222223
+[agedb_30][82000]XNorm: 22.492050
+[agedb_30][82000]Accuracy-Flip: 0.89150+-0.02736
+test accuracy is: 0.9818333333333336
+epoch 5, total_step 82000, total loss is 21.12 , inference loss is 12.30, weight deacy loss is 8.82, training accuracy is 0.218750, time 52.990 samples/sec
+epoch 5, total_step 82020, total loss is 19.41 , inference loss is 10.60, weight deacy loss is 8.82, training accuracy is 0.250000, time 56.530 samples/sec
+epoch 5, total_step 82040, total loss is 19.66 , inference loss is 10.84, weight deacy loss is 8.81, training accuracy is 0.406250, time 56.828 samples/sec
+epoch 5, total_step 82060, total loss is 19.58 , inference loss is 10.76, weight deacy loss is 8.81, training accuracy is 0.343750, time 56.481 samples/sec
+epoch 5, total_step 82080, total loss is 17.78 , inference loss is 8.97, weight deacy loss is 8.81, training accuracy is 0.375000, time 56.757 samples/sec
+epoch 5, total_step 82100, total loss is 17.81 , inference loss is 9.00, weight deacy loss is 8.81, training accuracy is 0.406250, time 57.576 samples/sec
+epoch 5, total_step 82120, total loss is 20.58 , inference loss is 11.77, weight deacy loss is 8.81, training accuracy is 0.375000, time 56.902 samples/sec
+epoch 5, total_step 82140, total loss is 20.99 , inference loss is 12.18, weight deacy loss is 8.81, training accuracy is 0.312500, time 56.412 samples/sec
+epoch 5, total_step 82160, total loss is 22.33 , inference loss is 13.52, weight deacy loss is 8.81, training accuracy is 0.312500, time 57.236 samples/sec
+epoch 5, total_step 82180, total loss is 22.99 , inference loss is 14.18, weight deacy loss is 8.81, training accuracy is 0.187500, time 55.607 samples/sec
+epoch 5, total_step 82200, total loss is 20.22 , inference loss is 11.41, weight deacy loss is 8.81, training accuracy is 0.406250, time 56.704 samples/sec
+epoch 5, total_step 82220, total loss is 19.51 , inference loss is 10.70, weight deacy loss is 8.81, training accuracy is 0.250000, time 56.044 samples/sec
+epoch 5, total_step 82240, total loss is 20.65 , inference loss is 11.85, weight deacy loss is 8.81, training accuracy is 0.250000, time 56.862 samples/sec
+epoch 5, total_step 82260, total loss is 19.90 , inference loss is 11.10, weight deacy loss is 8.80, training accuracy is 0.312500, time 56.370 samples/sec
+epoch 5, total_step 82280, total loss is 17.07 , inference loss is 8.27, weight deacy loss is 8.80, training accuracy is 0.312500, time 56.641 samples/sec
+epoch 5, total_step 82300, total loss is 19.25 , inference loss is 10.45, weight deacy loss is 8.80, training accuracy is 0.437500, time 56.109 samples/sec
+epoch 5, total_step 82320, total loss is 20.18 , inference loss is 11.38, weight deacy loss is 8.80, training accuracy is 0.343750, time 56.444 samples/sec
+epoch 5, total_step 82340, total loss is 18.31 , inference loss is 9.51, weight deacy loss is 8.80, training accuracy is 0.250000, time 55.988 samples/sec
+epoch 5, total_step 82360, total loss is 17.13 , inference loss is 8.33, weight deacy loss is 8.80, training accuracy is 0.437500, time 56.593 samples/sec
+epoch 5, total_step 82380, total loss is 13.85 , inference loss is 5.05, weight deacy loss is 8.80, training accuracy is 0.375000, time 56.512 samples/sec
+epoch 5, total_step 82400, total loss is 19.02 , inference loss is 10.22, weight deacy loss is 8.80, training accuracy is 0.156250, time 57.253 samples/sec
+epoch 5, total_step 82420, total loss is 21.67 , inference loss is 12.88, weight deacy loss is 8.80, training accuracy is 0.250000, time 56.754 samples/sec
+epoch 5, total_step 82440, total loss is 18.94 , inference loss is 10.14, weight deacy loss is 8.80, training accuracy is 0.375000, time 56.283 samples/sec
+epoch 5, total_step 82460, total loss is 21.75 , inference loss is 12.95, weight deacy loss is 8.80, training accuracy is 0.437500, time 57.255 samples/sec
+epoch 5, total_step 82480, total loss is 17.59 , inference loss is 8.80, weight deacy loss is 8.79, training accuracy is 0.343750, time 56.270 samples/sec
+epoch 5, total_step 82500, total loss is 17.90 , inference loss is 9.11, weight deacy loss is 8.79, training accuracy is 0.312500, time 55.902 samples/sec
+epoch 5, total_step 82520, total loss is 18.71 , inference loss is 9.91, weight deacy loss is 8.79, training accuracy is 0.468750, time 55.888 samples/sec
+epoch 5, total_step 82540, total loss is 19.90 , inference loss is 11.11, weight deacy loss is 8.79, training accuracy is 0.187500, time 57.014 samples/sec
+epoch 5, total_step 82560, total loss is 16.63 , inference loss is 7.84, weight deacy loss is 8.79, training accuracy is 0.281250, time 57.063 samples/sec
+epoch 5, total_step 82580, total loss is 18.03 , inference loss is 9.24, weight deacy loss is 8.79, training accuracy is 0.312500, time 55.900 samples/sec
+epoch 5, total_step 82600, total loss is 20.45 , inference loss is 11.66, weight deacy loss is 8.79, training accuracy is 0.250000, time 57.434 samples/sec
+epoch 5, total_step 82620, total loss is 16.80 , inference loss is 8.01, weight deacy loss is 8.79, training accuracy is 0.406250, time 56.571 samples/sec
+epoch 5, total_step 82640, total loss is 20.01 , inference loss is 11.23, weight deacy loss is 8.79, training accuracy is 0.250000, time 56.728 samples/sec
+epoch 5, total_step 82660, total loss is 18.04 , inference loss is 9.26, weight deacy loss is 8.79, training accuracy is 0.406250, time 56.694 samples/sec
+epoch 5, total_step 82680, total loss is 15.68 , inference loss is 6.90, weight deacy loss is 8.79, training accuracy is 0.437500, time 56.870 samples/sec
+epoch 5, total_step 82700, total loss is 16.59 , inference loss is 7.81, weight deacy loss is 8.78, training accuracy is 0.375000, time 54.804 samples/sec
+epoch 5, total_step 82720, total loss is 20.61 , inference loss is 11.83, weight deacy loss is 8.78, training accuracy is 0.218750, time 55.545 samples/sec
+epoch 5, total_step 82740, total loss is 20.29 , inference loss is 11.51, weight deacy loss is 8.78, training accuracy is 0.375000, time 56.712 samples/sec
+epoch 5, total_step 82760, total loss is 16.40 , inference loss is 7.62, weight deacy loss is 8.78, training accuracy is 0.312500, time 55.759 samples/sec
+epoch 5, total_step 82780, total loss is 19.53 , inference loss is 10.75, weight deacy loss is 8.78, training accuracy is 0.281250, time 56.574 samples/sec
+epoch 5, total_step 82800, total loss is 21.16 , inference loss is 12.38, weight deacy loss is 8.78, training accuracy is 0.187500, time 56.846 samples/sec
+epoch 5, total_step 82820, total loss is 20.71 , inference loss is 11.94, weight deacy loss is 8.78, training accuracy is 0.406250, time 56.611 samples/sec
+epoch 5, total_step 82840, total loss is 18.27 , inference loss is 9.49, weight deacy loss is 8.78, training accuracy is 0.312500, time 56.463 samples/sec
+epoch 5, total_step 82860, total loss is 24.51 , inference loss is 15.73, weight deacy loss is 8.78, training accuracy is 0.281250, time 56.709 samples/sec
+epoch 5, total_step 82880, total loss is 20.36 , inference loss is 11.58, weight deacy loss is 8.78, training accuracy is 0.218750, time 56.443 samples/sec
+epoch 5, total_step 82900, total loss is 19.69 , inference loss is 10.91, weight deacy loss is 8.78, training accuracy is 0.343750, time 56.679 samples/sec
+epoch 5, total_step 82920, total loss is 16.94 , inference loss is 8.17, weight deacy loss is 8.77, training accuracy is 0.250000, time 57.415 samples/sec
+epoch 5, total_step 82940, total loss is 19.52 , inference loss is 10.75, weight deacy loss is 8.77, training accuracy is 0.437500, time 57.209 samples/sec
+epoch 5, total_step 82960, total loss is 18.97 , inference loss is 10.20, weight deacy loss is 8.77, training accuracy is 0.343750, time 56.912 samples/sec
+epoch 5, total_step 82980, total loss is 15.81 , inference loss is 7.04, weight deacy loss is 8.77, training accuracy is 0.343750, time 56.812 samples/sec
+epoch 5, total_step 83000, total loss is 20.53 , inference loss is 11.76, weight deacy loss is 8.77, training accuracy is 0.406250, time 56.737 samples/sec
+epoch 5, total_step 83020, total loss is 20.90 , inference loss is 12.13, weight deacy loss is 8.77, training accuracy is 0.250000, time 56.780 samples/sec
+epoch 5, total_step 83040, total loss is 19.96 , inference loss is 11.19, weight deacy loss is 8.77, training accuracy is 0.281250, time 56.530 samples/sec
+epoch 5, total_step 83060, total loss is 21.29 , inference loss is 12.53, weight deacy loss is 8.77, training accuracy is 0.250000, time 56.325 samples/sec
+epoch 5, total_step 83080, total loss is 21.51 , inference loss is 12.74, weight deacy loss is 8.77, training accuracy is 0.250000, time 56.007 samples/sec
+epoch 5, total_step 83100, total loss is 22.40 , inference loss is 13.63, weight deacy loss is 8.77, training accuracy is 0.250000, time 55.852 samples/sec
+epoch 5, total_step 83120, total loss is 23.53 , inference loss is 14.76, weight deacy loss is 8.77, training accuracy is 0.218750, time 56.945 samples/sec
+epoch 5, total_step 83140, total loss is 21.15 , inference loss is 12.39, weight deacy loss is 8.76, training accuracy is 0.187500, time 56.250 samples/sec
+epoch 5, total_step 83160, total loss is 19.67 , inference loss is 10.91, weight deacy loss is 8.76, training accuracy is 0.312500, time 56.683 samples/sec
+epoch 5, total_step 83180, total loss is 19.46 , inference loss is 10.70, weight deacy loss is 8.76, training accuracy is 0.343750, time 56.550 samples/sec
+epoch 5, total_step 83200, total loss is 16.53 , inference loss is 7.77, weight deacy loss is 8.76, training accuracy is 0.500000, time 56.312 samples/sec
+epoch 5, total_step 83220, total loss is 18.47 , inference loss is 9.71, weight deacy loss is 8.76, training accuracy is 0.125000, time 57.635 samples/sec
+epoch 5, total_step 83240, total loss is 20.77 , inference loss is 12.01, weight deacy loss is 8.76, training accuracy is 0.437500, time 56.563 samples/sec
+epoch 5, total_step 83260, total loss is 18.48 , inference loss is 9.72, weight deacy loss is 8.76, training accuracy is 0.312500, time 56.414 samples/sec
+epoch 5, total_step 83280, total loss is 18.86 , inference loss is 10.11, weight deacy loss is 8.76, training accuracy is 0.312500, time 56.479 samples/sec
+epoch 5, total_step 83300, total loss is 18.89 , inference loss is 10.14, weight deacy loss is 8.76, training accuracy is 0.406250, time 56.553 samples/sec
+epoch 5, total_step 83320, total loss is 19.01 , inference loss is 10.26, weight deacy loss is 8.76, training accuracy is 0.250000, time 56.738 samples/sec
+epoch 5, total_step 83340, total loss is 22.47 , inference loss is 13.72, weight deacy loss is 8.76, training accuracy is 0.156250, time 56.707 samples/sec
+epoch 5, total_step 83360, total loss is 17.64 , inference loss is 8.89, weight deacy loss is 8.75, training accuracy is 0.375000, time 56.692 samples/sec
+epoch 5, total_step 83380, total loss is 21.53 , inference loss is 12.78, weight deacy loss is 8.75, training accuracy is 0.281250, time 56.766 samples/sec
+epoch 5, total_step 83400, total loss is 18.71 , inference loss is 9.96, weight deacy loss is 8.75, training accuracy is 0.281250, time 56.266 samples/sec
+epoch 5, total_step 83420, total loss is 18.38 , inference loss is 9.63, weight deacy loss is 8.75, training accuracy is 0.343750, time 56.970 samples/sec
+epoch 5, total_step 83440, total loss is 16.45 , inference loss is 7.70, weight deacy loss is 8.75, training accuracy is 0.500000, time 56.806 samples/sec
+epoch 5, total_step 83460, total loss is 18.63 , inference loss is 9.88, weight deacy loss is 8.75, training accuracy is 0.250000, time 57.052 samples/sec
+epoch 5, total_step 83480, total loss is 18.18 , inference loss is 9.43, weight deacy loss is 8.75, training accuracy is 0.250000, time 56.190 samples/sec
+epoch 5, total_step 83500, total loss is 19.74 , inference loss is 10.99, weight deacy loss is 8.75, training accuracy is 0.312500, time 57.386 samples/sec
+epoch 5, total_step 83520, total loss is 18.01 , inference loss is 9.26, weight deacy loss is 8.75, training accuracy is 0.312500, time 56.536 samples/sec
+epoch 5, total_step 83540, total loss is 22.41 , inference loss is 13.66, weight deacy loss is 8.75, training accuracy is 0.218750, time 56.783 samples/sec
+epoch 5, total_step 83560, total loss is 18.90 , inference loss is 10.15, weight deacy loss is 8.75, training accuracy is 0.406250, time 56.636 samples/sec
+epoch 5, total_step 83580, total loss is 18.83 , inference loss is 10.08, weight deacy loss is 8.74, training accuracy is 0.218750, time 56.115 samples/sec
+epoch 5, total_step 83600, total loss is 20.11 , inference loss is 11.37, weight deacy loss is 8.74, training accuracy is 0.187500, time 56.889 samples/sec
+epoch 5, total_step 83620, total loss is 19.71 , inference loss is 10.97, weight deacy loss is 8.74, training accuracy is 0.312500, time 56.077 samples/sec
+epoch 5, total_step 83640, total loss is 23.18 , inference loss is 14.44, weight deacy loss is 8.74, training accuracy is 0.187500, time 56.162 samples/sec
+epoch 5, total_step 83660, total loss is 20.59 , inference loss is 11.85, weight deacy loss is 8.74, training accuracy is 0.312500, time 56.239 samples/sec
+epoch 5, total_step 83680, total loss is 24.49 , inference loss is 15.75, weight deacy loss is 8.74, training accuracy is 0.093750, time 56.823 samples/sec
+epoch 5, total_step 83700, total loss is 19.16 , inference loss is 10.42, weight deacy loss is 8.74, training accuracy is 0.281250, time 55.919 samples/sec
+epoch 5, total_step 83720, total loss is 18.29 , inference loss is 9.55, weight deacy loss is 8.74, training accuracy is 0.281250, time 55.534 samples/sec
+epoch 5, total_step 83740, total loss is 16.97 , inference loss is 8.24, weight deacy loss is 8.74, training accuracy is 0.468750, time 56.082 samples/sec
+epoch 5, total_step 83760, total loss is 19.35 , inference loss is 10.61, weight deacy loss is 8.74, training accuracy is 0.375000, time 56.078 samples/sec
+epoch 5, total_step 83780, total loss is 26.07 , inference loss is 17.34, weight deacy loss is 8.74, training accuracy is 0.125000, time 56.812 samples/sec
+epoch 5, total_step 83800, total loss is 23.80 , inference loss is 15.06, weight deacy loss is 8.74, training accuracy is 0.218750, time 56.848 samples/sec
+epoch 5, total_step 83820, total loss is 18.06 , inference loss is 9.33, weight deacy loss is 8.73, training accuracy is 0.437500, time 56.525 samples/sec
+epoch 5, total_step 83840, total loss is 21.11 , inference loss is 12.37, weight deacy loss is 8.73, training accuracy is 0.156250, time 56.358 samples/sec
+epoch 5, total_step 83860, total loss is 21.33 , inference loss is 12.60, weight deacy loss is 8.73, training accuracy is 0.312500, time 56.646 samples/sec
+epoch 5, total_step 83880, total loss is 20.05 , inference loss is 11.31, weight deacy loss is 8.73, training accuracy is 0.375000, time 57.077 samples/sec
+epoch 5, total_step 83900, total loss is 18.47 , inference loss is 9.74, weight deacy loss is 8.73, training accuracy is 0.437500, time 56.979 samples/sec
+epoch 5, total_step 83920, total loss is 24.12 , inference loss is 15.39, weight deacy loss is 8.73, training accuracy is 0.125000, time 56.342 samples/sec
+epoch 5, total_step 83940, total loss is 18.12 , inference loss is 9.39, weight deacy loss is 8.73, training accuracy is 0.250000, time 55.567 samples/sec
+epoch 5, total_step 83960, total loss is 17.96 , inference loss is 9.24, weight deacy loss is 8.73, training accuracy is 0.312500, time 57.059 samples/sec
+epoch 5, total_step 83980, total loss is 19.06 , inference loss is 10.33, weight deacy loss is 8.73, training accuracy is 0.250000, time 57.598 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.34285800000005
+best_threshold_index 114 0.9885185185185185
+best_threshold_index 113 0.9877777777777778
+best_threshold_index 113 0.9877777777777778
+best_threshold_index 114 0.9879629629629629
+best_threshold_index 113 0.9885185185185185
+best_threshold_index 114 0.987037037037037
+best_threshold_index 115 0.9887037037037038
+best_threshold_index 114 0.9868518518518519
+best_threshold_index 115 0.9868518518518519
+best_threshold_index 114 0.987037037037037
+[lfw][84000]XNorm: 21.762353
+[lfw][84000]Accuracy-Flip: 0.98683+-0.00647
+testing verification..
+(14000, 512)
+infer time 85.33439299999996
+best_threshold_index 115 0.9857142857142858
+best_threshold_index 117 0.9842857142857143
+best_threshold_index 115 0.9844444444444445
+best_threshold_index 115 0.9842857142857143
+best_threshold_index 115 0.9846031746031746
+best_threshold_index 115 0.9844444444444445
+best_threshold_index 117 0.9852380952380952
+best_threshold_index 115 0.9842857142857143
+best_threshold_index 117 0.9847619047619047
+best_threshold_index 117 0.9857142857142858
+[cfp_ff][84000]XNorm: 21.305906
+[cfp_ff][84000]Accuracy-Flip: 0.98414+-0.00488
+testing verification..
+(14000, 512)
+infer time 85.88468900000007
+best_threshold_index 135 0.9171428571428571
+best_threshold_index 135 0.9163492063492064
+best_threshold_index 139 0.9158730158730158
+best_threshold_index 135 0.9163492063492064
+best_threshold_index 139 0.9193650793650794
+best_threshold_index 135 0.9155555555555556
+best_threshold_index 139 0.9131746031746032
+best_threshold_index 139 0.9158730158730158
+best_threshold_index 135 0.9161904761904762
+best_threshold_index 139 0.9160317460317461
+[cfp_fp][84000]XNorm: 20.808232
+[cfp_fp][84000]Accuracy-Flip: 0.91357+-0.01305
+testing verification..
+(12000, 512)
+infer time 73.45869400000001
+best_threshold_index 131 0.8955555555555555
+best_threshold_index 131 0.8935185185185185
+best_threshold_index 130 0.8966666666666666
+best_threshold_index 131 0.8968518518518519
+best_threshold_index 130 0.9020370370370371
+best_threshold_index 131 0.8951851851851852
+best_threshold_index 131 0.8966666666666666
+best_threshold_index 131 0.9009259259259259
+best_threshold_index 131 0.8992592592592593
+best_threshold_index 131 0.8985185185185185
+[agedb_30][84000]XNorm: 22.509831
+[agedb_30][84000]Accuracy-Flip: 0.89633+-0.02344
+test accuracy is: 0.9868333333333335
+epoch 5, total_step 84000, total loss is 17.88 , inference loss is 9.16, weight deacy loss is 8.73, training accuracy is 0.375000, time 53.938 samples/sec
+epoch 5, total_step 84020, total loss is 18.35 , inference loss is 9.63, weight deacy loss is 8.73, training accuracy is 0.218750, time 56.714 samples/sec
+epoch 5, total_step 84040, total loss is 18.75 , inference loss is 10.03, weight deacy loss is 8.72, training accuracy is 0.312500, time 56.128 samples/sec
+epoch 5, total_step 84060, total loss is 17.81 , inference loss is 9.08, weight deacy loss is 8.72, training accuracy is 0.312500, time 56.755 samples/sec
+epoch 5, total_step 84080, total loss is 17.87 , inference loss is 9.15, weight deacy loss is 8.72, training accuracy is 0.343750, time 55.793 samples/sec
+epoch 5, total_step 84100, total loss is 19.19 , inference loss is 10.47, weight deacy loss is 8.72, training accuracy is 0.406250, time 56.236 samples/sec
+epoch 5, total_step 84120, total loss is 16.07 , inference loss is 7.35, weight deacy loss is 8.72, training accuracy is 0.375000, time 56.110 samples/sec
+epoch 5, total_step 84140, total loss is 20.77 , inference loss is 12.05, weight deacy loss is 8.72, training accuracy is 0.343750, time 55.609 samples/sec
+epoch 5, total_step 84160, total loss is 18.80 , inference loss is 10.08, weight deacy loss is 8.72, training accuracy is 0.312500, time 56.648 samples/sec
+epoch 5, total_step 84180, total loss is 19.05 , inference loss is 10.33, weight deacy loss is 8.72, training accuracy is 0.281250, time 55.891 samples/sec
+epoch 5, total_step 84200, total loss is 20.34 , inference loss is 11.62, weight deacy loss is 8.72, training accuracy is 0.343750, time 56.558 samples/sec
+epoch 5, total_step 84220, total loss is 19.15 , inference loss is 10.44, weight deacy loss is 8.72, training accuracy is 0.375000, time 56.372 samples/sec
+epoch 5, total_step 84240, total loss is 19.68 , inference loss is 10.97, weight deacy loss is 8.72, training accuracy is 0.343750, time 56.490 samples/sec
+epoch 5, total_step 84260, total loss is 17.96 , inference loss is 9.25, weight deacy loss is 8.71, training accuracy is 0.312500, time 56.649 samples/sec
+epoch 5, total_step 84280, total loss is 18.44 , inference loss is 9.73, weight deacy loss is 8.71, training accuracy is 0.343750, time 56.573 samples/sec
+epoch 5, total_step 84300, total loss is 16.51 , inference loss is 7.80, weight deacy loss is 8.71, training accuracy is 0.406250, time 56.139 samples/sec
+epoch 5, total_step 84320, total loss is 18.58 , inference loss is 9.86, weight deacy loss is 8.71, training accuracy is 0.375000, time 56.957 samples/sec
+epoch 5, total_step 84340, total loss is 15.06 , inference loss is 6.35, weight deacy loss is 8.71, training accuracy is 0.406250, time 56.226 samples/sec
+epoch 5, total_step 84360, total loss is 19.82 , inference loss is 11.11, weight deacy loss is 8.71, training accuracy is 0.406250, time 56.236 samples/sec
+epoch 5, total_step 84380, total loss is 17.30 , inference loss is 8.59, weight deacy loss is 8.71, training accuracy is 0.281250, time 57.089 samples/sec
+epoch 5, total_step 84400, total loss is 18.11 , inference loss is 9.40, weight deacy loss is 8.71, training accuracy is 0.343750, time 56.133 samples/sec
+epoch 5, total_step 84420, total loss is 15.30 , inference loss is 6.59, weight deacy loss is 8.71, training accuracy is 0.468750, time 56.137 samples/sec
+epoch 5, total_step 84440, total loss is 18.20 , inference loss is 9.49, weight deacy loss is 8.71, training accuracy is 0.343750, time 55.930 samples/sec
+epoch 5, total_step 84460, total loss is 21.66 , inference loss is 12.95, weight deacy loss is 8.71, training accuracy is 0.281250, time 57.077 samples/sec
+epoch 5, total_step 84480, total loss is 18.44 , inference loss is 9.74, weight deacy loss is 8.70, training accuracy is 0.312500, time 55.887 samples/sec
+epoch 5, total_step 84500, total loss is 16.22 , inference loss is 7.51, weight deacy loss is 8.70, training accuracy is 0.437500, time 54.748 samples/sec
+epoch 5, total_step 84520, total loss is 16.58 , inference loss is 7.88, weight deacy loss is 8.70, training accuracy is 0.468750, time 56.946 samples/sec
+epoch 5, total_step 84540, total loss is 19.61 , inference loss is 10.91, weight deacy loss is 8.70, training accuracy is 0.281250, time 56.965 samples/sec
+epoch 5, total_step 84560, total loss is 19.10 , inference loss is 10.39, weight deacy loss is 8.70, training accuracy is 0.500000, time 56.417 samples/sec
+epoch 5, total_step 84580, total loss is 21.94 , inference loss is 13.24, weight deacy loss is 8.70, training accuracy is 0.250000, time 56.400 samples/sec
+epoch 5, total_step 84600, total loss is 18.43 , inference loss is 9.73, weight deacy loss is 8.70, training accuracy is 0.281250, time 57.013 samples/sec
+epoch 5, total_step 84620, total loss is 20.78 , inference loss is 12.09, weight deacy loss is 8.70, training accuracy is 0.218750, time 57.125 samples/sec
+epoch 5, total_step 84640, total loss is 22.55 , inference loss is 13.85, weight deacy loss is 8.70, training accuracy is 0.187500, time 55.324 samples/sec
+epoch 5, total_step 84660, total loss is 19.86 , inference loss is 11.16, weight deacy loss is 8.70, training accuracy is 0.156250, time 56.189 samples/sec
+epoch 5, total_step 84680, total loss is 18.60 , inference loss is 9.90, weight deacy loss is 8.70, training accuracy is 0.312500, time 56.055 samples/sec
+epoch 5, total_step 84700, total loss is 22.33 , inference loss is 13.63, weight deacy loss is 8.69, training accuracy is 0.218750, time 56.145 samples/sec
+epoch 5, total_step 84720, total loss is 17.43 , inference loss is 8.73, weight deacy loss is 8.69, training accuracy is 0.312500, time 57.329 samples/sec
+epoch 5, total_step 84740, total loss is 16.88 , inference loss is 8.19, weight deacy loss is 8.69, training accuracy is 0.468750, time 56.746 samples/sec
+epoch 5, total_step 84760, total loss is 18.14 , inference loss is 9.45, weight deacy loss is 8.69, training accuracy is 0.312500, time 56.144 samples/sec
+epoch 5, total_step 84780, total loss is 21.05 , inference loss is 12.36, weight deacy loss is 8.69, training accuracy is 0.281250, time 56.360 samples/sec
+epoch 5, total_step 84800, total loss is 16.85 , inference loss is 8.16, weight deacy loss is 8.69, training accuracy is 0.375000, time 56.766 samples/sec
+epoch 5, total_step 84820, total loss is 18.61 , inference loss is 9.92, weight deacy loss is 8.69, training accuracy is 0.281250, time 56.894 samples/sec
+epoch 5, total_step 84840, total loss is 20.45 , inference loss is 11.76, weight deacy loss is 8.69, training accuracy is 0.218750, time 56.025 samples/sec
+epoch 5, total_step 84860, total loss is 19.85 , inference loss is 11.16, weight deacy loss is 8.69, training accuracy is 0.281250, time 55.771 samples/sec
+epoch 5, total_step 84880, total loss is 16.76 , inference loss is 8.07, weight deacy loss is 8.69, training accuracy is 0.375000, time 55.611 samples/sec
+epoch 5, total_step 84900, total loss is 18.88 , inference loss is 10.19, weight deacy loss is 8.69, training accuracy is 0.375000, time 55.537 samples/sec
+epoch 5, total_step 84920, total loss is 17.53 , inference loss is 8.85, weight deacy loss is 8.68, training accuracy is 0.343750, time 56.706 samples/sec
+epoch 5, total_step 84940, total loss is 19.34 , inference loss is 10.65, weight deacy loss is 8.68, training accuracy is 0.343750, time 56.822 samples/sec
+epoch 5, total_step 84960, total loss is 18.27 , inference loss is 9.59, weight deacy loss is 8.68, training accuracy is 0.281250, time 55.416 samples/sec
+epoch 5, total_step 84980, total loss is 15.72 , inference loss is 7.04, weight deacy loss is 8.68, training accuracy is 0.406250, time 56.190 samples/sec
+epoch 5, total_step 85000, total loss is 17.74 , inference loss is 9.06, weight deacy loss is 8.68, training accuracy is 0.281250, time 56.698 samples/sec
+epoch 5, total_step 85020, total loss is 21.53 , inference loss is 12.85, weight deacy loss is 8.68, training accuracy is 0.343750, time 56.450 samples/sec
+epoch 5, total_step 85040, total loss is 18.47 , inference loss is 9.79, weight deacy loss is 8.68, training accuracy is 0.312500, time 57.045 samples/sec
+epoch 5, total_step 85060, total loss is 20.99 , inference loss is 12.31, weight deacy loss is 8.68, training accuracy is 0.312500, time 56.877 samples/sec
+epoch 5, total_step 85080, total loss is 18.64 , inference loss is 9.96, weight deacy loss is 8.68, training accuracy is 0.343750, time 56.917 samples/sec
+epoch 5, total_step 85100, total loss is 19.08 , inference loss is 10.40, weight deacy loss is 8.68, training accuracy is 0.218750, time 56.610 samples/sec
+epoch 5, total_step 85120, total loss is 24.57 , inference loss is 15.90, weight deacy loss is 8.68, training accuracy is 0.312500, time 56.128 samples/sec
+epoch 5, total_step 85140, total loss is 20.75 , inference loss is 12.08, weight deacy loss is 8.67, training accuracy is 0.218750, time 55.698 samples/sec
+epoch 5, total_step 85160, total loss is 20.17 , inference loss is 11.49, weight deacy loss is 8.67, training accuracy is 0.218750, time 56.990 samples/sec
+epoch 5, total_step 85180, total loss is 19.23 , inference loss is 10.56, weight deacy loss is 8.67, training accuracy is 0.312500, time 56.904 samples/sec
+epoch 5, total_step 85200, total loss is 20.92 , inference loss is 12.24, weight deacy loss is 8.67, training accuracy is 0.218750, time 56.215 samples/sec
+epoch 5, total_step 85220, total loss is 23.32 , inference loss is 14.65, weight deacy loss is 8.67, training accuracy is 0.187500, time 56.142 samples/sec
+epoch 5, total_step 85240, total loss is 25.19 , inference loss is 16.52, weight deacy loss is 8.67, training accuracy is 0.250000, time 56.375 samples/sec
+epoch 5, total_step 85260, total loss is 22.06 , inference loss is 13.39, weight deacy loss is 8.67, training accuracy is 0.062500, time 56.656 samples/sec
+epoch 5, total_step 85280, total loss is 23.95 , inference loss is 15.28, weight deacy loss is 8.67, training accuracy is 0.218750, time 56.445 samples/sec
+epoch 5, total_step 85300, total loss is 19.21 , inference loss is 10.54, weight deacy loss is 8.67, training accuracy is 0.281250, time 56.055 samples/sec
+epoch 5, total_step 85320, total loss is 19.54 , inference loss is 10.87, weight deacy loss is 8.67, training accuracy is 0.281250, time 56.166 samples/sec
+epoch 5, total_step 85340, total loss is 18.64 , inference loss is 9.97, weight deacy loss is 8.67, training accuracy is 0.281250, time 56.921 samples/sec
+epoch 5, total_step 85360, total loss is 18.75 , inference loss is 10.09, weight deacy loss is 8.67, training accuracy is 0.218750, time 56.286 samples/sec
+epoch 5, total_step 85380, total loss is 20.17 , inference loss is 11.51, weight deacy loss is 8.66, training accuracy is 0.312500, time 56.363 samples/sec
+epoch 5, total_step 85400, total loss is 18.49 , inference loss is 9.83, weight deacy loss is 8.66, training accuracy is 0.250000, time 55.788 samples/sec
+epoch 5, total_step 85420, total loss is 20.05 , inference loss is 11.39, weight deacy loss is 8.66, training accuracy is 0.250000, time 55.787 samples/sec
+epoch 5, total_step 85440, total loss is 18.94 , inference loss is 10.28, weight deacy loss is 8.66, training accuracy is 0.312500, time 56.463 samples/sec
+epoch 5, total_step 85460, total loss is 20.17 , inference loss is 11.51, weight deacy loss is 8.66, training accuracy is 0.312500, time 56.259 samples/sec
+epoch 5, total_step 85480, total loss is 19.97 , inference loss is 11.31, weight deacy loss is 8.66, training accuracy is 0.187500, time 56.955 samples/sec
+epoch 5, total_step 85500, total loss is 18.23 , inference loss is 9.57, weight deacy loss is 8.66, training accuracy is 0.375000, time 55.536 samples/sec
+epoch 5, total_step 85520, total loss is 18.36 , inference loss is 9.70, weight deacy loss is 8.66, training accuracy is 0.375000, time 56.501 samples/sec
+epoch 5, total_step 85540, total loss is 20.50 , inference loss is 11.85, weight deacy loss is 8.66, training accuracy is 0.375000, time 56.736 samples/sec
+epoch 5, total_step 85560, total loss is 21.78 , inference loss is 13.12, weight deacy loss is 8.66, training accuracy is 0.218750, time 55.735 samples/sec
+epoch 5, total_step 85580, total loss is 20.01 , inference loss is 11.35, weight deacy loss is 8.66, training accuracy is 0.281250, time 56.838 samples/sec
+epoch 5, total_step 85600, total loss is 16.85 , inference loss is 8.20, weight deacy loss is 8.66, training accuracy is 0.375000, time 56.657 samples/sec
+epoch 5, total_step 85620, total loss is 20.47 , inference loss is 11.82, weight deacy loss is 8.65, training accuracy is 0.312500, time 57.313 samples/sec
+epoch 5, total_step 85640, total loss is 20.96 , inference loss is 12.31, weight deacy loss is 8.65, training accuracy is 0.312500, time 56.744 samples/sec
+epoch 5, total_step 85660, total loss is 18.97 , inference loss is 10.32, weight deacy loss is 8.65, training accuracy is 0.281250, time 56.679 samples/sec
+epoch 5, total_step 85680, total loss is 26.27 , inference loss is 17.62, weight deacy loss is 8.65, training accuracy is 0.156250, time 56.673 samples/sec
+epoch 5, total_step 85700, total loss is 19.49 , inference loss is 10.84, weight deacy loss is 8.65, training accuracy is 0.250000, time 56.864 samples/sec
+epoch 5, total_step 85720, total loss is 23.02 , inference loss is 14.37, weight deacy loss is 8.65, training accuracy is 0.406250, time 56.109 samples/sec
+epoch 5, total_step 85740, total loss is 16.94 , inference loss is 8.29, weight deacy loss is 8.65, training accuracy is 0.312500, time 56.649 samples/sec
+epoch 5, total_step 85760, total loss is 21.67 , inference loss is 13.02, weight deacy loss is 8.65, training accuracy is 0.218750, time 56.382 samples/sec
+epoch 5, total_step 85780, total loss is 17.31 , inference loss is 8.67, weight deacy loss is 8.65, training accuracy is 0.312500, time 56.330 samples/sec
+epoch 5, total_step 85800, total loss is 16.84 , inference loss is 8.19, weight deacy loss is 8.65, training accuracy is 0.375000, time 56.955 samples/sec
+epoch 5, total_step 85820, total loss is 18.07 , inference loss is 9.42, weight deacy loss is 8.65, training accuracy is 0.312500, time 56.592 samples/sec
+epoch 5, total_step 85840, total loss is 18.09 , inference loss is 9.45, weight deacy loss is 8.64, training accuracy is 0.375000, time 56.513 samples/sec
+epoch 5, total_step 85860, total loss is 20.33 , inference loss is 11.69, weight deacy loss is 8.64, training accuracy is 0.312500, time 56.559 samples/sec
+epoch 5, total_step 85880, total loss is 19.77 , inference loss is 11.13, weight deacy loss is 8.64, training accuracy is 0.343750, time 56.777 samples/sec
+epoch 5, total_step 85900, total loss is 18.44 , inference loss is 9.80, weight deacy loss is 8.64, training accuracy is 0.281250, time 56.219 samples/sec
+epoch 5, total_step 85920, total loss is 21.37 , inference loss is 12.73, weight deacy loss is 8.64, training accuracy is 0.250000, time 56.411 samples/sec
+epoch 5, total_step 85940, total loss is 19.90 , inference loss is 11.26, weight deacy loss is 8.64, training accuracy is 0.312500, time 56.014 samples/sec
+epoch 5, total_step 85960, total loss is 20.39 , inference loss is 11.75, weight deacy loss is 8.64, training accuracy is 0.218750, time 56.630 samples/sec
+epoch 5, total_step 85980, total loss is 16.58 , inference loss is 7.94, weight deacy loss is 8.64, training accuracy is 0.312500, time 57.071 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.70988700000002
+best_threshold_index 115 0.9875925925925926
+best_threshold_index 115 0.9864814814814815
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9883333333333333
+best_threshold_index 115 0.9866666666666667
+best_threshold_index 115 0.9866666666666667
+best_threshold_index 115 0.987037037037037
+best_threshold_index 115 0.9862962962962963
+best_threshold_index 115 0.9855555555555555
+best_threshold_index 115 0.9862962962962963
+[lfw][86000]XNorm: 21.668538
+[lfw][86000]Accuracy-Flip: 0.98683+-0.00673
+testing verification..
+(14000, 512)
+infer time 84.20050200000006
+best_threshold_index 122 0.9866666666666667
+best_threshold_index 119 0.984920634920635
+best_threshold_index 121 0.9850793650793651
+best_threshold_index 119 0.9844444444444445
+best_threshold_index 122 0.9855555555555555
+best_threshold_index 122 0.9850793650793651
+best_threshold_index 122 0.9853968253968254
+best_threshold_index 122 0.9850793650793651
+best_threshold_index 122 0.9850793650793651
+best_threshold_index 122 0.9857142857142858
+[cfp_ff][86000]XNorm: 20.849141
+[cfp_ff][86000]Accuracy-Flip: 0.98443+-0.00449
+testing verification..
+(14000, 512)
+infer time 85.79586100000002
+best_threshold_index 141 0.9158730158730158
+best_threshold_index 140 0.9131746031746032
+best_threshold_index 141 0.9131746031746032
+best_threshold_index 140 0.9152380952380952
+best_threshold_index 143 0.9180952380952381
+best_threshold_index 141 0.9136507936507936
+best_threshold_index 140 0.9128571428571428
+best_threshold_index 140 0.913968253968254
+best_threshold_index 141 0.9142857142857143
+best_threshold_index 141 0.9152380952380952
+[cfp_fp][86000]XNorm: 20.938587
+[cfp_fp][86000]Accuracy-Flip: 0.91243+-0.01410
+testing verification..
+(12000, 512)
+infer time 73.73338600000002
+best_threshold_index 138 0.8894444444444445
+best_threshold_index 131 0.8861111111111111
+best_threshold_index 136 0.8888888888888888
+best_threshold_index 135 0.8888888888888888
+best_threshold_index 135 0.892962962962963
+best_threshold_index 138 0.8887037037037037
+best_threshold_index 139 0.8885185185185185
+best_threshold_index 136 0.8959259259259259
+best_threshold_index 135 0.8903703703703704
+best_threshold_index 135 0.8909259259259259
+[agedb_30][86000]XNorm: 21.608579
+[agedb_30][86000]Accuracy-Flip: 0.88567+-0.02095
+test accuracy is: 0.9868333333333335
+epoch 5, total_step 86000, total loss is 21.00 , inference loss is 12.36, weight deacy loss is 8.64, training accuracy is 0.250000, time 52.255 samples/sec
+epoch 5, total_step 86020, total loss is 17.37 , inference loss is 8.73, weight deacy loss is 8.64, training accuracy is 0.343750, time 56.127 samples/sec
+epoch 5, total_step 86040, total loss is 19.17 , inference loss is 10.54, weight deacy loss is 8.64, training accuracy is 0.125000, time 56.857 samples/sec
+epoch 5, total_step 86060, total loss is 20.20 , inference loss is 11.57, weight deacy loss is 8.63, training accuracy is 0.250000, time 56.328 samples/sec
+epoch 5, total_step 86080, total loss is 19.26 , inference loss is 10.63, weight deacy loss is 8.63, training accuracy is 0.312500, time 56.956 samples/sec
+epoch 5, total_step 86100, total loss is 17.01 , inference loss is 8.38, weight deacy loss is 8.63, training accuracy is 0.437500, time 56.761 samples/sec
+epoch 5, total_step 86120, total loss is 15.29 , inference loss is 6.66, weight deacy loss is 8.63, training accuracy is 0.375000, time 56.960 samples/sec
+epoch 5, total_step 86140, total loss is 22.86 , inference loss is 14.23, weight deacy loss is 8.63, training accuracy is 0.218750, time 55.696 samples/sec
+epoch 5, total_step 86160, total loss is 19.47 , inference loss is 10.84, weight deacy loss is 8.63, training accuracy is 0.187500, time 56.930 samples/sec
+epoch 5, total_step 86180, total loss is 18.74 , inference loss is 10.11, weight deacy loss is 8.63, training accuracy is 0.312500, time 57.438 samples/sec
+epoch 5, total_step 86200, total loss is 18.65 , inference loss is 10.02, weight deacy loss is 8.63, training accuracy is 0.312500, time 56.361 samples/sec
+epoch 5, total_step 86220, total loss is 20.84 , inference loss is 12.21, weight deacy loss is 8.63, training accuracy is 0.187500, time 56.174 samples/sec
+epoch 5, total_step 86240, total loss is 17.23 , inference loss is 8.60, weight deacy loss is 8.63, training accuracy is 0.250000, time 57.191 samples/sec
+epoch 5, total_step 86260, total loss is 19.70 , inference loss is 11.08, weight deacy loss is 8.63, training accuracy is 0.250000, time 55.117 samples/sec
+epoch 5, total_step 86280, total loss is 21.48 , inference loss is 12.85, weight deacy loss is 8.63, training accuracy is 0.250000, time 56.192 samples/sec
+epoch 5, total_step 86300, total loss is 20.75 , inference loss is 12.13, weight deacy loss is 8.62, training accuracy is 0.312500, time 56.765 samples/sec
+epoch 5, total_step 86320, total loss is 17.33 , inference loss is 8.71, weight deacy loss is 8.62, training accuracy is 0.406250, time 56.749 samples/sec
+epoch 5, total_step 86340, total loss is 17.30 , inference loss is 8.68, weight deacy loss is 8.62, training accuracy is 0.500000, time 57.489 samples/sec
+epoch 5, total_step 86360, total loss is 18.05 , inference loss is 9.43, weight deacy loss is 8.62, training accuracy is 0.281250, time 54.943 samples/sec
+epoch 5, total_step 86380, total loss is 23.05 , inference loss is 14.43, weight deacy loss is 8.62, training accuracy is 0.281250, time 56.499 samples/sec
+epoch 5, total_step 86400, total loss is 16.60 , inference loss is 7.98, weight deacy loss is 8.62, training accuracy is 0.406250, time 55.904 samples/sec
+epoch 5, total_step 86420, total loss is 22.30 , inference loss is 13.68, weight deacy loss is 8.62, training accuracy is 0.343750, time 56.748 samples/sec
+epoch 5, total_step 86440, total loss is 19.70 , inference loss is 11.08, weight deacy loss is 8.62, training accuracy is 0.343750, time 56.440 samples/sec
+epoch 5, total_step 86460, total loss is 20.98 , inference loss is 12.36, weight deacy loss is 8.62, training accuracy is 0.187500, time 55.992 samples/sec
+epoch 5, total_step 86480, total loss is 17.93 , inference loss is 9.31, weight deacy loss is 8.62, training accuracy is 0.312500, time 56.064 samples/sec
+epoch 5, total_step 86500, total loss is 20.86 , inference loss is 12.25, weight deacy loss is 8.62, training accuracy is 0.343750, time 56.569 samples/sec
+epoch 5, total_step 86520, total loss is 20.31 , inference loss is 11.69, weight deacy loss is 8.61, training accuracy is 0.125000, time 55.773 samples/sec
+epoch 5, total_step 86540, total loss is 23.09 , inference loss is 14.48, weight deacy loss is 8.61, training accuracy is 0.218750, time 56.133 samples/sec
+epoch 5, total_step 86560, total loss is 19.09 , inference loss is 10.48, weight deacy loss is 8.61, training accuracy is 0.281250, time 57.225 samples/sec
+epoch 5, total_step 86580, total loss is 20.15 , inference loss is 11.54, weight deacy loss is 8.61, training accuracy is 0.187500, time 55.884 samples/sec
+epoch 5, total_step 86600, total loss is 17.43 , inference loss is 8.82, weight deacy loss is 8.61, training accuracy is 0.343750, time 56.551 samples/sec
+epoch 5, total_step 86620, total loss is 20.72 , inference loss is 12.11, weight deacy loss is 8.61, training accuracy is 0.250000, time 55.597 samples/sec
+epoch 5, total_step 86640, total loss is 19.62 , inference loss is 11.01, weight deacy loss is 8.61, training accuracy is 0.250000, time 56.616 samples/sec
+epoch 5, total_step 86660, total loss is 19.67 , inference loss is 11.06, weight deacy loss is 8.61, training accuracy is 0.250000, time 56.564 samples/sec
+epoch 5, total_step 86680, total loss is 16.32 , inference loss is 7.71, weight deacy loss is 8.61, training accuracy is 0.375000, time 55.770 samples/sec
+epoch 5, total_step 86700, total loss is 20.35 , inference loss is 11.74, weight deacy loss is 8.61, training accuracy is 0.281250, time 56.873 samples/sec
+epoch 5, total_step 86720, total loss is 23.22 , inference loss is 14.62, weight deacy loss is 8.61, training accuracy is 0.156250, time 56.913 samples/sec
+epoch 5, total_step 86740, total loss is 21.61 , inference loss is 13.01, weight deacy loss is 8.61, training accuracy is 0.218750, time 57.097 samples/sec
+epoch 5, total_step 86760, total loss is 16.61 , inference loss is 8.00, weight deacy loss is 8.60, training accuracy is 0.375000, time 56.768 samples/sec
+epoch 5, total_step 86780, total loss is 21.36 , inference loss is 12.75, weight deacy loss is 8.60, training accuracy is 0.312500, time 56.902 samples/sec
+epoch 5, total_step 86800, total loss is 22.45 , inference loss is 13.85, weight deacy loss is 8.60, training accuracy is 0.343750, time 56.521 samples/sec
+epoch 5, total_step 86820, total loss is 21.05 , inference loss is 12.45, weight deacy loss is 8.60, training accuracy is 0.281250, time 56.400 samples/sec
+epoch 5, total_step 86840, total loss is 21.27 , inference loss is 12.67, weight deacy loss is 8.60, training accuracy is 0.218750, time 56.436 samples/sec
+epoch 5, total_step 86860, total loss is 18.36 , inference loss is 9.76, weight deacy loss is 8.60, training accuracy is 0.250000, time 57.707 samples/sec
+epoch 5, total_step 86880, total loss is 22.40 , inference loss is 13.80, weight deacy loss is 8.60, training accuracy is 0.218750, time 56.741 samples/sec
+epoch 5, total_step 86900, total loss is 16.81 , inference loss is 8.22, weight deacy loss is 8.60, training accuracy is 0.281250, time 56.378 samples/sec
+epoch 5, total_step 86920, total loss is 19.88 , inference loss is 11.28, weight deacy loss is 8.60, training accuracy is 0.437500, time 57.416 samples/sec
+epoch 5, total_step 86940, total loss is 16.33 , inference loss is 7.74, weight deacy loss is 8.60, training accuracy is 0.312500, time 56.761 samples/sec
+epoch 5, total_step 86960, total loss is 16.52 , inference loss is 7.93, weight deacy loss is 8.60, training accuracy is 0.375000, time 57.108 samples/sec
+epoch 5, total_step 86980, total loss is 21.39 , inference loss is 12.79, weight deacy loss is 8.59, training accuracy is 0.218750, time 56.519 samples/sec
+epoch 5, total_step 87000, total loss is 23.85 , inference loss is 15.25, weight deacy loss is 8.59, training accuracy is 0.281250, time 56.296 samples/sec
+epoch 5, total_step 87020, total loss is 15.05 , inference loss is 6.45, weight deacy loss is 8.59, training accuracy is 0.500000, time 56.445 samples/sec
+epoch 5, total_step 87040, total loss is 19.33 , inference loss is 10.74, weight deacy loss is 8.59, training accuracy is 0.375000, time 56.807 samples/sec
+epoch 5, total_step 87060, total loss is 16.83 , inference loss is 8.24, weight deacy loss is 8.59, training accuracy is 0.343750, time 56.138 samples/sec
+epoch 5, total_step 87080, total loss is 21.00 , inference loss is 12.41, weight deacy loss is 8.59, training accuracy is 0.281250, time 56.975 samples/sec
+epoch 5, total_step 87100, total loss is 18.87 , inference loss is 10.28, weight deacy loss is 8.59, training accuracy is 0.312500, time 56.614 samples/sec
+epoch 5, total_step 87120, total loss is 14.67 , inference loss is 6.08, weight deacy loss is 8.59, training accuracy is 0.500000, time 56.117 samples/sec
+epoch 5, total_step 87140, total loss is 20.03 , inference loss is 11.44, weight deacy loss is 8.59, training accuracy is 0.281250, time 56.300 samples/sec
+epoch 5, total_step 87160, total loss is 20.17 , inference loss is 11.58, weight deacy loss is 8.59, training accuracy is 0.312500, time 56.679 samples/sec
+epoch 5, total_step 87180, total loss is 16.65 , inference loss is 8.07, weight deacy loss is 8.59, training accuracy is 0.468750, time 56.773 samples/sec
+epoch 5, total_step 87200, total loss is 17.32 , inference loss is 8.73, weight deacy loss is 8.58, training accuracy is 0.375000, time 56.750 samples/sec
+epoch 5, total_step 87220, total loss is 20.91 , inference loss is 12.33, weight deacy loss is 8.58, training accuracy is 0.187500, time 56.918 samples/sec
+epoch 5, total_step 87240, total loss is 18.19 , inference loss is 9.61, weight deacy loss is 8.58, training accuracy is 0.343750, time 57.221 samples/sec
+epoch 5, total_step 87260, total loss is 20.47 , inference loss is 11.89, weight deacy loss is 8.58, training accuracy is 0.312500, time 55.109 samples/sec
+epoch 5, total_step 87280, total loss is 21.52 , inference loss is 12.94, weight deacy loss is 8.58, training accuracy is 0.093750, time 55.694 samples/sec
+epoch 5, total_step 87300, total loss is 15.19 , inference loss is 6.61, weight deacy loss is 8.58, training accuracy is 0.406250, time 56.668 samples/sec
+epoch 5, total_step 87320, total loss is 20.08 , inference loss is 11.50, weight deacy loss is 8.58, training accuracy is 0.250000, time 56.749 samples/sec
+epoch 5, total_step 87340, total loss is 18.71 , inference loss is 10.13, weight deacy loss is 8.58, training accuracy is 0.343750, time 56.143 samples/sec
+epoch 5, total_step 87360, total loss is 20.82 , inference loss is 12.24, weight deacy loss is 8.58, training accuracy is 0.187500, time 56.308 samples/sec
+epoch 5, total_step 87380, total loss is 20.47 , inference loss is 11.89, weight deacy loss is 8.58, training accuracy is 0.218750, time 56.199 samples/sec
+epoch 5, total_step 87400, total loss is 16.07 , inference loss is 7.49, weight deacy loss is 8.58, training accuracy is 0.281250, time 56.014 samples/sec
+epoch 5, total_step 87420, total loss is 19.89 , inference loss is 11.32, weight deacy loss is 8.58, training accuracy is 0.250000, time 56.440 samples/sec
+epoch 5, total_step 87440, total loss is 22.49 , inference loss is 13.92, weight deacy loss is 8.57, training accuracy is 0.343750, time 56.869 samples/sec
+epoch 5, total_step 87460, total loss is 21.29 , inference loss is 12.72, weight deacy loss is 8.57, training accuracy is 0.218750, time 56.882 samples/sec
+epoch 5, total_step 87480, total loss is 18.16 , inference loss is 9.59, weight deacy loss is 8.57, training accuracy is 0.343750, time 56.813 samples/sec
+epoch 5, total_step 87500, total loss is 19.35 , inference loss is 10.77, weight deacy loss is 8.57, training accuracy is 0.312500, time 55.916 samples/sec
+epoch 5, total_step 87520, total loss is 16.96 , inference loss is 8.39, weight deacy loss is 8.57, training accuracy is 0.375000, time 55.831 samples/sec
+epoch 5, total_step 87540, total loss is 20.60 , inference loss is 12.03, weight deacy loss is 8.57, training accuracy is 0.281250, time 56.106 samples/sec
+epoch 5, total_step 87560, total loss is 19.30 , inference loss is 10.73, weight deacy loss is 8.57, training accuracy is 0.343750, time 56.526 samples/sec
+epoch 5, total_step 87580, total loss is 24.23 , inference loss is 15.66, weight deacy loss is 8.57, training accuracy is 0.281250, time 56.644 samples/sec
+epoch 5, total_step 87600, total loss is 20.89 , inference loss is 12.32, weight deacy loss is 8.57, training accuracy is 0.312500, time 55.967 samples/sec
+epoch 5, total_step 87620, total loss is 19.55 , inference loss is 10.98, weight deacy loss is 8.57, training accuracy is 0.250000, time 56.759 samples/sec
+epoch 5, total_step 87640, total loss is 19.79 , inference loss is 11.23, weight deacy loss is 8.57, training accuracy is 0.250000, time 57.750 samples/sec
+epoch 5, total_step 87660, total loss is 20.88 , inference loss is 12.32, weight deacy loss is 8.57, training accuracy is 0.312500, time 55.921 samples/sec
+epoch 5, total_step 87680, total loss is 20.72 , inference loss is 12.16, weight deacy loss is 8.56, training accuracy is 0.218750, time 56.322 samples/sec
+epoch 5, total_step 87700, total loss is 18.13 , inference loss is 9.56, weight deacy loss is 8.56, training accuracy is 0.468750, time 56.015 samples/sec
+epoch 5, total_step 87720, total loss is 19.86 , inference loss is 11.30, weight deacy loss is 8.56, training accuracy is 0.250000, time 56.525 samples/sec
+epoch 5, total_step 87740, total loss is 18.27 , inference loss is 9.71, weight deacy loss is 8.56, training accuracy is 0.250000, time 56.602 samples/sec
+epoch 5, total_step 87760, total loss is 19.68 , inference loss is 11.12, weight deacy loss is 8.56, training accuracy is 0.281250, time 56.179 samples/sec
+epoch 5, total_step 87780, total loss is 23.60 , inference loss is 15.04, weight deacy loss is 8.56, training accuracy is 0.375000, time 56.719 samples/sec
+epoch 5, total_step 87800, total loss is 20.30 , inference loss is 11.74, weight deacy loss is 8.56, training accuracy is 0.312500, time 56.292 samples/sec
+epoch 5, total_step 87820, total loss is 16.13 , inference loss is 7.57, weight deacy loss is 8.56, training accuracy is 0.312500, time 56.885 samples/sec
+epoch 5, total_step 87840, total loss is 22.18 , inference loss is 13.62, weight deacy loss is 8.56, training accuracy is 0.218750, time 57.140 samples/sec
+epoch 5, total_step 87860, total loss is 18.78 , inference loss is 10.22, weight deacy loss is 8.56, training accuracy is 0.250000, time 57.188 samples/sec
+epoch 5, total_step 87880, total loss is 21.60 , inference loss is 13.04, weight deacy loss is 8.56, training accuracy is 0.312500, time 56.648 samples/sec
+epoch 5, total_step 87900, total loss is 21.30 , inference loss is 12.75, weight deacy loss is 8.56, training accuracy is 0.343750, time 55.585 samples/sec
+epoch 5, total_step 87920, total loss is 19.10 , inference loss is 10.55, weight deacy loss is 8.55, training accuracy is 0.250000, time 57.014 samples/sec
+epoch 5, total_step 87940, total loss is 20.72 , inference loss is 12.17, weight deacy loss is 8.55, training accuracy is 0.406250, time 56.465 samples/sec
+epoch 5, total_step 87960, total loss is 22.35 , inference loss is 13.80, weight deacy loss is 8.55, training accuracy is 0.281250, time 56.108 samples/sec
+epoch 5, total_step 87980, total loss is 16.32 , inference loss is 7.77, weight deacy loss is 8.55, training accuracy is 0.187500, time 56.306 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.07247100000002
+best_threshold_index 107 0.9835185185185186
+best_threshold_index 112 0.9822222222222222
+best_threshold_index 103 0.9833333333333333
+best_threshold_index 103 0.9838888888888889
+best_threshold_index 107 0.9833333333333333
+best_threshold_index 111 0.9816666666666667
+best_threshold_index 107 0.9825925925925926
+best_threshold_index 111 0.9814814814814815
+best_threshold_index 107 0.980925925925926
+best_threshold_index 107 0.9814814814814815
+[lfw][88000]XNorm: 22.557550
+[lfw][88000]Accuracy-Flip: 0.98017+-0.00938
+testing verification..
+(14000, 512)
+infer time 85.87088599999997
+best_threshold_index 115 0.9861904761904762
+best_threshold_index 113 0.9853968253968254
+best_threshold_index 113 0.9858730158730159
+best_threshold_index 113 0.9852380952380952
+best_threshold_index 113 0.9865079365079366
+best_threshold_index 111 0.9847619047619047
+best_threshold_index 116 0.9861904761904762
+best_threshold_index 111 0.984920634920635
+best_threshold_index 115 0.9852380952380952
+best_threshold_index 115 0.9865079365079366
+[cfp_ff][88000]XNorm: 22.479632
+[cfp_ff][88000]Accuracy-Flip: 0.98414+-0.00532
+testing verification..
+(14000, 512)
+infer time 85.96455299999997
+best_threshold_index 137 0.9125396825396825
+best_threshold_index 137 0.9122222222222223
+best_threshold_index 137 0.910952380952381
+best_threshold_index 137 0.9131746031746032
+best_threshold_index 137 0.9157142857142857
+best_threshold_index 137 0.9101587301587302
+best_threshold_index 137 0.9104761904761904
+best_threshold_index 133 0.9107936507936508
+best_threshold_index 137 0.912063492063492
+best_threshold_index 137 0.9133333333333333
+[cfp_fp][88000]XNorm: 21.938769
+[cfp_fp][88000]Accuracy-Flip: 0.91100+-0.01376
+testing verification..
+(12000, 512)
+infer time 73.55749800000002
+best_threshold_index 139 0.8901851851851852
+best_threshold_index 140 0.8887037037037037
+best_threshold_index 135 0.8877777777777778
+best_threshold_index 135 0.887962962962963
+best_threshold_index 135 0.8938888888888888
+best_threshold_index 135 0.8883333333333333
+best_threshold_index 135 0.8894444444444445
+best_threshold_index 140 0.8957407407407407
+best_threshold_index 140 0.8925925925925926
+best_threshold_index 140 0.8922222222222222
+[agedb_30][88000]XNorm: 22.916166
+[agedb_30][88000]Accuracy-Flip: 0.88617+-0.02255
+test accuracy is: 0.9801666666666666
+epoch 5, total_step 88000, total loss is 16.12 , inference loss is 7.57, weight deacy loss is 8.55, training accuracy is 0.343750, time 54.860 samples/sec
+epoch 5, total_step 88020, total loss is 20.00 , inference loss is 11.45, weight deacy loss is 8.55, training accuracy is 0.281250, time 56.559 samples/sec
+epoch 5, total_step 88040, total loss is 17.02 , inference loss is 8.47, weight deacy loss is 8.55, training accuracy is 0.281250, time 56.496 samples/sec
+epoch 5, total_step 88060, total loss is 18.27 , inference loss is 9.72, weight deacy loss is 8.55, training accuracy is 0.281250, time 55.905 samples/sec
+epoch 5, total_step 88080, total loss is 20.45 , inference loss is 11.90, weight deacy loss is 8.55, training accuracy is 0.187500, time 55.990 samples/sec
+epoch 5, total_step 88100, total loss is 17.01 , inference loss is 8.46, weight deacy loss is 8.55, training accuracy is 0.437500, time 56.684 samples/sec
+epoch 5, total_step 88120, total loss is 22.54 , inference loss is 13.99, weight deacy loss is 8.55, training accuracy is 0.281250, time 56.515 samples/sec
+epoch 5, total_step 88140, total loss is 17.78 , inference loss is 9.24, weight deacy loss is 8.55, training accuracy is 0.312500, time 56.839 samples/sec
+epoch 5, total_step 88160, total loss is 23.59 , inference loss is 15.04, weight deacy loss is 8.54, training accuracy is 0.312500, time 56.517 samples/sec
+epoch 5, total_step 88180, total loss is 19.02 , inference loss is 10.48, weight deacy loss is 8.54, training accuracy is 0.281250, time 56.636 samples/sec
+epoch 5, total_step 88200, total loss is 17.10 , inference loss is 8.56, weight deacy loss is 8.54, training accuracy is 0.437500, time 56.986 samples/sec
+epoch 5, total_step 88220, total loss is 19.54 , inference loss is 11.00, weight deacy loss is 8.54, training accuracy is 0.218750, time 56.577 samples/sec
+epoch 5, total_step 88240, total loss is 18.01 , inference loss is 9.47, weight deacy loss is 8.54, training accuracy is 0.218750, time 56.960 samples/sec
+epoch 5, total_step 88260, total loss is 20.77 , inference loss is 12.23, weight deacy loss is 8.54, training accuracy is 0.281250, time 56.245 samples/sec
+epoch 5, total_step 88280, total loss is 25.38 , inference loss is 16.84, weight deacy loss is 8.54, training accuracy is 0.156250, time 57.300 samples/sec
+epoch 5, total_step 88300, total loss is 18.22 , inference loss is 9.68, weight deacy loss is 8.54, training accuracy is 0.281250, time 56.882 samples/sec
+epoch 5, total_step 88320, total loss is 20.02 , inference loss is 11.48, weight deacy loss is 8.54, training accuracy is 0.187500, time 56.120 samples/sec
+epoch 5, total_step 88340, total loss is 18.02 , inference loss is 9.49, weight deacy loss is 8.54, training accuracy is 0.218750, time 56.770 samples/sec
+epoch 5, total_step 88360, total loss is 16.34 , inference loss is 7.80, weight deacy loss is 8.54, training accuracy is 0.312500, time 56.547 samples/sec
+epoch 5, total_step 88380, total loss is 20.88 , inference loss is 12.35, weight deacy loss is 8.53, training accuracy is 0.281250, time 56.418 samples/sec
+epoch 5, total_step 88400, total loss is 22.33 , inference loss is 13.80, weight deacy loss is 8.53, training accuracy is 0.156250, time 56.582 samples/sec
+epoch 5, total_step 88420, total loss is 23.07 , inference loss is 14.53, weight deacy loss is 8.53, training accuracy is 0.281250, time 56.874 samples/sec
+epoch 5, total_step 88440, total loss is 23.89 , inference loss is 15.35, weight deacy loss is 8.53, training accuracy is 0.156250, time 56.788 samples/sec
+epoch 5, total_step 88460, total loss is 20.80 , inference loss is 12.27, weight deacy loss is 8.53, training accuracy is 0.187500, time 56.246 samples/sec
+epoch 5, total_step 88480, total loss is 19.16 , inference loss is 10.63, weight deacy loss is 8.53, training accuracy is 0.250000, time 56.672 samples/sec
+epoch 5, total_step 88500, total loss is 21.81 , inference loss is 13.28, weight deacy loss is 8.53, training accuracy is 0.218750, time 56.649 samples/sec
+epoch 5, total_step 88520, total loss is 21.23 , inference loss is 12.70, weight deacy loss is 8.53, training accuracy is 0.250000, time 56.896 samples/sec
+epoch 5, total_step 88540, total loss is 14.75 , inference loss is 6.22, weight deacy loss is 8.53, training accuracy is 0.375000, time 56.068 samples/sec
+epoch 5, total_step 88560, total loss is 18.00 , inference loss is 9.47, weight deacy loss is 8.53, training accuracy is 0.218750, time 56.139 samples/sec
+epoch 5, total_step 88580, total loss is 16.63 , inference loss is 8.10, weight deacy loss is 8.53, training accuracy is 0.406250, time 55.361 samples/sec
+epoch 5, total_step 88600, total loss is 19.26 , inference loss is 10.73, weight deacy loss is 8.53, training accuracy is 0.218750, time 56.229 samples/sec
+epoch 5, total_step 88620, total loss is 19.37 , inference loss is 10.84, weight deacy loss is 8.52, training accuracy is 0.343750, time 57.049 samples/sec
+epoch 5, total_step 88640, total loss is 17.04 , inference loss is 8.52, weight deacy loss is 8.52, training accuracy is 0.312500, time 56.161 samples/sec
+epoch 5, total_step 88660, total loss is 24.82 , inference loss is 16.29, weight deacy loss is 8.52, training accuracy is 0.218750, time 56.612 samples/sec
+epoch 5, total_step 88680, total loss is 20.47 , inference loss is 11.95, weight deacy loss is 8.52, training accuracy is 0.281250, time 57.518 samples/sec
+epoch 5, total_step 88700, total loss is 19.27 , inference loss is 10.75, weight deacy loss is 8.52, training accuracy is 0.312500, time 56.203 samples/sec
+epoch 5, total_step 88720, total loss is 19.10 , inference loss is 10.58, weight deacy loss is 8.52, training accuracy is 0.187500, time 56.260 samples/sec
+epoch 5, total_step 88740, total loss is 16.93 , inference loss is 8.41, weight deacy loss is 8.52, training accuracy is 0.218750, time 56.642 samples/sec
+epoch 5, total_step 88760, total loss is 20.52 , inference loss is 12.00, weight deacy loss is 8.52, training accuracy is 0.218750, time 57.293 samples/sec
+epoch 5, total_step 88780, total loss is 18.73 , inference loss is 10.21, weight deacy loss is 8.52, training accuracy is 0.281250, time 55.470 samples/sec
+epoch 5, total_step 88800, total loss is 18.55 , inference loss is 10.04, weight deacy loss is 8.52, training accuracy is 0.281250, time 56.484 samples/sec
+epoch 5, total_step 88820, total loss is 21.15 , inference loss is 12.64, weight deacy loss is 8.52, training accuracy is 0.281250, time 56.065 samples/sec
+epoch 5, total_step 88840, total loss is 18.31 , inference loss is 9.80, weight deacy loss is 8.52, training accuracy is 0.250000, time 56.685 samples/sec
+epoch 5, total_step 88860, total loss is 19.63 , inference loss is 11.12, weight deacy loss is 8.51, training accuracy is 0.187500, time 56.396 samples/sec
+epoch 5, total_step 88880, total loss is 18.99 , inference loss is 10.47, weight deacy loss is 8.51, training accuracy is 0.250000, time 56.567 samples/sec
+epoch 5, total_step 88900, total loss is 22.57 , inference loss is 14.06, weight deacy loss is 8.51, training accuracy is 0.156250, time 57.165 samples/sec
+epoch 5, total_step 88920, total loss is 18.16 , inference loss is 9.65, weight deacy loss is 8.51, training accuracy is 0.281250, time 56.650 samples/sec
+epoch 5, total_step 88940, total loss is 17.95 , inference loss is 9.44, weight deacy loss is 8.51, training accuracy is 0.281250, time 56.740 samples/sec
+epoch 5, total_step 88960, total loss is 19.32 , inference loss is 10.81, weight deacy loss is 8.51, training accuracy is 0.125000, time 56.482 samples/sec
+epoch 5, total_step 88980, total loss is 17.36 , inference loss is 8.85, weight deacy loss is 8.51, training accuracy is 0.343750, time 56.980 samples/sec
+epoch 5, total_step 89000, total loss is 16.71 , inference loss is 8.20, weight deacy loss is 8.51, training accuracy is 0.281250, time 56.948 samples/sec
+epoch 5, total_step 89020, total loss is 21.41 , inference loss is 12.90, weight deacy loss is 8.51, training accuracy is 0.187500, time 57.415 samples/sec
+epoch 5, total_step 89040, total loss is 19.62 , inference loss is 11.11, weight deacy loss is 8.51, training accuracy is 0.250000, time 56.702 samples/sec
+epoch 5, total_step 89060, total loss is 19.79 , inference loss is 11.28, weight deacy loss is 8.51, training accuracy is 0.031250, time 56.420 samples/sec
+epoch 5, total_step 89080, total loss is 18.73 , inference loss is 10.23, weight deacy loss is 8.51, training accuracy is 0.187500, time 56.401 samples/sec
+epoch 5, total_step 89100, total loss is 17.71 , inference loss is 9.21, weight deacy loss is 8.50, training accuracy is 0.250000, time 56.524 samples/sec
+epoch 5, total_step 89120, total loss is 25.07 , inference loss is 16.57, weight deacy loss is 8.50, training accuracy is 0.093750, time 56.714 samples/sec
+epoch 5, total_step 89140, total loss is 20.90 , inference loss is 12.40, weight deacy loss is 8.50, training accuracy is 0.156250, time 56.337 samples/sec
+epoch 5, total_step 89160, total loss is 18.30 , inference loss is 9.80, weight deacy loss is 8.50, training accuracy is 0.281250, time 56.455 samples/sec
+epoch 5, total_step 89180, total loss is 20.33 , inference loss is 11.83, weight deacy loss is 8.50, training accuracy is 0.250000, time 54.286 samples/sec
+epoch 5, total_step 89200, total loss is 20.75 , inference loss is 12.25, weight deacy loss is 8.50, training accuracy is 0.218750, time 56.798 samples/sec
+epoch 5, total_step 89220, total loss is 17.66 , inference loss is 9.16, weight deacy loss is 8.50, training accuracy is 0.375000, time 56.683 samples/sec
+epoch 5, total_step 89240, total loss is 19.87 , inference loss is 11.37, weight deacy loss is 8.50, training accuracy is 0.218750, time 57.012 samples/sec
+epoch 5, total_step 89260, total loss is 18.86 , inference loss is 10.36, weight deacy loss is 8.50, training accuracy is 0.281250, time 56.341 samples/sec
+epoch 5, total_step 89280, total loss is 22.01 , inference loss is 13.51, weight deacy loss is 8.50, training accuracy is 0.156250, time 56.463 samples/sec
+epoch 5, total_step 89300, total loss is 19.71 , inference loss is 11.22, weight deacy loss is 8.50, training accuracy is 0.187500, time 56.219 samples/sec
+epoch 5, total_step 89320, total loss is 20.47 , inference loss is 11.97, weight deacy loss is 8.50, training accuracy is 0.218750, time 56.554 samples/sec
+epoch 5, total_step 89340, total loss is 16.82 , inference loss is 8.33, weight deacy loss is 8.49, training accuracy is 0.250000, time 56.654 samples/sec
+epoch 5, total_step 89360, total loss is 18.51 , inference loss is 10.02, weight deacy loss is 8.49, training accuracy is 0.281250, time 56.591 samples/sec
+epoch 5, total_step 89380, total loss is 22.74 , inference loss is 14.25, weight deacy loss is 8.49, training accuracy is 0.156250, time 56.825 samples/sec
+epoch 5, total_step 89400, total loss is 25.06 , inference loss is 16.57, weight deacy loss is 8.49, training accuracy is 0.125000, time 56.952 samples/sec
+epoch 5, total_step 89420, total loss is 18.48 , inference loss is 9.99, weight deacy loss is 8.49, training accuracy is 0.218750, time 56.388 samples/sec
+epoch 5, total_step 89440, total loss is 17.31 , inference loss is 8.82, weight deacy loss is 8.49, training accuracy is 0.312500, time 55.987 samples/sec
+epoch 5, total_step 89460, total loss is 21.35 , inference loss is 12.86, weight deacy loss is 8.49, training accuracy is 0.125000, time 56.399 samples/sec
+epoch 5, total_step 89480, total loss is 19.71 , inference loss is 11.22, weight deacy loss is 8.49, training accuracy is 0.187500, time 56.971 samples/sec
+epoch 5, total_step 89500, total loss is 14.87 , inference loss is 6.38, weight deacy loss is 8.49, training accuracy is 0.312500, time 56.745 samples/sec
+epoch 5, total_step 89520, total loss is 19.68 , inference loss is 11.19, weight deacy loss is 8.49, training accuracy is 0.187500, time 56.199 samples/sec
+epoch 5, total_step 89540, total loss is 20.54 , inference loss is 12.05, weight deacy loss is 8.49, training accuracy is 0.218750, time 57.033 samples/sec
+epoch 5, total_step 89560, total loss is 23.69 , inference loss is 15.20, weight deacy loss is 8.49, training accuracy is 0.281250, time 56.820 samples/sec
+epoch 5, total_step 89580, total loss is 18.17 , inference loss is 9.69, weight deacy loss is 8.48, training accuracy is 0.218750, time 55.750 samples/sec
+epoch 5, total_step 89600, total loss is 17.36 , inference loss is 8.88, weight deacy loss is 8.48, training accuracy is 0.343750, time 56.243 samples/sec
+epoch 5, total_step 89620, total loss is 20.49 , inference loss is 12.00, weight deacy loss is 8.48, training accuracy is 0.187500, time 56.708 samples/sec
+epoch 5, total_step 89640, total loss is 19.21 , inference loss is 10.73, weight deacy loss is 8.48, training accuracy is 0.218750, time 56.805 samples/sec
+epoch 5, total_step 89660, total loss is 20.87 , inference loss is 12.39, weight deacy loss is 8.48, training accuracy is 0.343750, time 56.234 samples/sec
+epoch 5, total_step 89680, total loss is 20.66 , inference loss is 12.18, weight deacy loss is 8.48, training accuracy is 0.375000, time 56.075 samples/sec
+epoch 5, total_step 89700, total loss is 20.49 , inference loss is 12.01, weight deacy loss is 8.48, training accuracy is 0.250000, time 56.629 samples/sec
+epoch 5, total_step 89720, total loss is 22.47 , inference loss is 13.99, weight deacy loss is 8.48, training accuracy is 0.187500, time 57.609 samples/sec
+epoch 5, total_step 89740, total loss is 27.51 , inference loss is 19.03, weight deacy loss is 8.48, training accuracy is 0.093750, time 55.101 samples/sec
+epoch 5, total_step 89760, total loss is 22.56 , inference loss is 14.09, weight deacy loss is 8.48, training accuracy is 0.062500, time 56.511 samples/sec
+epoch 5, total_step 89780, total loss is 21.09 , inference loss is 12.61, weight deacy loss is 8.48, training accuracy is 0.218750, time 55.996 samples/sec
+epoch 5, total_step 89800, total loss is 19.00 , inference loss is 10.52, weight deacy loss is 8.48, training accuracy is 0.281250, time 56.253 samples/sec
+epoch 5, total_step 89820, total loss is 24.45 , inference loss is 15.98, weight deacy loss is 8.47, training accuracy is 0.187500, time 57.337 samples/sec
+epoch 5, total_step 89840, total loss is 17.30 , inference loss is 8.82, weight deacy loss is 8.47, training accuracy is 0.343750, time 56.913 samples/sec
+epoch 5, total_step 89860, total loss is 20.93 , inference loss is 12.45, weight deacy loss is 8.47, training accuracy is 0.218750, time 56.962 samples/sec
+epoch 5, total_step 89880, total loss is 19.08 , inference loss is 10.61, weight deacy loss is 8.47, training accuracy is 0.218750, time 57.071 samples/sec
+epoch 5, total_step 89900, total loss is 22.87 , inference loss is 14.40, weight deacy loss is 8.47, training accuracy is 0.187500, time 56.830 samples/sec
+epoch 5, total_step 89920, total loss is 20.82 , inference loss is 12.35, weight deacy loss is 8.47, training accuracy is 0.187500, time 57.078 samples/sec
+epoch 5, total_step 89940, total loss is 17.43 , inference loss is 8.97, weight deacy loss is 8.47, training accuracy is 0.218750, time 56.716 samples/sec
+epoch 5, total_step 89960, total loss is 17.82 , inference loss is 9.35, weight deacy loss is 8.47, training accuracy is 0.281250, time 57.021 samples/sec
+epoch 5, total_step 89980, total loss is 17.58 , inference loss is 9.11, weight deacy loss is 8.47, training accuracy is 0.312500, time 56.876 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.61259500000003
+best_threshold_index 104 0.9829629629629629
+best_threshold_index 107 0.9816666666666667
+best_threshold_index 104 0.9822222222222222
+best_threshold_index 104 0.9833333333333333
+best_threshold_index 107 0.9827777777777778
+best_threshold_index 104 0.9812962962962963
+best_threshold_index 104 0.982037037037037
+best_threshold_index 104 0.9812962962962963
+best_threshold_index 107 0.980925925925926
+best_threshold_index 107 0.980925925925926
+[lfw][90000]XNorm: 23.559421
+[lfw][90000]Accuracy-Flip: 0.98083+-0.00739
+testing verification..
+(14000, 512)
+infer time 85.09515700000001
+best_threshold_index 112 0.9841269841269841
+best_threshold_index 112 0.983015873015873
+best_threshold_index 113 0.9825396825396825
+best_threshold_index 112 0.9825396825396825
+best_threshold_index 114 0.9846031746031746
+best_threshold_index 112 0.9817460317460317
+best_threshold_index 112 0.9831746031746031
+best_threshold_index 112 0.9820634920634921
+best_threshold_index 112 0.983015873015873
+best_threshold_index 112 0.9834920634920635
+[cfp_ff][90000]XNorm: 23.470100
+[cfp_ff][90000]Accuracy-Flip: 0.98214+-0.00796
+testing verification..
+(14000, 512)
+infer time 84.91222099999996
+best_threshold_index 132 0.9034920634920635
+best_threshold_index 132 0.9026984126984127
+best_threshold_index 132 0.9011111111111111
+best_threshold_index 132 0.9053968253968254
+best_threshold_index 132 0.9057142857142857
+best_threshold_index 132 0.9007936507936508
+best_threshold_index 132 0.9001587301587302
+best_threshold_index 132 0.9006349206349207
+best_threshold_index 132 0.903015873015873
+best_threshold_index 133 0.9017460317460317
+[cfp_fp][90000]XNorm: 22.973805
+[cfp_fp][90000]Accuracy-Flip: 0.90143+-0.01655
+testing verification..
+(12000, 512)
+infer time 72.64788899999996
+best_threshold_index 138 0.8870370370370371
+best_threshold_index 138 0.8846296296296297
+best_threshold_index 138 0.8848148148148148
+best_threshold_index 136 0.8862962962962962
+best_threshold_index 136 0.892962962962963
+best_threshold_index 138 0.8861111111111111
+best_threshold_index 138 0.8859259259259259
+best_threshold_index 136 0.8911111111111111
+best_threshold_index 138 0.8903703703703704
+best_threshold_index 136 0.8870370370370371
+[agedb_30][90000]XNorm: 23.372903
+[agedb_30][90000]Accuracy-Flip: 0.88433+-0.02563
+test accuracy is: 0.9808333333333333
+epoch 5, total_step 90000, total loss is 23.72 , inference loss is 15.25, weight deacy loss is 8.47, training accuracy is 0.156250, time 54.302 samples/sec
+epoch 5, total_step 90020, total loss is 24.92 , inference loss is 16.45, weight deacy loss is 8.47, training accuracy is 0.125000, time 56.735 samples/sec
+epoch 5, total_step 90040, total loss is 19.33 , inference loss is 10.87, weight deacy loss is 8.47, training accuracy is 0.281250, time 56.101 samples/sec
+epoch 5, total_step 90060, total loss is 21.20 , inference loss is 12.73, weight deacy loss is 8.46, training accuracy is 0.343750, time 56.337 samples/sec
+epoch 5, total_step 90080, total loss is 17.12 , inference loss is 8.65, weight deacy loss is 8.46, training accuracy is 0.343750, time 56.519 samples/sec
+epoch 5, total_step 90100, total loss is 20.46 , inference loss is 12.00, weight deacy loss is 8.46, training accuracy is 0.187500, time 56.489 samples/sec
+epoch 5, total_step 90120, total loss is 16.25 , inference loss is 7.79, weight deacy loss is 8.46, training accuracy is 0.312500, time 56.113 samples/sec
+epoch 5, total_step 90140, total loss is 20.94 , inference loss is 12.47, weight deacy loss is 8.46, training accuracy is 0.218750, time 56.386 samples/sec
+epoch 5, total_step 90160, total loss is 24.13 , inference loss is 15.67, weight deacy loss is 8.46, training accuracy is 0.125000, time 56.520 samples/sec
+epoch 5, total_step 90180, total loss is 20.14 , inference loss is 11.68, weight deacy loss is 8.46, training accuracy is 0.187500, time 55.860 samples/sec
+epoch 5, total_step 90200, total loss is 21.65 , inference loss is 13.19, weight deacy loss is 8.46, training accuracy is 0.250000, time 56.900 samples/sec
+epoch 5, total_step 90220, total loss is 20.65 , inference loss is 12.19, weight deacy loss is 8.46, training accuracy is 0.343750, time 56.827 samples/sec
+epoch 5, total_step 90240, total loss is 17.24 , inference loss is 8.78, weight deacy loss is 8.46, training accuracy is 0.187500, time 56.110 samples/sec
+epoch 5, total_step 90260, total loss is 20.27 , inference loss is 11.81, weight deacy loss is 8.46, training accuracy is 0.125000, time 56.880 samples/sec
+epoch 5, total_step 90280, total loss is 21.24 , inference loss is 12.78, weight deacy loss is 8.46, training accuracy is 0.187500, time 57.006 samples/sec
+epoch 5, total_step 90300, total loss is 19.23 , inference loss is 10.77, weight deacy loss is 8.45, training accuracy is 0.375000, time 57.061 samples/sec
+epoch 5, total_step 90320, total loss is 24.26 , inference loss is 15.81, weight deacy loss is 8.45, training accuracy is 0.250000, time 55.740 samples/sec
+epoch 5, total_step 90340, total loss is 19.45 , inference loss is 10.99, weight deacy loss is 8.45, training accuracy is 0.250000, time 56.765 samples/sec
+epoch 5, total_step 90360, total loss is 24.03 , inference loss is 15.58, weight deacy loss is 8.45, training accuracy is 0.218750, time 56.051 samples/sec
+epoch 5, total_step 90380, total loss is 16.22 , inference loss is 7.76, weight deacy loss is 8.45, training accuracy is 0.156250, time 56.839 samples/sec
+epoch 5, total_step 90400, total loss is 20.85 , inference loss is 12.40, weight deacy loss is 8.45, training accuracy is 0.218750, time 56.910 samples/sec
+epoch 5, total_step 90420, total loss is 19.58 , inference loss is 11.13, weight deacy loss is 8.45, training accuracy is 0.312500, time 56.311 samples/sec
+epoch 5, total_step 90440, total loss is 19.45 , inference loss is 11.00, weight deacy loss is 8.45, training accuracy is 0.250000, time 56.054 samples/sec
+epoch 5, total_step 90460, total loss is 21.84 , inference loss is 13.39, weight deacy loss is 8.45, training accuracy is 0.281250, time 56.502 samples/sec
+epoch 5, total_step 90480, total loss is 20.57 , inference loss is 12.13, weight deacy loss is 8.45, training accuracy is 0.281250, time 56.077 samples/sec
+epoch 5, total_step 90500, total loss is 16.42 , inference loss is 7.97, weight deacy loss is 8.45, training accuracy is 0.250000, time 56.582 samples/sec
+epoch 5, total_step 90520, total loss is 16.33 , inference loss is 7.89, weight deacy loss is 8.45, training accuracy is 0.406250, time 56.797 samples/sec
+epoch 5, total_step 90540, total loss is 21.57 , inference loss is 13.12, weight deacy loss is 8.44, training accuracy is 0.156250, time 56.073 samples/sec
+epoch 5, total_step 90560, total loss is 20.59 , inference loss is 12.15, weight deacy loss is 8.44, training accuracy is 0.187500, time 55.081 samples/sec
+epoch 5, total_step 90580, total loss is 19.23 , inference loss is 10.79, weight deacy loss is 8.44, training accuracy is 0.187500, time 56.631 samples/sec
+epoch 5, total_step 90600, total loss is 23.69 , inference loss is 15.25, weight deacy loss is 8.44, training accuracy is 0.250000, time 56.494 samples/sec
+epoch 5, total_step 90620, total loss is 19.05 , inference loss is 10.61, weight deacy loss is 8.44, training accuracy is 0.312500, time 56.835 samples/sec
+epoch 5, total_step 90640, total loss is 22.17 , inference loss is 13.73, weight deacy loss is 8.44, training accuracy is 0.156250, time 56.730 samples/sec
+epoch 5, total_step 90660, total loss is 20.12 , inference loss is 11.68, weight deacy loss is 8.44, training accuracy is 0.125000, time 56.828 samples/sec
+epoch 5, total_step 90680, total loss is 22.71 , inference loss is 14.27, weight deacy loss is 8.44, training accuracy is 0.156250, time 56.768 samples/sec
+epoch 5, total_step 90700, total loss is 22.69 , inference loss is 14.25, weight deacy loss is 8.44, training accuracy is 0.218750, time 56.217 samples/sec
+epoch 5, total_step 90720, total loss is 19.96 , inference loss is 11.53, weight deacy loss is 8.44, training accuracy is 0.125000, time 56.140 samples/sec
+epoch 5, total_step 90740, total loss is 22.90 , inference loss is 14.47, weight deacy loss is 8.44, training accuracy is 0.093750, time 56.656 samples/sec
+epoch 5, total_step 90760, total loss is 20.09 , inference loss is 11.65, weight deacy loss is 8.44, training accuracy is 0.375000, time 56.109 samples/sec
+epoch 5, total_step 90780, total loss is 20.44 , inference loss is 12.00, weight deacy loss is 8.43, training accuracy is 0.187500, time 56.376 samples/sec
+epoch 5, total_step 90800, total loss is 20.19 , inference loss is 11.76, weight deacy loss is 8.43, training accuracy is 0.187500, time 57.071 samples/sec
+epoch 5, total_step 90820, total loss is 18.38 , inference loss is 9.95, weight deacy loss is 8.43, training accuracy is 0.281250, time 56.641 samples/sec
+epoch 5, total_step 90840, total loss is 23.58 , inference loss is 15.14, weight deacy loss is 8.43, training accuracy is 0.125000, time 56.100 samples/sec
+epoch 5, total_step 90860, total loss is 19.70 , inference loss is 11.27, weight deacy loss is 8.43, training accuracy is 0.218750, time 56.636 samples/sec
+epoch 5, total_step 90880, total loss is 22.03 , inference loss is 13.60, weight deacy loss is 8.43, training accuracy is 0.156250, time 56.957 samples/sec
+epoch 5, total_step 90900, total loss is 22.57 , inference loss is 14.14, weight deacy loss is 8.43, training accuracy is 0.125000, time 56.172 samples/sec
+epoch 5, total_step 90920, total loss is 21.57 , inference loss is 13.14, weight deacy loss is 8.43, training accuracy is 0.218750, time 56.718 samples/sec
+epoch 5, total_step 90940, total loss is 19.16 , inference loss is 10.73, weight deacy loss is 8.43, training accuracy is 0.187500, time 56.351 samples/sec
+epoch 5, total_step 90960, total loss is 17.46 , inference loss is 9.03, weight deacy loss is 8.43, training accuracy is 0.281250, time 57.256 samples/sec
+epoch 5, total_step 90980, total loss is 24.28 , inference loss is 15.85, weight deacy loss is 8.43, training accuracy is 0.218750, time 56.639 samples/sec
+epoch 5, total_step 91000, total loss is 17.72 , inference loss is 9.30, weight deacy loss is 8.43, training accuracy is 0.250000, time 56.779 samples/sec
+epoch 5, total_step 91020, total loss is 18.05 , inference loss is 9.63, weight deacy loss is 8.42, training accuracy is 0.250000, time 56.874 samples/sec
+epoch 5, total_step 91040, total loss is 26.12 , inference loss is 17.70, weight deacy loss is 8.42, training accuracy is 0.156250, time 56.786 samples/sec
+epoch 5, total_step 91060, total loss is 22.66 , inference loss is 14.23, weight deacy loss is 8.42, training accuracy is 0.125000, time 55.967 samples/sec
+epoch 5, total_step 91080, total loss is 21.78 , inference loss is 13.35, weight deacy loss is 8.42, training accuracy is 0.218750, time 56.418 samples/sec
+epoch 5, total_step 91100, total loss is 17.98 , inference loss is 9.56, weight deacy loss is 8.42, training accuracy is 0.281250, time 56.729 samples/sec
+epoch 5, total_step 91120, total loss is 20.72 , inference loss is 12.29, weight deacy loss is 8.42, training accuracy is 0.187500, time 56.775 samples/sec
+epoch 5, total_step 91140, total loss is 15.45 , inference loss is 7.03, weight deacy loss is 8.42, training accuracy is 0.375000, time 57.400 samples/sec
+epoch 5, total_step 91160, total loss is 17.98 , inference loss is 9.56, weight deacy loss is 8.42, training accuracy is 0.281250, time 55.798 samples/sec
+epoch 5, total_step 91180, total loss is 19.47 , inference loss is 11.05, weight deacy loss is 8.42, training accuracy is 0.250000, time 56.496 samples/sec
+epoch 5, total_step 91200, total loss is 19.02 , inference loss is 10.60, weight deacy loss is 8.42, training accuracy is 0.218750, time 56.803 samples/sec
+epoch 5, total_step 91220, total loss is 23.11 , inference loss is 14.69, weight deacy loss is 8.42, training accuracy is 0.250000, time 56.212 samples/sec
+epoch 5, total_step 91240, total loss is 20.07 , inference loss is 11.65, weight deacy loss is 8.42, training accuracy is 0.125000, time 55.966 samples/sec
+epoch 5, total_step 91260, total loss is 22.89 , inference loss is 14.47, weight deacy loss is 8.41, training accuracy is 0.218750, time 56.440 samples/sec
+epoch 5, total_step 91280, total loss is 17.76 , inference loss is 9.35, weight deacy loss is 8.41, training accuracy is 0.218750, time 57.550 samples/sec
+epoch 5, total_step 91300, total loss is 18.52 , inference loss is 10.11, weight deacy loss is 8.41, training accuracy is 0.250000, time 56.216 samples/sec
+epoch 5, total_step 91320, total loss is 19.73 , inference loss is 11.32, weight deacy loss is 8.41, training accuracy is 0.343750, time 56.466 samples/sec
+epoch 5, total_step 91340, total loss is 23.49 , inference loss is 15.08, weight deacy loss is 8.41, training accuracy is 0.218750, time 56.742 samples/sec
+epoch 5, total_step 91360, total loss is 19.35 , inference loss is 10.94, weight deacy loss is 8.41, training accuracy is 0.281250, time 56.957 samples/sec
+epoch 5, total_step 91380, total loss is 21.99 , inference loss is 13.58, weight deacy loss is 8.41, training accuracy is 0.156250, time 56.723 samples/sec
+epoch 5, total_step 91400, total loss is 18.59 , inference loss is 10.18, weight deacy loss is 8.41, training accuracy is 0.156250, time 56.530 samples/sec
+epoch 5, total_step 91420, total loss is 20.65 , inference loss is 12.24, weight deacy loss is 8.41, training accuracy is 0.093750, time 56.850 samples/sec
+epoch 5, total_step 91440, total loss is 22.51 , inference loss is 14.10, weight deacy loss is 8.41, training accuracy is 0.125000, time 57.278 samples/sec
+epoch 5, total_step 91460, total loss is 20.11 , inference loss is 11.70, weight deacy loss is 8.41, training accuracy is 0.062500, time 56.140 samples/sec
+epoch 5, total_step 91480, total loss is 20.02 , inference loss is 11.61, weight deacy loss is 8.41, training accuracy is 0.125000, time 56.677 samples/sec
+epoch 5, total_step 91500, total loss is 19.73 , inference loss is 11.32, weight deacy loss is 8.41, training accuracy is 0.156250, time 56.317 samples/sec
+epoch 5, total_step 91520, total loss is 19.44 , inference loss is 11.03, weight deacy loss is 8.40, training accuracy is 0.156250, time 56.259 samples/sec
+epoch 5, total_step 91540, total loss is 19.29 , inference loss is 10.89, weight deacy loss is 8.40, training accuracy is 0.187500, time 56.913 samples/sec
+epoch 5, total_step 91560, total loss is 21.80 , inference loss is 13.39, weight deacy loss is 8.40, training accuracy is 0.250000, time 56.160 samples/sec
+epoch 5, total_step 91580, total loss is 21.04 , inference loss is 12.64, weight deacy loss is 8.40, training accuracy is 0.125000, time 56.663 samples/sec
+epoch 5, total_step 91600, total loss is 19.89 , inference loss is 11.49, weight deacy loss is 8.40, training accuracy is 0.250000, time 56.755 samples/sec
+epoch 5, total_step 91620, total loss is 26.56 , inference loss is 18.16, weight deacy loss is 8.40, training accuracy is 0.187500, time 56.350 samples/sec
+epoch 5, total_step 91640, total loss is 20.62 , inference loss is 12.22, weight deacy loss is 8.40, training accuracy is 0.093750, time 56.258 samples/sec
+epoch 5, total_step 91660, total loss is 24.79 , inference loss is 16.39, weight deacy loss is 8.40, training accuracy is 0.187500, time 56.662 samples/sec
+epoch 5, total_step 91680, total loss is 20.56 , inference loss is 12.16, weight deacy loss is 8.40, training accuracy is 0.187500, time 56.934 samples/sec
+epoch 5, total_step 91700, total loss is 22.11 , inference loss is 13.71, weight deacy loss is 8.40, training accuracy is 0.187500, time 56.541 samples/sec
+epoch 5, total_step 91720, total loss is 21.51 , inference loss is 13.11, weight deacy loss is 8.40, training accuracy is 0.218750, time 56.775 samples/sec
+epoch 5, total_step 91740, total loss is 20.05 , inference loss is 11.66, weight deacy loss is 8.40, training accuracy is 0.187500, time 56.331 samples/sec
+epoch 5, total_step 91760, total loss is 17.53 , inference loss is 9.14, weight deacy loss is 8.40, training accuracy is 0.218750, time 56.728 samples/sec
+epoch 5, total_step 91780, total loss is 17.17 , inference loss is 8.78, weight deacy loss is 8.39, training accuracy is 0.156250, time 56.744 samples/sec
+epoch 5, total_step 91800, total loss is 23.21 , inference loss is 14.82, weight deacy loss is 8.39, training accuracy is 0.250000, time 55.017 samples/sec
+epoch 5, total_step 91820, total loss is 22.06 , inference loss is 13.67, weight deacy loss is 8.39, training accuracy is 0.187500, time 55.623 samples/sec
+epoch 5, total_step 91840, total loss is 16.79 , inference loss is 8.40, weight deacy loss is 8.39, training accuracy is 0.218750, time 56.616 samples/sec
+epoch 5, total_step 91860, total loss is 20.11 , inference loss is 11.72, weight deacy loss is 8.39, training accuracy is 0.156250, time 56.990 samples/sec
+epoch 5, total_step 91880, total loss is 20.99 , inference loss is 12.60, weight deacy loss is 8.39, training accuracy is 0.281250, time 55.425 samples/sec
+epoch 5, total_step 91900, total loss is 19.39 , inference loss is 11.00, weight deacy loss is 8.39, training accuracy is 0.281250, time 57.508 samples/sec
+epoch 5, total_step 91920, total loss is 14.91 , inference loss is 6.52, weight deacy loss is 8.39, training accuracy is 0.218750, time 57.093 samples/sec
+epoch 5, total_step 91940, total loss is 23.86 , inference loss is 15.47, weight deacy loss is 8.39, training accuracy is 0.187500, time 56.494 samples/sec
+epoch 5, total_step 91960, total loss is 16.57 , inference loss is 8.18, weight deacy loss is 8.39, training accuracy is 0.375000, time 56.272 samples/sec
+epoch 5, total_step 91980, total loss is 25.86 , inference loss is 17.47, weight deacy loss is 8.39, training accuracy is 0.156250, time 56.755 samples/sec
+
+
+
+
+End of epoch 5
+testing verification..
+(12000, 512)
+infer time 72.66550200000007
+best_threshold_index 118 0.9881481481481481
+best_threshold_index 120 0.987037037037037
+best_threshold_index 120 0.9881481481481481
+best_threshold_index 120 0.9879629629629629
+best_threshold_index 120 0.9879629629629629
+best_threshold_index 120 0.9872222222222222
+best_threshold_index 119 0.9883333333333333
+best_threshold_index 120 0.9872222222222222
+best_threshold_index 120 0.9861111111111112
+best_threshold_index 119 0.9868518518518519
+[lfw][92000]XNorm: 22.143844
+[lfw][92000]Accuracy-Flip: 0.98683+-0.00660
+testing verification..
+(14000, 512)
+infer time 85.14384800000002
+best_threshold_index 124 0.986984126984127
+best_threshold_index 122 0.9866666666666667
+best_threshold_index 121 0.9873015873015873
+best_threshold_index 125 0.9868253968253968
+best_threshold_index 118 0.9863492063492063
+best_threshold_index 124 0.9857142857142858
+best_threshold_index 124 0.9876190476190476
+best_threshold_index 124 0.9863492063492063
+best_threshold_index 118 0.9863492063492063
+best_threshold_index 122 0.9873015873015873
+[cfp_ff][92000]XNorm: 21.888622
+[cfp_ff][92000]Accuracy-Flip: 0.98514+-0.00508
+testing verification..
+(14000, 512)
+infer time 85.28048800000008
+best_threshold_index 143 0.9161904761904762
+best_threshold_index 143 0.9153968253968254
+best_threshold_index 143 0.9144444444444444
+best_threshold_index 143 0.9174603174603174
+best_threshold_index 143 0.9177777777777778
+best_threshold_index 143 0.915079365079365
+best_threshold_index 143 0.9138095238095238
+best_threshold_index 143 0.9146031746031746
+best_threshold_index 143 0.9173015873015873
+best_threshold_index 143 0.9165079365079365
+[cfp_fp][92000]XNorm: 21.110546
+[cfp_fp][92000]Accuracy-Flip: 0.91586+-0.01189
+testing verification..
+(12000, 512)
+infer time 73.01945500000002
+best_threshold_index 145 0.8874074074074074
+best_threshold_index 145 0.8868518518518519
+best_threshold_index 145 0.8872222222222222
+best_threshold_index 145 0.8892592592592593
+best_threshold_index 145 0.8927777777777778
+best_threshold_index 145 0.8877777777777778
+best_threshold_index 145 0.8898148148148148
+best_threshold_index 145 0.8942592592592593
+best_threshold_index 145 0.8909259259259259
+best_threshold_index 145 0.8903703703703704
+[agedb_30][92000]XNorm: 22.115537
+[agedb_30][92000]Accuracy-Flip: 0.88967+-0.02122
+test accuracy is: 0.9868333333333335
+epoch 6, total_step 92000, total loss is 24.34 , inference loss is 15.96, weight deacy loss is 8.39, training accuracy is 0.031250, time 53.053 samples/sec
+epoch 6, total_step 92020, total loss is 19.74 , inference loss is 11.35, weight deacy loss is 8.38, training accuracy is 0.375000, time 56.559 samples/sec
+epoch 6, total_step 92040, total loss is 14.25 , inference loss is 5.86, weight deacy loss is 8.38, training accuracy is 0.531250, time 56.462 samples/sec
+epoch 6, total_step 92060, total loss is 17.91 , inference loss is 9.53, weight deacy loss is 8.38, training accuracy is 0.312500, time 56.345 samples/sec
+epoch 6, total_step 92080, total loss is 15.59 , inference loss is 7.21, weight deacy loss is 8.38, training accuracy is 0.468750, time 55.917 samples/sec
+epoch 6, total_step 92100, total loss is 14.90 , inference loss is 6.52, weight deacy loss is 8.38, training accuracy is 0.468750, time 56.204 samples/sec
+epoch 6, total_step 92120, total loss is 14.10 , inference loss is 5.72, weight deacy loss is 8.38, training accuracy is 0.468750, time 56.862 samples/sec
+epoch 6, total_step 92140, total loss is 12.99 , inference loss is 4.61, weight deacy loss is 8.38, training accuracy is 0.625000, time 57.022 samples/sec
+epoch 6, total_step 92160, total loss is 21.17 , inference loss is 12.79, weight deacy loss is 8.38, training accuracy is 0.406250, time 56.294 samples/sec
+epoch 6, total_step 92180, total loss is 15.46 , inference loss is 7.08, weight deacy loss is 8.38, training accuracy is 0.531250, time 56.240 samples/sec
+epoch 6, total_step 92200, total loss is 19.33 , inference loss is 10.95, weight deacy loss is 8.38, training accuracy is 0.531250, time 56.537 samples/sec
+epoch 6, total_step 92220, total loss is 15.36 , inference loss is 6.98, weight deacy loss is 8.38, training accuracy is 0.531250, time 56.389 samples/sec
+epoch 6, total_step 92240, total loss is 12.42 , inference loss is 4.05, weight deacy loss is 8.38, training accuracy is 0.687500, time 56.107 samples/sec
+epoch 6, total_step 92260, total loss is 17.50 , inference loss is 9.12, weight deacy loss is 8.37, training accuracy is 0.343750, time 56.417 samples/sec
+epoch 6, total_step 92280, total loss is 15.09 , inference loss is 6.72, weight deacy loss is 8.37, training accuracy is 0.531250, time 56.132 samples/sec
+epoch 6, total_step 92300, total loss is 13.14 , inference loss is 4.77, weight deacy loss is 8.37, training accuracy is 0.500000, time 56.243 samples/sec
+epoch 6, total_step 92320, total loss is 17.12 , inference loss is 8.75, weight deacy loss is 8.37, training accuracy is 0.468750, time 57.044 samples/sec
+epoch 6, total_step 92340, total loss is 16.66 , inference loss is 8.29, weight deacy loss is 8.37, training accuracy is 0.531250, time 56.517 samples/sec
+epoch 6, total_step 92360, total loss is 18.13 , inference loss is 9.76, weight deacy loss is 8.37, training accuracy is 0.437500, time 56.851 samples/sec
+epoch 6, total_step 92380, total loss is 15.86 , inference loss is 7.49, weight deacy loss is 8.37, training accuracy is 0.406250, time 56.141 samples/sec
+epoch 6, total_step 92400, total loss is 15.15 , inference loss is 6.78, weight deacy loss is 8.37, training accuracy is 0.531250, time 57.005 samples/sec
+epoch 6, total_step 92420, total loss is 17.29 , inference loss is 8.92, weight deacy loss is 8.37, training accuracy is 0.406250, time 55.797 samples/sec
+epoch 6, total_step 92440, total loss is 15.81 , inference loss is 7.44, weight deacy loss is 8.37, training accuracy is 0.406250, time 57.106 samples/sec
+epoch 6, total_step 92460, total loss is 15.85 , inference loss is 7.48, weight deacy loss is 8.37, training accuracy is 0.437500, time 56.294 samples/sec
+epoch 6, total_step 92480, total loss is 19.92 , inference loss is 11.55, weight deacy loss is 8.37, training accuracy is 0.218750, time 57.446 samples/sec
+epoch 6, total_step 92500, total loss is 15.43 , inference loss is 7.07, weight deacy loss is 8.36, training accuracy is 0.437500, time 56.801 samples/sec
+epoch 6, total_step 92520, total loss is 17.38 , inference loss is 9.01, weight deacy loss is 8.36, training accuracy is 0.312500, time 55.698 samples/sec
+epoch 6, total_step 92540, total loss is 16.49 , inference loss is 8.13, weight deacy loss is 8.36, training accuracy is 0.406250, time 56.763 samples/sec
+epoch 6, total_step 92560, total loss is 15.70 , inference loss is 7.34, weight deacy loss is 8.36, training accuracy is 0.500000, time 56.056 samples/sec
+epoch 6, total_step 92580, total loss is 20.12 , inference loss is 11.75, weight deacy loss is 8.36, training accuracy is 0.375000, time 54.673 samples/sec
+epoch 6, total_step 92600, total loss is 16.31 , inference loss is 7.95, weight deacy loss is 8.36, training accuracy is 0.531250, time 56.592 samples/sec
+epoch 6, total_step 92620, total loss is 16.61 , inference loss is 8.25, weight deacy loss is 8.36, training accuracy is 0.375000, time 56.743 samples/sec
+epoch 6, total_step 92640, total loss is 19.10 , inference loss is 10.74, weight deacy loss is 8.36, training accuracy is 0.375000, time 55.791 samples/sec
+epoch 6, total_step 92660, total loss is 18.38 , inference loss is 10.02, weight deacy loss is 8.36, training accuracy is 0.375000, time 55.917 samples/sec
+epoch 6, total_step 92680, total loss is 14.62 , inference loss is 6.26, weight deacy loss is 8.36, training accuracy is 0.531250, time 57.187 samples/sec
+epoch 6, total_step 92700, total loss is 20.81 , inference loss is 12.45, weight deacy loss is 8.36, training accuracy is 0.281250, time 55.972 samples/sec
+epoch 6, total_step 92720, total loss is 16.67 , inference loss is 8.31, weight deacy loss is 8.36, training accuracy is 0.437500, time 56.747 samples/sec
+epoch 6, total_step 92740, total loss is 19.08 , inference loss is 10.73, weight deacy loss is 8.36, training accuracy is 0.406250, time 55.594 samples/sec
+epoch 6, total_step 92760, total loss is 18.11 , inference loss is 9.75, weight deacy loss is 8.36, training accuracy is 0.437500, time 55.930 samples/sec
+epoch 6, total_step 92780, total loss is 18.87 , inference loss is 10.51, weight deacy loss is 8.36, training accuracy is 0.406250, time 56.628 samples/sec
+epoch 6, total_step 92800, total loss is 15.00 , inference loss is 6.65, weight deacy loss is 8.35, training accuracy is 0.437500, time 57.182 samples/sec
+epoch 6, total_step 92820, total loss is 21.00 , inference loss is 12.64, weight deacy loss is 8.35, training accuracy is 0.375000, time 56.385 samples/sec
+epoch 6, total_step 92840, total loss is 15.92 , inference loss is 7.57, weight deacy loss is 8.35, training accuracy is 0.406250, time 55.838 samples/sec
+epoch 6, total_step 92860, total loss is 17.05 , inference loss is 8.70, weight deacy loss is 8.35, training accuracy is 0.375000, time 56.084 samples/sec
+epoch 6, total_step 92880, total loss is 15.49 , inference loss is 7.14, weight deacy loss is 8.35, training accuracy is 0.562500, time 55.093 samples/sec
+epoch 6, total_step 92900, total loss is 21.85 , inference loss is 13.50, weight deacy loss is 8.35, training accuracy is 0.343750, time 56.709 samples/sec
+epoch 6, total_step 92920, total loss is 23.13 , inference loss is 14.78, weight deacy loss is 8.35, training accuracy is 0.375000, time 57.157 samples/sec
+epoch 6, total_step 92940, total loss is 18.11 , inference loss is 9.76, weight deacy loss is 8.35, training accuracy is 0.437500, time 56.954 samples/sec
+epoch 6, total_step 92960, total loss is 17.86 , inference loss is 9.51, weight deacy loss is 8.35, training accuracy is 0.375000, time 56.342 samples/sec
+epoch 6, total_step 92980, total loss is 18.20 , inference loss is 9.85, weight deacy loss is 8.35, training accuracy is 0.343750, time 56.671 samples/sec
+epoch 6, total_step 93000, total loss is 15.92 , inference loss is 7.57, weight deacy loss is 8.35, training accuracy is 0.343750, time 56.297 samples/sec
+epoch 6, total_step 93020, total loss is 13.47 , inference loss is 5.13, weight deacy loss is 8.35, training accuracy is 0.531250, time 56.445 samples/sec
+epoch 6, total_step 93040, total loss is 19.11 , inference loss is 10.76, weight deacy loss is 8.35, training accuracy is 0.437500, time 57.170 samples/sec
+epoch 6, total_step 93060, total loss is 20.63 , inference loss is 12.28, weight deacy loss is 8.35, training accuracy is 0.500000, time 56.573 samples/sec
+epoch 6, total_step 93080, total loss is 20.36 , inference loss is 12.01, weight deacy loss is 8.35, training accuracy is 0.281250, time 57.427 samples/sec
+epoch 6, total_step 93100, total loss is 18.47 , inference loss is 10.13, weight deacy loss is 8.34, training accuracy is 0.406250, time 56.628 samples/sec
+epoch 6, total_step 93120, total loss is 17.58 , inference loss is 9.23, weight deacy loss is 8.34, training accuracy is 0.375000, time 56.583 samples/sec
+epoch 6, total_step 93140, total loss is 17.79 , inference loss is 9.45, weight deacy loss is 8.34, training accuracy is 0.375000, time 56.608 samples/sec
+epoch 6, total_step 93160, total loss is 16.16 , inference loss is 7.81, weight deacy loss is 8.34, training accuracy is 0.343750, time 57.345 samples/sec
+epoch 6, total_step 93180, total loss is 23.79 , inference loss is 15.45, weight deacy loss is 8.34, training accuracy is 0.250000, time 57.429 samples/sec
+epoch 6, total_step 93200, total loss is 20.64 , inference loss is 12.30, weight deacy loss is 8.34, training accuracy is 0.468750, time 56.298 samples/sec
+epoch 6, total_step 93220, total loss is 17.70 , inference loss is 9.36, weight deacy loss is 8.34, training accuracy is 0.437500, time 57.379 samples/sec
+epoch 6, total_step 93240, total loss is 21.21 , inference loss is 12.87, weight deacy loss is 8.34, training accuracy is 0.281250, time 56.614 samples/sec
+epoch 6, total_step 93260, total loss is 16.80 , inference loss is 8.46, weight deacy loss is 8.34, training accuracy is 0.531250, time 56.474 samples/sec
+epoch 6, total_step 93280, total loss is 19.87 , inference loss is 11.53, weight deacy loss is 8.34, training accuracy is 0.250000, time 56.561 samples/sec
+epoch 6, total_step 93300, total loss is 15.26 , inference loss is 6.93, weight deacy loss is 8.34, training accuracy is 0.562500, time 55.984 samples/sec
+epoch 6, total_step 93320, total loss is 15.58 , inference loss is 7.25, weight deacy loss is 8.34, training accuracy is 0.437500, time 56.747 samples/sec
+epoch 6, total_step 93340, total loss is 13.63 , inference loss is 5.29, weight deacy loss is 8.34, training accuracy is 0.562500, time 56.342 samples/sec
+epoch 6, total_step 93360, total loss is 17.88 , inference loss is 9.54, weight deacy loss is 8.34, training accuracy is 0.468750, time 56.067 samples/sec
+epoch 6, total_step 93380, total loss is 19.04 , inference loss is 10.70, weight deacy loss is 8.33, training accuracy is 0.406250, time 57.034 samples/sec
+epoch 6, total_step 93400, total loss is 18.21 , inference loss is 9.87, weight deacy loss is 8.33, training accuracy is 0.406250, time 57.273 samples/sec
+epoch 6, total_step 93420, total loss is 21.43 , inference loss is 13.10, weight deacy loss is 8.33, training accuracy is 0.312500, time 56.715 samples/sec
+epoch 6, total_step 93440, total loss is 18.07 , inference loss is 9.73, weight deacy loss is 8.33, training accuracy is 0.406250, time 57.245 samples/sec
+epoch 6, total_step 93460, total loss is 18.16 , inference loss is 9.83, weight deacy loss is 8.33, training accuracy is 0.500000, time 55.871 samples/sec
+epoch 6, total_step 93480, total loss is 17.91 , inference loss is 9.57, weight deacy loss is 8.33, training accuracy is 0.375000, time 56.939 samples/sec
+epoch 6, total_step 93500, total loss is 17.67 , inference loss is 9.34, weight deacy loss is 8.33, training accuracy is 0.343750, time 55.861 samples/sec
+epoch 6, total_step 93520, total loss is 17.07 , inference loss is 8.74, weight deacy loss is 8.33, training accuracy is 0.250000, time 56.701 samples/sec
+epoch 6, total_step 93540, total loss is 22.24 , inference loss is 13.91, weight deacy loss is 8.33, training accuracy is 0.218750, time 56.287 samples/sec
+epoch 6, total_step 93560, total loss is 16.68 , inference loss is 8.35, weight deacy loss is 8.33, training accuracy is 0.500000, time 57.273 samples/sec
+epoch 6, total_step 93580, total loss is 16.17 , inference loss is 7.84, weight deacy loss is 8.33, training accuracy is 0.218750, time 56.135 samples/sec
+epoch 6, total_step 93600, total loss is 16.02 , inference loss is 7.69, weight deacy loss is 8.33, training accuracy is 0.406250, time 56.289 samples/sec
+epoch 6, total_step 93620, total loss is 18.40 , inference loss is 10.08, weight deacy loss is 8.33, training accuracy is 0.375000, time 56.791 samples/sec
+epoch 6, total_step 93640, total loss is 17.76 , inference loss is 9.43, weight deacy loss is 8.33, training accuracy is 0.375000, time 56.726 samples/sec
+epoch 6, total_step 93660, total loss is 16.07 , inference loss is 7.74, weight deacy loss is 8.32, training accuracy is 0.406250, time 56.984 samples/sec
+epoch 6, total_step 93680, total loss is 16.46 , inference loss is 8.14, weight deacy loss is 8.32, training accuracy is 0.375000, time 56.224 samples/sec
+epoch 6, total_step 93700, total loss is 18.33 , inference loss is 10.00, weight deacy loss is 8.32, training accuracy is 0.531250, time 57.374 samples/sec
+epoch 6, total_step 93720, total loss is 19.62 , inference loss is 11.30, weight deacy loss is 8.32, training accuracy is 0.250000, time 56.849 samples/sec
+epoch 6, total_step 93740, total loss is 18.67 , inference loss is 10.34, weight deacy loss is 8.32, training accuracy is 0.281250, time 56.374 samples/sec
+epoch 6, total_step 93760, total loss is 18.92 , inference loss is 10.60, weight deacy loss is 8.32, training accuracy is 0.312500, time 57.017 samples/sec
+epoch 6, total_step 93780, total loss is 16.89 , inference loss is 8.57, weight deacy loss is 8.32, training accuracy is 0.218750, time 56.944 samples/sec
+epoch 6, total_step 93800, total loss is 15.81 , inference loss is 7.49, weight deacy loss is 8.32, training accuracy is 0.406250, time 56.591 samples/sec
+epoch 6, total_step 93820, total loss is 16.52 , inference loss is 8.20, weight deacy loss is 8.32, training accuracy is 0.250000, time 56.045 samples/sec
+epoch 6, total_step 93840, total loss is 21.44 , inference loss is 13.12, weight deacy loss is 8.32, training accuracy is 0.437500, time 57.700 samples/sec
+epoch 6, total_step 93860, total loss is 15.85 , inference loss is 7.53, weight deacy loss is 8.32, training accuracy is 0.468750, time 55.668 samples/sec
+epoch 6, total_step 93880, total loss is 22.07 , inference loss is 13.76, weight deacy loss is 8.32, training accuracy is 0.312500, time 56.687 samples/sec
+epoch 6, total_step 93900, total loss is 19.87 , inference loss is 11.56, weight deacy loss is 8.32, training accuracy is 0.312500, time 55.807 samples/sec
+epoch 6, total_step 93920, total loss is 15.98 , inference loss is 7.67, weight deacy loss is 8.32, training accuracy is 0.312500, time 56.533 samples/sec
+epoch 6, total_step 93940, total loss is 18.65 , inference loss is 10.33, weight deacy loss is 8.31, training accuracy is 0.312500, time 56.883 samples/sec
+epoch 6, total_step 93960, total loss is 13.91 , inference loss is 5.60, weight deacy loss is 8.31, training accuracy is 0.468750, time 55.772 samples/sec
+epoch 6, total_step 93980, total loss is 18.02 , inference loss is 9.71, weight deacy loss is 8.31, training accuracy is 0.281250, time 56.283 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.96029299999995
+best_threshold_index 119 0.9868518518518519
+best_threshold_index 119 0.9851851851851852
+best_threshold_index 116 0.9857407407407407
+best_threshold_index 110 0.9861111111111112
+best_threshold_index 119 0.9855555555555555
+best_threshold_index 119 0.9853703703703703
+best_threshold_index 111 0.9855555555555555
+best_threshold_index 111 0.9853703703703703
+best_threshold_index 119 0.985
+best_threshold_index 116 0.9844444444444445
+[lfw][94000]XNorm: 21.486613
+[lfw][94000]Accuracy-Flip: 0.98300+-0.00591
+testing verification..
+(14000, 512)
+infer time 85.42720699999994
+best_threshold_index 117 0.9857142857142858
+best_threshold_index 117 0.9842857142857143
+best_threshold_index 117 0.9842857142857143
+best_threshold_index 117 0.9834920634920635
+best_threshold_index 117 0.9842857142857143
+best_threshold_index 117 0.9834920634920635
+best_threshold_index 117 0.984920634920635
+best_threshold_index 117 0.9838095238095238
+best_threshold_index 117 0.9838095238095238
+best_threshold_index 117 0.9847619047619047
+[cfp_ff][94000]XNorm: 20.619378
+[cfp_ff][94000]Accuracy-Flip: 0.98429+-0.00596
+testing verification..
+(14000, 512)
+infer time 85.26136199999999
+best_threshold_index 131 0.906984126984127
+best_threshold_index 133 0.9074603174603174
+best_threshold_index 133 0.903968253968254
+best_threshold_index 133 0.9076190476190477
+best_threshold_index 133 0.9082539682539682
+best_threshold_index 133 0.9034920634920635
+best_threshold_index 133 0.903968253968254
+best_threshold_index 133 0.9038095238095238
+best_threshold_index 133 0.9049206349206349
+best_threshold_index 133 0.9066666666666666
+[cfp_fp][94000]XNorm: 21.602207
+[cfp_fp][94000]Accuracy-Flip: 0.90514+-0.01633
+testing verification..
+(12000, 512)
+infer time 72.927316
+best_threshold_index 135 0.905
+best_threshold_index 135 0.9044444444444445
+best_threshold_index 135 0.9059259259259259
+best_threshold_index 135 0.9070370370370371
+best_threshold_index 135 0.9118518518518518
+best_threshold_index 135 0.9031481481481481
+best_threshold_index 135 0.9062962962962963
+best_threshold_index 135 0.9120370370370371
+best_threshold_index 135 0.9101851851851852
+best_threshold_index 135 0.9074074074074074
+[agedb_30][94000]XNorm: 22.173319
+[agedb_30][94000]Accuracy-Flip: 0.90733+-0.02626
+test accuracy is: 0.983
+epoch 6, total_step 94000, total loss is 16.23 , inference loss is 7.91, weight deacy loss is 8.31, training accuracy is 0.437500, time 53.231 samples/sec
+epoch 6, total_step 94020, total loss is 13.97 , inference loss is 5.66, weight deacy loss is 8.31, training accuracy is 0.531250, time 56.204 samples/sec
+epoch 6, total_step 94040, total loss is 15.58 , inference loss is 7.27, weight deacy loss is 8.31, training accuracy is 0.531250, time 56.777 samples/sec
+epoch 6, total_step 94060, total loss is 16.72 , inference loss is 8.41, weight deacy loss is 8.31, training accuracy is 0.375000, time 56.513 samples/sec
+epoch 6, total_step 94080, total loss is 16.51 , inference loss is 8.20, weight deacy loss is 8.31, training accuracy is 0.531250, time 56.438 samples/sec
+epoch 6, total_step 94100, total loss is 18.08 , inference loss is 9.77, weight deacy loss is 8.31, training accuracy is 0.406250, time 56.594 samples/sec
+epoch 6, total_step 94120, total loss is 20.66 , inference loss is 12.35, weight deacy loss is 8.31, training accuracy is 0.375000, time 56.410 samples/sec
+epoch 6, total_step 94140, total loss is 14.59 , inference loss is 6.28, weight deacy loss is 8.31, training accuracy is 0.562500, time 57.256 samples/sec
+epoch 6, total_step 94160, total loss is 15.06 , inference loss is 6.75, weight deacy loss is 8.31, training accuracy is 0.406250, time 57.203 samples/sec
+epoch 6, total_step 94180, total loss is 14.11 , inference loss is 5.81, weight deacy loss is 8.30, training accuracy is 0.562500, time 56.556 samples/sec
+epoch 6, total_step 94200, total loss is 19.48 , inference loss is 11.18, weight deacy loss is 8.30, training accuracy is 0.343750, time 56.905 samples/sec
+epoch 6, total_step 94220, total loss is 14.01 , inference loss is 5.71, weight deacy loss is 8.30, training accuracy is 0.437500, time 56.493 samples/sec
+epoch 6, total_step 94240, total loss is 17.61 , inference loss is 9.30, weight deacy loss is 8.30, training accuracy is 0.500000, time 56.370 samples/sec
+epoch 6, total_step 94260, total loss is 18.26 , inference loss is 9.96, weight deacy loss is 8.30, training accuracy is 0.406250, time 55.878 samples/sec
+epoch 6, total_step 94280, total loss is 15.74 , inference loss is 7.44, weight deacy loss is 8.30, training accuracy is 0.500000, time 56.523 samples/sec
+epoch 6, total_step 94300, total loss is 15.04 , inference loss is 6.74, weight deacy loss is 8.30, training accuracy is 0.406250, time 56.572 samples/sec
+epoch 6, total_step 94320, total loss is 15.96 , inference loss is 7.66, weight deacy loss is 8.30, training accuracy is 0.406250, time 56.889 samples/sec
+epoch 6, total_step 94340, total loss is 15.36 , inference loss is 7.06, weight deacy loss is 8.30, training accuracy is 0.500000, time 56.728 samples/sec
+epoch 6, total_step 94360, total loss is 14.47 , inference loss is 6.17, weight deacy loss is 8.30, training accuracy is 0.531250, time 56.723 samples/sec
+epoch 6, total_step 94380, total loss is 14.52 , inference loss is 6.23, weight deacy loss is 8.30, training accuracy is 0.468750, time 55.592 samples/sec
+epoch 6, total_step 94400, total loss is 13.22 , inference loss is 4.92, weight deacy loss is 8.30, training accuracy is 0.500000, time 56.325 samples/sec
+epoch 6, total_step 94420, total loss is 18.30 , inference loss is 10.01, weight deacy loss is 8.29, training accuracy is 0.375000, time 55.934 samples/sec
+epoch 6, total_step 94440, total loss is 16.74 , inference loss is 8.45, weight deacy loss is 8.29, training accuracy is 0.437500, time 55.005 samples/sec
+epoch 6, total_step 94460, total loss is 19.39 , inference loss is 11.10, weight deacy loss is 8.29, training accuracy is 0.406250, time 56.025 samples/sec
+epoch 6, total_step 94480, total loss is 16.25 , inference loss is 7.96, weight deacy loss is 8.29, training accuracy is 0.406250, time 56.987 samples/sec
+epoch 6, total_step 94500, total loss is 13.56 , inference loss is 5.27, weight deacy loss is 8.29, training accuracy is 0.531250, time 56.494 samples/sec
+epoch 6, total_step 94520, total loss is 17.82 , inference loss is 9.53, weight deacy loss is 8.29, training accuracy is 0.343750, time 57.366 samples/sec
+epoch 6, total_step 94540, total loss is 15.69 , inference loss is 7.40, weight deacy loss is 8.29, training accuracy is 0.406250, time 56.041 samples/sec
+epoch 6, total_step 94560, total loss is 18.68 , inference loss is 10.39, weight deacy loss is 8.29, training accuracy is 0.375000, time 56.451 samples/sec
+epoch 6, total_step 94580, total loss is 17.44 , inference loss is 9.15, weight deacy loss is 8.29, training accuracy is 0.375000, time 56.186 samples/sec
+epoch 6, total_step 94600, total loss is 14.14 , inference loss is 5.85, weight deacy loss is 8.29, training accuracy is 0.562500, time 55.850 samples/sec
+epoch 6, total_step 94620, total loss is 17.48 , inference loss is 9.19, weight deacy loss is 8.29, training accuracy is 0.406250, time 55.671 samples/sec
+epoch 6, total_step 94640, total loss is 16.85 , inference loss is 8.57, weight deacy loss is 8.29, training accuracy is 0.312500, time 56.672 samples/sec
+epoch 6, total_step 94660, total loss is 15.09 , inference loss is 6.80, weight deacy loss is 8.29, training accuracy is 0.531250, time 55.773 samples/sec
+epoch 6, total_step 94680, total loss is 14.94 , inference loss is 6.66, weight deacy loss is 8.28, training accuracy is 0.468750, time 55.720 samples/sec
+epoch 6, total_step 94700, total loss is 18.24 , inference loss is 9.96, weight deacy loss is 8.28, training accuracy is 0.250000, time 56.385 samples/sec
+epoch 6, total_step 94720, total loss is 12.31 , inference loss is 4.03, weight deacy loss is 8.28, training accuracy is 0.375000, time 56.406 samples/sec
+epoch 6, total_step 94740, total loss is 12.42 , inference loss is 4.14, weight deacy loss is 8.28, training accuracy is 0.468750, time 56.190 samples/sec
+epoch 6, total_step 94760, total loss is 15.59 , inference loss is 7.31, weight deacy loss is 8.28, training accuracy is 0.531250, time 56.696 samples/sec
+epoch 6, total_step 94780, total loss is 16.05 , inference loss is 7.77, weight deacy loss is 8.28, training accuracy is 0.437500, time 56.812 samples/sec
+epoch 6, total_step 94800, total loss is 15.94 , inference loss is 7.66, weight deacy loss is 8.28, training accuracy is 0.437500, time 56.018 samples/sec
+epoch 6, total_step 94820, total loss is 21.13 , inference loss is 12.85, weight deacy loss is 8.28, training accuracy is 0.375000, time 56.265 samples/sec
+epoch 6, total_step 94840, total loss is 15.56 , inference loss is 7.29, weight deacy loss is 8.28, training accuracy is 0.375000, time 56.808 samples/sec
+epoch 6, total_step 94860, total loss is 13.91 , inference loss is 5.63, weight deacy loss is 8.28, training accuracy is 0.437500, time 56.884 samples/sec
+epoch 6, total_step 94880, total loss is 18.27 , inference loss is 9.99, weight deacy loss is 8.28, training accuracy is 0.406250, time 56.859 samples/sec
+epoch 6, total_step 94900, total loss is 14.28 , inference loss is 6.01, weight deacy loss is 8.28, training accuracy is 0.500000, time 56.739 samples/sec
+epoch 6, total_step 94920, total loss is 14.71 , inference loss is 6.43, weight deacy loss is 8.27, training accuracy is 0.500000, time 56.478 samples/sec
+epoch 6, total_step 94940, total loss is 13.96 , inference loss is 5.69, weight deacy loss is 8.27, training accuracy is 0.468750, time 56.968 samples/sec
+epoch 6, total_step 94960, total loss is 13.76 , inference loss is 5.49, weight deacy loss is 8.27, training accuracy is 0.468750, time 56.602 samples/sec
+epoch 6, total_step 94980, total loss is 21.19 , inference loss is 12.92, weight deacy loss is 8.27, training accuracy is 0.375000, time 55.335 samples/sec
+epoch 6, total_step 95000, total loss is 13.01 , inference loss is 4.74, weight deacy loss is 8.27, training accuracy is 0.562500, time 55.884 samples/sec
+epoch 6, total_step 95020, total loss is 21.14 , inference loss is 12.87, weight deacy loss is 8.27, training accuracy is 0.312500, time 56.312 samples/sec
+epoch 6, total_step 95040, total loss is 15.66 , inference loss is 7.39, weight deacy loss is 8.27, training accuracy is 0.375000, time 56.186 samples/sec
+epoch 6, total_step 95060, total loss is 14.42 , inference loss is 6.15, weight deacy loss is 8.27, training accuracy is 0.500000, time 55.837 samples/sec
+epoch 6, total_step 95080, total loss is 18.81 , inference loss is 10.54, weight deacy loss is 8.27, training accuracy is 0.437500, time 55.902 samples/sec
+epoch 6, total_step 95100, total loss is 13.86 , inference loss is 5.59, weight deacy loss is 8.27, training accuracy is 0.437500, time 54.253 samples/sec
+epoch 6, total_step 95120, total loss is 17.72 , inference loss is 9.45, weight deacy loss is 8.27, training accuracy is 0.437500, time 57.111 samples/sec
+epoch 6, total_step 95140, total loss is 14.80 , inference loss is 6.53, weight deacy loss is 8.27, training accuracy is 0.593750, time 57.200 samples/sec
+epoch 6, total_step 95160, total loss is 12.83 , inference loss is 4.56, weight deacy loss is 8.27, training accuracy is 0.531250, time 56.375 samples/sec
+epoch 6, total_step 95180, total loss is 15.53 , inference loss is 7.27, weight deacy loss is 8.26, training accuracy is 0.562500, time 56.683 samples/sec
+epoch 6, total_step 95200, total loss is 20.75 , inference loss is 12.49, weight deacy loss is 8.26, training accuracy is 0.281250, time 56.728 samples/sec
+epoch 6, total_step 95220, total loss is 19.52 , inference loss is 11.25, weight deacy loss is 8.26, training accuracy is 0.406250, time 56.345 samples/sec
+epoch 6, total_step 95240, total loss is 16.34 , inference loss is 8.08, weight deacy loss is 8.26, training accuracy is 0.406250, time 56.616 samples/sec
+epoch 6, total_step 95260, total loss is 13.81 , inference loss is 5.55, weight deacy loss is 8.26, training accuracy is 0.406250, time 56.447 samples/sec
+epoch 6, total_step 95280, total loss is 18.60 , inference loss is 10.33, weight deacy loss is 8.26, training accuracy is 0.437500, time 55.747 samples/sec
+epoch 6, total_step 95300, total loss is 15.79 , inference loss is 7.53, weight deacy loss is 8.26, training accuracy is 0.437500, time 55.806 samples/sec
+epoch 6, total_step 95320, total loss is 15.14 , inference loss is 6.88, weight deacy loss is 8.26, training accuracy is 0.437500, time 56.659 samples/sec
+epoch 6, total_step 95340, total loss is 19.50 , inference loss is 11.25, weight deacy loss is 8.26, training accuracy is 0.312500, time 55.353 samples/sec
+epoch 6, total_step 95360, total loss is 13.36 , inference loss is 5.10, weight deacy loss is 8.26, training accuracy is 0.468750, time 57.366 samples/sec
+epoch 6, total_step 95380, total loss is 17.49 , inference loss is 9.23, weight deacy loss is 8.26, training accuracy is 0.375000, time 56.717 samples/sec
+epoch 6, total_step 95400, total loss is 16.47 , inference loss is 8.21, weight deacy loss is 8.26, training accuracy is 0.437500, time 56.693 samples/sec
+epoch 6, total_step 95420, total loss is 15.28 , inference loss is 7.03, weight deacy loss is 8.26, training accuracy is 0.375000, time 55.644 samples/sec
+epoch 6, total_step 95440, total loss is 24.64 , inference loss is 16.39, weight deacy loss is 8.26, training accuracy is 0.218750, time 56.434 samples/sec
+epoch 6, total_step 95460, total loss is 20.81 , inference loss is 12.55, weight deacy loss is 8.25, training accuracy is 0.375000, time 55.782 samples/sec
+epoch 6, total_step 95480, total loss is 17.74 , inference loss is 9.49, weight deacy loss is 8.25, training accuracy is 0.531250, time 56.840 samples/sec
+epoch 6, total_step 95500, total loss is 15.99 , inference loss is 7.74, weight deacy loss is 8.25, training accuracy is 0.500000, time 56.439 samples/sec
+epoch 6, total_step 95520, total loss is 12.05 , inference loss is 3.80, weight deacy loss is 8.25, training accuracy is 0.531250, time 56.631 samples/sec
+epoch 6, total_step 95540, total loss is 19.55 , inference loss is 11.30, weight deacy loss is 8.25, training accuracy is 0.343750, time 57.333 samples/sec
+epoch 6, total_step 95560, total loss is 19.19 , inference loss is 10.94, weight deacy loss is 8.25, training accuracy is 0.250000, time 56.228 samples/sec
+epoch 6, total_step 95580, total loss is 14.13 , inference loss is 5.88, weight deacy loss is 8.25, training accuracy is 0.531250, time 56.798 samples/sec
+epoch 6, total_step 95600, total loss is 18.55 , inference loss is 10.30, weight deacy loss is 8.25, training accuracy is 0.468750, time 56.409 samples/sec
+epoch 6, total_step 95620, total loss is 18.65 , inference loss is 10.41, weight deacy loss is 8.25, training accuracy is 0.343750, time 57.180 samples/sec
+epoch 6, total_step 95640, total loss is 17.33 , inference loss is 9.09, weight deacy loss is 8.25, training accuracy is 0.406250, time 56.673 samples/sec
+epoch 6, total_step 95660, total loss is 16.73 , inference loss is 8.48, weight deacy loss is 8.25, training accuracy is 0.375000, time 57.056 samples/sec
+epoch 6, total_step 95680, total loss is 16.12 , inference loss is 7.87, weight deacy loss is 8.25, training accuracy is 0.343750, time 56.214 samples/sec
+epoch 6, total_step 95700, total loss is 20.51 , inference loss is 12.27, weight deacy loss is 8.25, training accuracy is 0.250000, time 56.302 samples/sec
+epoch 6, total_step 95720, total loss is 13.01 , inference loss is 4.77, weight deacy loss is 8.24, training accuracy is 0.468750, time 56.109 samples/sec
+epoch 6, total_step 95740, total loss is 16.07 , inference loss is 7.83, weight deacy loss is 8.24, training accuracy is 0.437500, time 56.803 samples/sec
+epoch 6, total_step 95760, total loss is 17.42 , inference loss is 9.18, weight deacy loss is 8.24, training accuracy is 0.281250, time 56.599 samples/sec
+epoch 6, total_step 95780, total loss is 17.90 , inference loss is 9.66, weight deacy loss is 8.24, training accuracy is 0.562500, time 56.790 samples/sec
+epoch 6, total_step 95800, total loss is 22.01 , inference loss is 13.77, weight deacy loss is 8.24, training accuracy is 0.312500, time 56.091 samples/sec
+epoch 6, total_step 95820, total loss is 17.39 , inference loss is 9.15, weight deacy loss is 8.24, training accuracy is 0.468750, time 56.945 samples/sec
+epoch 6, total_step 95840, total loss is 18.87 , inference loss is 10.63, weight deacy loss is 8.24, training accuracy is 0.312500, time 56.178 samples/sec
+epoch 6, total_step 95860, total loss is 15.54 , inference loss is 7.30, weight deacy loss is 8.24, training accuracy is 0.375000, time 56.389 samples/sec
+epoch 6, total_step 95880, total loss is 21.90 , inference loss is 13.67, weight deacy loss is 8.24, training accuracy is 0.281250, time 57.349 samples/sec
+epoch 6, total_step 95900, total loss is 16.83 , inference loss is 8.59, weight deacy loss is 8.24, training accuracy is 0.375000, time 56.294 samples/sec
+epoch 6, total_step 95920, total loss is 18.00 , inference loss is 9.76, weight deacy loss is 8.24, training accuracy is 0.312500, time 57.753 samples/sec
+epoch 6, total_step 95940, total loss is 21.06 , inference loss is 12.82, weight deacy loss is 8.24, training accuracy is 0.281250, time 56.220 samples/sec
+epoch 6, total_step 95960, total loss is 20.73 , inference loss is 12.49, weight deacy loss is 8.24, training accuracy is 0.343750, time 57.177 samples/sec
+epoch 6, total_step 95980, total loss is 17.97 , inference loss is 9.73, weight deacy loss is 8.24, training accuracy is 0.312500, time 56.656 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.73988200000004
+best_threshold_index 119 0.987037037037037
+best_threshold_index 119 0.9851851851851852
+best_threshold_index 119 0.9866666666666667
+best_threshold_index 118 0.9872222222222222
+best_threshold_index 119 0.987037037037037
+best_threshold_index 122 0.9857407407407407
+best_threshold_index 118 0.9861111111111112
+best_threshold_index 119 0.9862962962962963
+best_threshold_index 119 0.985
+best_threshold_index 119 0.9855555555555555
+[lfw][96000]XNorm: 21.718951
+[lfw][96000]Accuracy-Flip: 0.98533+-0.00694
+testing verification..
+(14000, 512)
+infer time 84.74993800000013
+best_threshold_index 125 0.9880952380952381
+best_threshold_index 125 0.9866666666666667
+best_threshold_index 125 0.9873015873015873
+best_threshold_index 123 0.9866666666666667
+best_threshold_index 125 0.9876190476190476
+best_threshold_index 125 0.9858730158730159
+best_threshold_index 123 0.9871428571428571
+best_threshold_index 121 0.9861904761904762
+best_threshold_index 121 0.9858730158730159
+best_threshold_index 125 0.9873015873015873
+[cfp_ff][96000]XNorm: 21.433831
+[cfp_ff][96000]Accuracy-Flip: 0.98586+-0.00580
+testing verification..
+(14000, 512)
+infer time 85.25747799999996
+best_threshold_index 141 0.9163492063492064
+best_threshold_index 141 0.9146031746031746
+best_threshold_index 141 0.9149206349206349
+best_threshold_index 140 0.9163492063492064
+best_threshold_index 141 0.9171428571428571
+best_threshold_index 141 0.9128571428571428
+best_threshold_index 141 0.9115873015873016
+best_threshold_index 141 0.913968253968254
+best_threshold_index 141 0.915079365079365
+best_threshold_index 141 0.9142857142857143
+[cfp_fp][96000]XNorm: 21.501058
+[cfp_fp][96000]Accuracy-Flip: 0.91429+-0.01483
+testing verification..
+(12000, 512)
+infer time 73.357388
+best_threshold_index 143 0.8883333333333333
+best_threshold_index 143 0.8885185185185185
+best_threshold_index 142 0.8896296296296297
+best_threshold_index 142 0.8905555555555555
+best_threshold_index 142 0.8937037037037037
+best_threshold_index 142 0.8868518518518519
+best_threshold_index 142 0.887962962962963
+best_threshold_index 142 0.8962962962962963
+best_threshold_index 142 0.8935185185185185
+best_threshold_index 142 0.8935185185185185
+[agedb_30][96000]XNorm: 22.208219
+[agedb_30][96000]Accuracy-Flip: 0.89000+-0.02669
+test accuracy is: 0.9853333333333334
+epoch 6, total_step 96000, total loss is 18.87 , inference loss is 10.64, weight deacy loss is 8.23, training accuracy is 0.375000, time 54.321 samples/sec
+epoch 6, total_step 96020, total loss is 16.81 , inference loss is 8.57, weight deacy loss is 8.23, training accuracy is 0.375000, time 56.127 samples/sec
+epoch 6, total_step 96040, total loss is 18.56 , inference loss is 10.33, weight deacy loss is 8.23, training accuracy is 0.375000, time 56.626 samples/sec
+epoch 6, total_step 96060, total loss is 15.01 , inference loss is 6.78, weight deacy loss is 8.23, training accuracy is 0.406250, time 56.261 samples/sec
+epoch 6, total_step 96080, total loss is 18.67 , inference loss is 10.44, weight deacy loss is 8.23, training accuracy is 0.312500, time 57.495 samples/sec
+epoch 6, total_step 96100, total loss is 18.04 , inference loss is 9.80, weight deacy loss is 8.23, training accuracy is 0.406250, time 57.043 samples/sec
+epoch 6, total_step 96120, total loss is 19.25 , inference loss is 11.02, weight deacy loss is 8.23, training accuracy is 0.343750, time 56.591 samples/sec
+epoch 6, total_step 96140, total loss is 17.60 , inference loss is 9.37, weight deacy loss is 8.23, training accuracy is 0.343750, time 56.430 samples/sec
+epoch 6, total_step 96160, total loss is 18.23 , inference loss is 10.00, weight deacy loss is 8.23, training accuracy is 0.375000, time 56.476 samples/sec
+epoch 6, total_step 96180, total loss is 17.07 , inference loss is 8.84, weight deacy loss is 8.23, training accuracy is 0.343750, time 56.684 samples/sec
+epoch 6, total_step 96200, total loss is 15.36 , inference loss is 7.13, weight deacy loss is 8.23, training accuracy is 0.437500, time 56.451 samples/sec
+epoch 6, total_step 96220, total loss is 18.56 , inference loss is 10.33, weight deacy loss is 8.23, training accuracy is 0.375000, time 56.662 samples/sec
+epoch 6, total_step 96240, total loss is 14.68 , inference loss is 6.45, weight deacy loss is 8.23, training accuracy is 0.437500, time 56.590 samples/sec
+epoch 6, total_step 96260, total loss is 19.73 , inference loss is 11.51, weight deacy loss is 8.23, training accuracy is 0.281250, time 56.666 samples/sec
+epoch 6, total_step 96280, total loss is 18.00 , inference loss is 9.77, weight deacy loss is 8.22, training accuracy is 0.406250, time 55.840 samples/sec
+epoch 6, total_step 96300, total loss is 18.18 , inference loss is 9.96, weight deacy loss is 8.22, training accuracy is 0.437500, time 56.109 samples/sec
+epoch 6, total_step 96320, total loss is 16.23 , inference loss is 8.01, weight deacy loss is 8.22, training accuracy is 0.406250, time 55.296 samples/sec
+epoch 6, total_step 96340, total loss is 19.47 , inference loss is 11.25, weight deacy loss is 8.22, training accuracy is 0.406250, time 57.272 samples/sec
+epoch 6, total_step 96360, total loss is 21.45 , inference loss is 13.23, weight deacy loss is 8.22, training accuracy is 0.250000, time 56.085 samples/sec
+epoch 6, total_step 96380, total loss is 18.61 , inference loss is 10.38, weight deacy loss is 8.22, training accuracy is 0.312500, time 56.154 samples/sec
+epoch 6, total_step 96400, total loss is 13.77 , inference loss is 5.55, weight deacy loss is 8.22, training accuracy is 0.500000, time 56.739 samples/sec
+epoch 6, total_step 96420, total loss is 19.61 , inference loss is 11.39, weight deacy loss is 8.22, training accuracy is 0.250000, time 56.718 samples/sec
+epoch 6, total_step 96440, total loss is 18.92 , inference loss is 10.70, weight deacy loss is 8.22, training accuracy is 0.375000, time 57.432 samples/sec
+epoch 6, total_step 96460, total loss is 17.90 , inference loss is 9.68, weight deacy loss is 8.22, training accuracy is 0.312500, time 56.796 samples/sec
+epoch 6, total_step 96480, total loss is 15.73 , inference loss is 7.51, weight deacy loss is 8.22, training accuracy is 0.437500, time 57.977 samples/sec
+epoch 6, total_step 96500, total loss is 21.08 , inference loss is 12.86, weight deacy loss is 8.22, training accuracy is 0.312500, time 56.929 samples/sec
+epoch 6, total_step 96520, total loss is 15.76 , inference loss is 7.54, weight deacy loss is 8.22, training accuracy is 0.375000, time 55.482 samples/sec
+epoch 6, total_step 96540, total loss is 17.79 , inference loss is 9.58, weight deacy loss is 8.21, training accuracy is 0.343750, time 56.467 samples/sec
+epoch 6, total_step 96560, total loss is 18.26 , inference loss is 10.05, weight deacy loss is 8.21, training accuracy is 0.437500, time 55.969 samples/sec
+epoch 6, total_step 96580, total loss is 17.08 , inference loss is 8.87, weight deacy loss is 8.21, training accuracy is 0.406250, time 56.864 samples/sec
+epoch 6, total_step 96600, total loss is 16.04 , inference loss is 7.83, weight deacy loss is 8.21, training accuracy is 0.375000, time 56.528 samples/sec
+epoch 6, total_step 96620, total loss is 21.24 , inference loss is 13.03, weight deacy loss is 8.21, training accuracy is 0.218750, time 56.153 samples/sec
+epoch 6, total_step 96640, total loss is 18.61 , inference loss is 10.40, weight deacy loss is 8.21, training accuracy is 0.375000, time 56.424 samples/sec
+epoch 6, total_step 96660, total loss is 21.09 , inference loss is 12.88, weight deacy loss is 8.21, training accuracy is 0.406250, time 56.911 samples/sec
+epoch 6, total_step 96680, total loss is 20.27 , inference loss is 12.06, weight deacy loss is 8.21, training accuracy is 0.250000, time 56.549 samples/sec
+epoch 6, total_step 96700, total loss is 16.23 , inference loss is 8.02, weight deacy loss is 8.21, training accuracy is 0.531250, time 56.468 samples/sec
+epoch 6, total_step 96720, total loss is 14.24 , inference loss is 6.03, weight deacy loss is 8.21, training accuracy is 0.375000, time 57.026 samples/sec
+epoch 6, total_step 96740, total loss is 20.04 , inference loss is 11.83, weight deacy loss is 8.21, training accuracy is 0.406250, time 56.076 samples/sec
+epoch 6, total_step 96760, total loss is 18.78 , inference loss is 10.57, weight deacy loss is 8.21, training accuracy is 0.312500, time 56.012 samples/sec
+epoch 6, total_step 96780, total loss is 22.31 , inference loss is 14.10, weight deacy loss is 8.21, training accuracy is 0.343750, time 56.265 samples/sec
+epoch 6, total_step 96800, total loss is 15.28 , inference loss is 7.08, weight deacy loss is 8.21, training accuracy is 0.343750, time 56.860 samples/sec
+epoch 6, total_step 96820, total loss is 20.44 , inference loss is 12.23, weight deacy loss is 8.20, training accuracy is 0.312500, time 56.121 samples/sec
+epoch 6, total_step 96840, total loss is 16.96 , inference loss is 8.76, weight deacy loss is 8.20, training accuracy is 0.343750, time 56.818 samples/sec
+epoch 6, total_step 96860, total loss is 14.34 , inference loss is 6.14, weight deacy loss is 8.20, training accuracy is 0.500000, time 56.815 samples/sec
+epoch 6, total_step 96880, total loss is 19.85 , inference loss is 11.64, weight deacy loss is 8.20, training accuracy is 0.312500, time 56.759 samples/sec
+epoch 6, total_step 96900, total loss is 19.03 , inference loss is 10.83, weight deacy loss is 8.20, training accuracy is 0.343750, time 56.476 samples/sec
+epoch 6, total_step 96920, total loss is 18.77 , inference loss is 10.57, weight deacy loss is 8.20, training accuracy is 0.250000, time 56.554 samples/sec
+epoch 6, total_step 96940, total loss is 17.23 , inference loss is 9.03, weight deacy loss is 8.20, training accuracy is 0.218750, time 57.014 samples/sec
+epoch 6, total_step 96960, total loss is 15.30 , inference loss is 7.10, weight deacy loss is 8.20, training accuracy is 0.406250, time 56.563 samples/sec
+epoch 6, total_step 96980, total loss is 19.53 , inference loss is 11.33, weight deacy loss is 8.20, training accuracy is 0.406250, time 56.898 samples/sec
+epoch 6, total_step 97000, total loss is 21.20 , inference loss is 13.00, weight deacy loss is 8.20, training accuracy is 0.218750, time 56.550 samples/sec
+epoch 6, total_step 97020, total loss is 17.54 , inference loss is 9.35, weight deacy loss is 8.20, training accuracy is 0.375000, time 56.438 samples/sec
+epoch 6, total_step 97040, total loss is 15.81 , inference loss is 7.61, weight deacy loss is 8.20, training accuracy is 0.406250, time 57.168 samples/sec
+epoch 6, total_step 97060, total loss is 17.19 , inference loss is 9.00, weight deacy loss is 8.20, training accuracy is 0.406250, time 56.665 samples/sec
+epoch 6, total_step 97080, total loss is 17.94 , inference loss is 9.74, weight deacy loss is 8.19, training accuracy is 0.468750, time 55.788 samples/sec
+epoch 6, total_step 97100, total loss is 20.01 , inference loss is 11.81, weight deacy loss is 8.19, training accuracy is 0.343750, time 55.911 samples/sec
+epoch 6, total_step 97120, total loss is 16.78 , inference loss is 8.58, weight deacy loss is 8.19, training accuracy is 0.406250, time 57.092 samples/sec
+epoch 6, total_step 97140, total loss is 18.01 , inference loss is 9.82, weight deacy loss is 8.19, training accuracy is 0.312500, time 55.959 samples/sec
+epoch 6, total_step 97160, total loss is 16.35 , inference loss is 8.16, weight deacy loss is 8.19, training accuracy is 0.375000, time 56.514 samples/sec
+epoch 6, total_step 97180, total loss is 13.59 , inference loss is 5.40, weight deacy loss is 8.19, training accuracy is 0.468750, time 56.739 samples/sec
+epoch 6, total_step 97200, total loss is 15.00 , inference loss is 6.81, weight deacy loss is 8.19, training accuracy is 0.437500, time 56.521 samples/sec
+epoch 6, total_step 97220, total loss is 20.80 , inference loss is 12.61, weight deacy loss is 8.19, training accuracy is 0.375000, time 56.123 samples/sec
+epoch 6, total_step 97240, total loss is 18.50 , inference loss is 10.31, weight deacy loss is 8.19, training accuracy is 0.406250, time 55.696 samples/sec
+epoch 6, total_step 97260, total loss is 14.34 , inference loss is 6.15, weight deacy loss is 8.19, training accuracy is 0.468750, time 56.515 samples/sec
+epoch 6, total_step 97280, total loss is 17.93 , inference loss is 9.74, weight deacy loss is 8.19, training accuracy is 0.375000, time 56.280 samples/sec
+epoch 6, total_step 97300, total loss is 21.36 , inference loss is 13.18, weight deacy loss is 8.19, training accuracy is 0.312500, time 56.410 samples/sec
+epoch 6, total_step 97320, total loss is 20.15 , inference loss is 11.96, weight deacy loss is 8.19, training accuracy is 0.312500, time 56.666 samples/sec
+epoch 6, total_step 97340, total loss is 17.55 , inference loss is 9.37, weight deacy loss is 8.19, training accuracy is 0.343750, time 57.002 samples/sec
+epoch 6, total_step 97360, total loss is 16.82 , inference loss is 8.64, weight deacy loss is 8.18, training accuracy is 0.343750, time 56.030 samples/sec
+epoch 6, total_step 97380, total loss is 13.38 , inference loss is 5.19, weight deacy loss is 8.18, training accuracy is 0.468750, time 56.822 samples/sec
+epoch 6, total_step 97400, total loss is 18.18 , inference loss is 10.00, weight deacy loss is 8.18, training accuracy is 0.343750, time 55.326 samples/sec
+epoch 6, total_step 97420, total loss is 14.97 , inference loss is 6.78, weight deacy loss is 8.18, training accuracy is 0.375000, time 56.212 samples/sec
+epoch 6, total_step 97440, total loss is 16.46 , inference loss is 8.28, weight deacy loss is 8.18, training accuracy is 0.218750, time 56.272 samples/sec
+epoch 6, total_step 97460, total loss is 19.28 , inference loss is 11.10, weight deacy loss is 8.18, training accuracy is 0.375000, time 56.604 samples/sec
+epoch 6, total_step 97480, total loss is 17.88 , inference loss is 9.70, weight deacy loss is 8.18, training accuracy is 0.343750, time 56.551 samples/sec
+epoch 6, total_step 97500, total loss is 17.66 , inference loss is 9.48, weight deacy loss is 8.18, training accuracy is 0.375000, time 56.082 samples/sec
+epoch 6, total_step 97520, total loss is 15.38 , inference loss is 7.20, weight deacy loss is 8.18, training accuracy is 0.468750, time 56.962 samples/sec
+epoch 6, total_step 97540, total loss is 18.39 , inference loss is 10.21, weight deacy loss is 8.18, training accuracy is 0.250000, time 56.748 samples/sec
+epoch 6, total_step 97560, total loss is 15.96 , inference loss is 7.78, weight deacy loss is 8.18, training accuracy is 0.343750, time 56.667 samples/sec
+epoch 6, total_step 97580, total loss is 18.88 , inference loss is 10.70, weight deacy loss is 8.18, training accuracy is 0.343750, time 57.052 samples/sec
+epoch 6, total_step 97600, total loss is 17.02 , inference loss is 8.84, weight deacy loss is 8.18, training accuracy is 0.312500, time 56.965 samples/sec
+epoch 6, total_step 97620, total loss is 18.20 , inference loss is 10.02, weight deacy loss is 8.17, training accuracy is 0.406250, time 56.640 samples/sec
+epoch 6, total_step 97640, total loss is 16.56 , inference loss is 8.39, weight deacy loss is 8.17, training accuracy is 0.250000, time 57.082 samples/sec
+epoch 6, total_step 97660, total loss is 20.49 , inference loss is 12.32, weight deacy loss is 8.17, training accuracy is 0.281250, time 55.330 samples/sec
+epoch 6, total_step 97680, total loss is 20.74 , inference loss is 12.57, weight deacy loss is 8.17, training accuracy is 0.281250, time 55.909 samples/sec
+epoch 6, total_step 97700, total loss is 16.48 , inference loss is 8.31, weight deacy loss is 8.17, training accuracy is 0.312500, time 56.227 samples/sec
+epoch 6, total_step 97720, total loss is 18.96 , inference loss is 10.79, weight deacy loss is 8.17, training accuracy is 0.375000, time 56.303 samples/sec
+epoch 6, total_step 97740, total loss is 23.17 , inference loss is 15.00, weight deacy loss is 8.17, training accuracy is 0.218750, time 56.748 samples/sec
+epoch 6, total_step 97760, total loss is 19.33 , inference loss is 11.16, weight deacy loss is 8.17, training accuracy is 0.218750, time 57.066 samples/sec
+epoch 6, total_step 97780, total loss is 18.24 , inference loss is 10.07, weight deacy loss is 8.17, training accuracy is 0.250000, time 56.480 samples/sec
+epoch 6, total_step 97800, total loss is 18.21 , inference loss is 10.04, weight deacy loss is 8.17, training accuracy is 0.375000, time 56.034 samples/sec
+epoch 6, total_step 97820, total loss is 17.99 , inference loss is 9.82, weight deacy loss is 8.17, training accuracy is 0.437500, time 55.962 samples/sec
+epoch 6, total_step 97840, total loss is 22.54 , inference loss is 14.38, weight deacy loss is 8.17, training accuracy is 0.281250, time 56.593 samples/sec
+epoch 6, total_step 97860, total loss is 20.42 , inference loss is 12.25, weight deacy loss is 8.17, training accuracy is 0.437500, time 55.951 samples/sec
+epoch 6, total_step 97880, total loss is 15.31 , inference loss is 7.15, weight deacy loss is 8.17, training accuracy is 0.406250, time 56.484 samples/sec
+epoch 6, total_step 97900, total loss is 19.09 , inference loss is 10.93, weight deacy loss is 8.16, training accuracy is 0.218750, time 56.675 samples/sec
+epoch 6, total_step 97920, total loss is 19.65 , inference loss is 11.49, weight deacy loss is 8.16, training accuracy is 0.250000, time 56.933 samples/sec
+epoch 6, total_step 97940, total loss is 17.16 , inference loss is 8.99, weight deacy loss is 8.16, training accuracy is 0.468750, time 57.034 samples/sec
+epoch 6, total_step 97960, total loss is 17.44 , inference loss is 9.28, weight deacy loss is 8.16, training accuracy is 0.343750, time 56.870 samples/sec
+epoch 6, total_step 97980, total loss is 18.13 , inference loss is 9.97, weight deacy loss is 8.16, training accuracy is 0.375000, time 56.585 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.44845899999996
+best_threshold_index 121 0.9885185185185185
+best_threshold_index 121 0.9866666666666667
+best_threshold_index 121 0.9879629629629629
+best_threshold_index 121 0.9881481481481481
+best_threshold_index 121 0.9881481481481481
+best_threshold_index 122 0.9874074074074074
+best_threshold_index 121 0.9879629629629629
+best_threshold_index 121 0.9868518518518519
+best_threshold_index 121 0.9866666666666667
+best_threshold_index 121 0.9868518518518519
+[lfw][98000]XNorm: 20.820239
+[lfw][98000]Accuracy-Flip: 0.98700+-0.00609
+testing verification..
+(14000, 512)
+infer time 85.00140600000009
+best_threshold_index 121 0.9873015873015873
+best_threshold_index 121 0.9858730158730159
+best_threshold_index 121 0.9853968253968254
+best_threshold_index 121 0.9847619047619047
+best_threshold_index 121 0.986031746031746
+best_threshold_index 121 0.984920634920635
+best_threshold_index 121 0.9865079365079366
+best_threshold_index 121 0.9853968253968254
+best_threshold_index 121 0.9855555555555555
+best_threshold_index 121 0.9868253968253968
+[cfp_ff][98000]XNorm: 20.514410
+[cfp_ff][98000]Accuracy-Flip: 0.98586+-0.00701
+testing verification..
+(14000, 512)
+infer time 84.97502100000005
+best_threshold_index 139 0.9177777777777778
+best_threshold_index 139 0.916984126984127
+best_threshold_index 139 0.9144444444444444
+best_threshold_index 135 0.9171428571428571
+best_threshold_index 139 0.9198412698412698
+best_threshold_index 139 0.9138095238095238
+best_threshold_index 135 0.913968253968254
+best_threshold_index 139 0.9166666666666666
+best_threshold_index 139 0.9171428571428571
+best_threshold_index 138 0.9168253968253969
+[cfp_fp][98000]XNorm: 20.575833
+[cfp_fp][98000]Accuracy-Flip: 0.91486+-0.01583
+testing verification..
+(12000, 512)
+infer time 73.04720600000002
+best_threshold_index 136 0.8961111111111111
+best_threshold_index 136 0.8942592592592593
+best_threshold_index 136 0.8994444444444445
+best_threshold_index 136 0.8981481481481481
+best_threshold_index 136 0.9024074074074074
+best_threshold_index 136 0.8953703703703704
+best_threshold_index 133 0.8975925925925926
+best_threshold_index 136 0.9024074074074074
+best_threshold_index 133 0.8994444444444445
+best_threshold_index 136 0.8968518518518519
+[agedb_30][98000]XNorm: 21.473991
+[agedb_30][98000]Accuracy-Flip: 0.89617+-0.02400
+test accuracy is: 0.9870000000000001
+epoch 6, total_step 98000, total loss is 21.41 , inference loss is 13.25, weight deacy loss is 8.16, training accuracy is 0.250000, time 53.549 samples/sec
+epoch 6, total_step 98020, total loss is 18.99 , inference loss is 10.83, weight deacy loss is 8.16, training accuracy is 0.343750, time 56.844 samples/sec
+epoch 6, total_step 98040, total loss is 17.76 , inference loss is 9.60, weight deacy loss is 8.16, training accuracy is 0.343750, time 56.886 samples/sec
+epoch 6, total_step 98060, total loss is 17.69 , inference loss is 9.53, weight deacy loss is 8.16, training accuracy is 0.343750, time 57.045 samples/sec
+epoch 6, total_step 98080, total loss is 21.92 , inference loss is 13.76, weight deacy loss is 8.16, training accuracy is 0.187500, time 56.517 samples/sec
+epoch 6, total_step 98100, total loss is 21.37 , inference loss is 13.22, weight deacy loss is 8.16, training accuracy is 0.281250, time 56.201 samples/sec
+epoch 6, total_step 98120, total loss is 18.23 , inference loss is 10.07, weight deacy loss is 8.16, training accuracy is 0.250000, time 56.689 samples/sec
+epoch 6, total_step 98140, total loss is 17.40 , inference loss is 9.24, weight deacy loss is 8.16, training accuracy is 0.406250, time 57.424 samples/sec
+epoch 6, total_step 98160, total loss is 21.15 , inference loss is 12.99, weight deacy loss is 8.16, training accuracy is 0.250000, time 56.914 samples/sec
+epoch 6, total_step 98180, total loss is 16.80 , inference loss is 8.64, weight deacy loss is 8.15, training accuracy is 0.375000, time 57.348 samples/sec
+epoch 6, total_step 98200, total loss is 20.08 , inference loss is 11.93, weight deacy loss is 8.15, training accuracy is 0.343750, time 56.507 samples/sec
+epoch 6, total_step 98220, total loss is 20.96 , inference loss is 12.81, weight deacy loss is 8.15, training accuracy is 0.187500, time 56.540 samples/sec
+epoch 6, total_step 98240, total loss is 18.11 , inference loss is 9.96, weight deacy loss is 8.15, training accuracy is 0.312500, time 56.330 samples/sec
+epoch 6, total_step 98260, total loss is 16.08 , inference loss is 7.93, weight deacy loss is 8.15, training accuracy is 0.593750, time 56.361 samples/sec
+epoch 6, total_step 98280, total loss is 15.58 , inference loss is 7.43, weight deacy loss is 8.15, training accuracy is 0.375000, time 57.072 samples/sec
+epoch 6, total_step 98300, total loss is 16.13 , inference loss is 7.98, weight deacy loss is 8.15, training accuracy is 0.312500, time 55.718 samples/sec
+epoch 6, total_step 98320, total loss is 16.85 , inference loss is 8.70, weight deacy loss is 8.15, training accuracy is 0.375000, time 56.408 samples/sec
+epoch 6, total_step 98340, total loss is 16.93 , inference loss is 8.79, weight deacy loss is 8.15, training accuracy is 0.375000, time 56.937 samples/sec
+epoch 6, total_step 98360, total loss is 14.08 , inference loss is 5.93, weight deacy loss is 8.15, training accuracy is 0.500000, time 55.811 samples/sec
+epoch 6, total_step 98380, total loss is 18.95 , inference loss is 10.80, weight deacy loss is 8.15, training accuracy is 0.375000, time 54.992 samples/sec
+epoch 6, total_step 98400, total loss is 21.24 , inference loss is 13.09, weight deacy loss is 8.15, training accuracy is 0.187500, time 56.630 samples/sec
+epoch 6, total_step 98420, total loss is 15.55 , inference loss is 7.40, weight deacy loss is 8.15, training accuracy is 0.406250, time 56.800 samples/sec
+epoch 6, total_step 98440, total loss is 20.05 , inference loss is 11.90, weight deacy loss is 8.14, training accuracy is 0.218750, time 56.339 samples/sec
+epoch 6, total_step 98460, total loss is 13.63 , inference loss is 5.49, weight deacy loss is 8.14, training accuracy is 0.312500, time 56.639 samples/sec
+epoch 6, total_step 98480, total loss is 19.81 , inference loss is 11.67, weight deacy loss is 8.14, training accuracy is 0.218750, time 56.368 samples/sec
+epoch 6, total_step 98500, total loss is 20.05 , inference loss is 11.91, weight deacy loss is 8.14, training accuracy is 0.312500, time 56.262 samples/sec
+epoch 6, total_step 98520, total loss is 14.37 , inference loss is 6.23, weight deacy loss is 8.14, training accuracy is 0.468750, time 56.518 samples/sec
+epoch 6, total_step 98540, total loss is 14.90 , inference loss is 6.76, weight deacy loss is 8.14, training accuracy is 0.531250, time 57.101 samples/sec
+epoch 6, total_step 98560, total loss is 21.21 , inference loss is 13.07, weight deacy loss is 8.14, training accuracy is 0.312500, time 55.911 samples/sec
+epoch 6, total_step 98580, total loss is 17.80 , inference loss is 9.66, weight deacy loss is 8.14, training accuracy is 0.250000, time 56.568 samples/sec
+epoch 6, total_step 98600, total loss is 15.54 , inference loss is 7.40, weight deacy loss is 8.14, training accuracy is 0.312500, time 56.547 samples/sec
+epoch 6, total_step 98620, total loss is 16.52 , inference loss is 8.38, weight deacy loss is 8.14, training accuracy is 0.250000, time 55.075 samples/sec
+epoch 6, total_step 98640, total loss is 20.18 , inference loss is 12.04, weight deacy loss is 8.14, training accuracy is 0.218750, time 56.961 samples/sec
+epoch 6, total_step 98660, total loss is 15.34 , inference loss is 7.20, weight deacy loss is 8.14, training accuracy is 0.437500, time 56.030 samples/sec
+epoch 6, total_step 98680, total loss is 20.72 , inference loss is 12.59, weight deacy loss is 8.14, training accuracy is 0.218750, time 56.467 samples/sec
+epoch 6, total_step 98700, total loss is 21.04 , inference loss is 12.91, weight deacy loss is 8.14, training accuracy is 0.437500, time 55.604 samples/sec
+epoch 6, total_step 98720, total loss is 14.72 , inference loss is 6.59, weight deacy loss is 8.13, training accuracy is 0.312500, time 56.061 samples/sec
+epoch 6, total_step 98740, total loss is 15.66 , inference loss is 7.52, weight deacy loss is 8.13, training accuracy is 0.406250, time 55.606 samples/sec
+epoch 6, total_step 98760, total loss is 18.64 , inference loss is 10.51, weight deacy loss is 8.13, training accuracy is 0.250000, time 56.541 samples/sec
+epoch 6, total_step 98780, total loss is 15.96 , inference loss is 7.83, weight deacy loss is 8.13, training accuracy is 0.250000, time 56.769 samples/sec
+epoch 6, total_step 98800, total loss is 16.37 , inference loss is 8.24, weight deacy loss is 8.13, training accuracy is 0.343750, time 56.717 samples/sec
+epoch 6, total_step 98820, total loss is 15.45 , inference loss is 7.32, weight deacy loss is 8.13, training accuracy is 0.468750, time 56.785 samples/sec
+epoch 6, total_step 98840, total loss is 17.39 , inference loss is 9.26, weight deacy loss is 8.13, training accuracy is 0.343750, time 56.725 samples/sec
+epoch 6, total_step 98860, total loss is 15.66 , inference loss is 7.53, weight deacy loss is 8.13, training accuracy is 0.406250, time 56.304 samples/sec
+epoch 6, total_step 98880, total loss is 23.08 , inference loss is 14.95, weight deacy loss is 8.13, training accuracy is 0.218750, time 56.181 samples/sec
+epoch 6, total_step 98900, total loss is 19.39 , inference loss is 11.26, weight deacy loss is 8.13, training accuracy is 0.218750, time 56.231 samples/sec
+epoch 6, total_step 98920, total loss is 19.75 , inference loss is 11.63, weight deacy loss is 8.13, training accuracy is 0.218750, time 57.157 samples/sec
+epoch 6, total_step 98940, total loss is 19.57 , inference loss is 11.45, weight deacy loss is 8.13, training accuracy is 0.375000, time 56.611 samples/sec
+epoch 6, total_step 98960, total loss is 20.52 , inference loss is 12.40, weight deacy loss is 8.13, training accuracy is 0.281250, time 56.850 samples/sec
+epoch 6, total_step 98980, total loss is 14.08 , inference loss is 5.95, weight deacy loss is 8.13, training accuracy is 0.468750, time 56.053 samples/sec
+epoch 6, total_step 99000, total loss is 13.87 , inference loss is 5.75, weight deacy loss is 8.12, training accuracy is 0.500000, time 56.477 samples/sec
+epoch 6, total_step 99020, total loss is 19.24 , inference loss is 11.12, weight deacy loss is 8.12, training accuracy is 0.250000, time 57.323 samples/sec
+epoch 6, total_step 99040, total loss is 19.38 , inference loss is 11.25, weight deacy loss is 8.12, training accuracy is 0.250000, time 56.045 samples/sec
+epoch 6, total_step 99060, total loss is 19.20 , inference loss is 11.08, weight deacy loss is 8.12, training accuracy is 0.281250, time 54.917 samples/sec
+epoch 6, total_step 99080, total loss is 16.67 , inference loss is 8.55, weight deacy loss is 8.12, training accuracy is 0.406250, time 56.924 samples/sec
+epoch 6, total_step 99100, total loss is 17.43 , inference loss is 9.31, weight deacy loss is 8.12, training accuracy is 0.375000, time 56.807 samples/sec
+epoch 6, total_step 99120, total loss is 21.75 , inference loss is 13.63, weight deacy loss is 8.12, training accuracy is 0.343750, time 56.614 samples/sec
+epoch 6, total_step 99140, total loss is 14.58 , inference loss is 6.46, weight deacy loss is 8.12, training accuracy is 0.375000, time 56.917 samples/sec
+epoch 6, total_step 99160, total loss is 17.99 , inference loss is 9.87, weight deacy loss is 8.12, training accuracy is 0.312500, time 55.899 samples/sec
+epoch 6, total_step 99180, total loss is 17.63 , inference loss is 9.51, weight deacy loss is 8.12, training accuracy is 0.250000, time 56.558 samples/sec
+epoch 6, total_step 99200, total loss is 15.04 , inference loss is 6.93, weight deacy loss is 8.12, training accuracy is 0.343750, time 56.795 samples/sec
+epoch 6, total_step 99220, total loss is 19.09 , inference loss is 10.98, weight deacy loss is 8.12, training accuracy is 0.281250, time 56.731 samples/sec
+epoch 6, total_step 99240, total loss is 18.93 , inference loss is 10.81, weight deacy loss is 8.12, training accuracy is 0.218750, time 56.449 samples/sec
+epoch 6, total_step 99260, total loss is 14.15 , inference loss is 6.03, weight deacy loss is 8.12, training accuracy is 0.375000, time 56.675 samples/sec
+epoch 6, total_step 99280, total loss is 19.33 , inference loss is 11.22, weight deacy loss is 8.11, training accuracy is 0.250000, time 55.838 samples/sec
+epoch 6, total_step 99300, total loss is 16.91 , inference loss is 8.79, weight deacy loss is 8.11, training accuracy is 0.343750, time 56.616 samples/sec
+epoch 6, total_step 99320, total loss is 15.89 , inference loss is 7.78, weight deacy loss is 8.11, training accuracy is 0.500000, time 57.421 samples/sec
+epoch 6, total_step 99340, total loss is 17.61 , inference loss is 9.50, weight deacy loss is 8.11, training accuracy is 0.375000, time 56.609 samples/sec
+epoch 6, total_step 99360, total loss is 17.93 , inference loss is 9.82, weight deacy loss is 8.11, training accuracy is 0.187500, time 56.881 samples/sec
+epoch 6, total_step 99380, total loss is 13.70 , inference loss is 5.59, weight deacy loss is 8.11, training accuracy is 0.562500, time 57.137 samples/sec
+epoch 6, total_step 99400, total loss is 17.60 , inference loss is 9.49, weight deacy loss is 8.11, training accuracy is 0.406250, time 55.479 samples/sec
+epoch 6, total_step 99420, total loss is 16.86 , inference loss is 8.75, weight deacy loss is 8.11, training accuracy is 0.437500, time 56.361 samples/sec
+epoch 6, total_step 99440, total loss is 16.57 , inference loss is 8.46, weight deacy loss is 8.11, training accuracy is 0.468750, time 56.962 samples/sec
+epoch 6, total_step 99460, total loss is 15.49 , inference loss is 7.38, weight deacy loss is 8.11, training accuracy is 0.500000, time 56.444 samples/sec
+epoch 6, total_step 99480, total loss is 17.27 , inference loss is 9.17, weight deacy loss is 8.11, training accuracy is 0.375000, time 57.068 samples/sec
+epoch 6, total_step 99500, total loss is 16.43 , inference loss is 8.33, weight deacy loss is 8.11, training accuracy is 0.468750, time 55.904 samples/sec
+epoch 6, total_step 99520, total loss is 17.57 , inference loss is 9.47, weight deacy loss is 8.11, training accuracy is 0.406250, time 56.425 samples/sec
+epoch 6, total_step 99540, total loss is 21.86 , inference loss is 13.75, weight deacy loss is 8.10, training accuracy is 0.187500, time 56.744 samples/sec
+epoch 6, total_step 99560, total loss is 17.97 , inference loss is 9.87, weight deacy loss is 8.10, training accuracy is 0.375000, time 56.467 samples/sec
+epoch 6, total_step 99580, total loss is 16.91 , inference loss is 8.81, weight deacy loss is 8.10, training accuracy is 0.406250, time 56.379 samples/sec
+epoch 6, total_step 99600, total loss is 16.25 , inference loss is 8.15, weight deacy loss is 8.10, training accuracy is 0.437500, time 56.368 samples/sec
+epoch 6, total_step 99620, total loss is 14.20 , inference loss is 6.10, weight deacy loss is 8.10, training accuracy is 0.343750, time 56.081 samples/sec
+epoch 6, total_step 99640, total loss is 14.77 , inference loss is 6.67, weight deacy loss is 8.10, training accuracy is 0.468750, time 56.946 samples/sec
+epoch 6, total_step 99660, total loss is 16.46 , inference loss is 8.36, weight deacy loss is 8.10, training accuracy is 0.406250, time 56.894 samples/sec
+epoch 6, total_step 99680, total loss is 19.09 , inference loss is 10.99, weight deacy loss is 8.10, training accuracy is 0.375000, time 55.897 samples/sec
+epoch 6, total_step 99700, total loss is 20.61 , inference loss is 12.51, weight deacy loss is 8.10, training accuracy is 0.187500, time 56.067 samples/sec
+epoch 6, total_step 99720, total loss is 13.91 , inference loss is 5.81, weight deacy loss is 8.10, training accuracy is 0.406250, time 56.355 samples/sec
+epoch 6, total_step 99740, total loss is 18.99 , inference loss is 10.89, weight deacy loss is 8.10, training accuracy is 0.281250, time 56.631 samples/sec
+epoch 6, total_step 99760, total loss is 18.01 , inference loss is 9.91, weight deacy loss is 8.10, training accuracy is 0.343750, time 55.949 samples/sec
+epoch 6, total_step 99780, total loss is 21.00 , inference loss is 12.90, weight deacy loss is 8.10, training accuracy is 0.281250, time 55.961 samples/sec
+epoch 6, total_step 99800, total loss is 22.64 , inference loss is 14.55, weight deacy loss is 8.09, training accuracy is 0.250000, time 55.305 samples/sec
+epoch 6, total_step 99820, total loss is 19.49 , inference loss is 11.39, weight deacy loss is 8.09, training accuracy is 0.281250, time 56.893 samples/sec
+epoch 6, total_step 99840, total loss is 18.90 , inference loss is 10.81, weight deacy loss is 8.09, training accuracy is 0.218750, time 57.115 samples/sec
+epoch 6, total_step 99860, total loss is 16.48 , inference loss is 8.39, weight deacy loss is 8.09, training accuracy is 0.437500, time 57.272 samples/sec
+epoch 6, total_step 99880, total loss is 16.11 , inference loss is 8.02, weight deacy loss is 8.09, training accuracy is 0.375000, time 57.404 samples/sec
+epoch 6, total_step 99900, total loss is 17.18 , inference loss is 9.09, weight deacy loss is 8.09, training accuracy is 0.500000, time 57.150 samples/sec
+epoch 6, total_step 99920, total loss is 18.31 , inference loss is 10.22, weight deacy loss is 8.09, training accuracy is 0.375000, time 57.035 samples/sec
+epoch 6, total_step 99940, total loss is 17.89 , inference loss is 9.80, weight deacy loss is 8.09, training accuracy is 0.250000, time 56.284 samples/sec
+epoch 6, total_step 99960, total loss is 20.32 , inference loss is 12.23, weight deacy loss is 8.09, training accuracy is 0.250000, time 56.800 samples/sec
+epoch 6, total_step 99980, total loss is 19.19 , inference loss is 11.10, weight deacy loss is 8.09, training accuracy is 0.218750, time 55.496 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.8506879999999
+best_threshold_index 112 0.9864814814814815
+best_threshold_index 121 0.985
+best_threshold_index 116 0.9857407407407407
+best_threshold_index 116 0.9862962962962963
+best_threshold_index 116 0.9855555555555555
+best_threshold_index 112 0.985
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 116 0.9853703703703703
+best_threshold_index 112 0.9842592592592593
+best_threshold_index 112 0.9840740740740741
+[lfw][100000]XNorm: 21.841007
+[lfw][100000]Accuracy-Flip: 0.98400+-0.00638
+testing verification..
+(14000, 512)
+infer time 86.11353199999994
+best_threshold_index 118 0.9853968253968254
+best_threshold_index 118 0.9842857142857143
+best_threshold_index 118 0.9838095238095238
+best_threshold_index 118 0.9841269841269841
+best_threshold_index 118 0.9842857142857143
+best_threshold_index 118 0.9839682539682539
+best_threshold_index 121 0.984920634920635
+best_threshold_index 118 0.9844444444444445
+best_threshold_index 118 0.9833333333333333
+best_threshold_index 118 0.9844444444444445
+[cfp_ff][100000]XNorm: 21.458234
+[cfp_ff][100000]Accuracy-Flip: 0.98371+-0.00550
+testing verification..
+(14000, 512)
+infer time 85.73046499999997
+best_threshold_index 142 0.9193650793650794
+best_threshold_index 142 0.9176190476190477
+best_threshold_index 142 0.9166666666666666
+best_threshold_index 142 0.9204761904761904
+best_threshold_index 142 0.9188888888888889
+best_threshold_index 142 0.9166666666666666
+best_threshold_index 142 0.9153968253968254
+best_threshold_index 142 0.9168253968253969
+best_threshold_index 142 0.9171428571428571
+best_threshold_index 142 0.9180952380952381
+[cfp_fp][100000]XNorm: 21.499028
+[cfp_fp][100000]Accuracy-Flip: 0.91771+-0.01291
+testing verification..
+(12000, 512)
+infer time 73.98843399999996
+best_threshold_index 141 0.9094444444444445
+best_threshold_index 141 0.9096296296296297
+best_threshold_index 141 0.9120370370370371
+best_threshold_index 141 0.9111111111111111
+best_threshold_index 141 0.914074074074074
+best_threshold_index 141 0.909074074074074
+best_threshold_index 141 0.9112962962962963
+best_threshold_index 141 0.915
+best_threshold_index 141 0.9125925925925926
+best_threshold_index 141 0.914074074074074
+[agedb_30][100000]XNorm: 22.696280
+[agedb_30][100000]Accuracy-Flip: 0.91183+-0.01793
+test accuracy is: 0.984
+epoch 6, total_step 100000, total loss is 18.55 , inference loss is 10.46, weight deacy loss is 8.09, training accuracy is 0.406250, time 55.076 samples/sec
+epoch 6, total_step 100020, total loss is 20.82 , inference loss is 12.73, weight deacy loss is 8.09, training accuracy is 0.281250, time 56.643 samples/sec
+epoch 6, total_step 100040, total loss is 17.87 , inference loss is 9.78, weight deacy loss is 8.09, training accuracy is 0.406250, time 57.775 samples/sec
+epoch 6, total_step 100060, total loss is 18.92 , inference loss is 10.84, weight deacy loss is 8.09, training accuracy is 0.250000, time 55.239 samples/sec
+epoch 6, total_step 100080, total loss is 16.17 , inference loss is 8.08, weight deacy loss is 8.08, training accuracy is 0.468750, time 56.858 samples/sec
+epoch 6, total_step 100100, total loss is 16.25 , inference loss is 8.16, weight deacy loss is 8.08, training accuracy is 0.250000, time 56.328 samples/sec
+epoch 6, total_step 100120, total loss is 15.26 , inference loss is 7.18, weight deacy loss is 8.08, training accuracy is 0.375000, time 56.346 samples/sec
+epoch 6, total_step 100140, total loss is 19.18 , inference loss is 11.09, weight deacy loss is 8.08, training accuracy is 0.187500, time 56.261 samples/sec
+epoch 6, total_step 100160, total loss is 20.36 , inference loss is 12.28, weight deacy loss is 8.08, training accuracy is 0.312500, time 56.848 samples/sec
+epoch 6, total_step 100180, total loss is 19.66 , inference loss is 11.58, weight deacy loss is 8.08, training accuracy is 0.375000, time 56.935 samples/sec
+epoch 6, total_step 100200, total loss is 17.13 , inference loss is 9.05, weight deacy loss is 8.08, training accuracy is 0.343750, time 55.687 samples/sec
+epoch 6, total_step 100220, total loss is 21.08 , inference loss is 13.01, weight deacy loss is 8.08, training accuracy is 0.281250, time 56.618 samples/sec
+epoch 6, total_step 100240, total loss is 17.15 , inference loss is 9.07, weight deacy loss is 8.08, training accuracy is 0.343750, time 56.492 samples/sec
+epoch 6, total_step 100260, total loss is 19.26 , inference loss is 11.19, weight deacy loss is 8.08, training accuracy is 0.343750, time 56.959 samples/sec
+epoch 6, total_step 100280, total loss is 18.42 , inference loss is 10.34, weight deacy loss is 8.08, training accuracy is 0.281250, time 56.707 samples/sec
+epoch 6, total_step 100300, total loss is 15.31 , inference loss is 7.23, weight deacy loss is 8.08, training accuracy is 0.500000, time 55.589 samples/sec
+epoch 6, total_step 100320, total loss is 16.91 , inference loss is 8.83, weight deacy loss is 8.08, training accuracy is 0.437500, time 56.689 samples/sec
+epoch 6, total_step 100340, total loss is 21.55 , inference loss is 13.47, weight deacy loss is 8.07, training accuracy is 0.312500, time 56.916 samples/sec
+epoch 6, total_step 100360, total loss is 18.90 , inference loss is 10.83, weight deacy loss is 8.07, training accuracy is 0.312500, time 56.253 samples/sec
+epoch 6, total_step 100380, total loss is 15.06 , inference loss is 6.99, weight deacy loss is 8.07, training accuracy is 0.218750, time 55.797 samples/sec
+epoch 6, total_step 100400, total loss is 20.64 , inference loss is 12.57, weight deacy loss is 8.07, training accuracy is 0.312500, time 56.431 samples/sec
+epoch 6, total_step 100420, total loss is 16.95 , inference loss is 8.88, weight deacy loss is 8.07, training accuracy is 0.375000, time 56.664 samples/sec
+epoch 6, total_step 100440, total loss is 16.38 , inference loss is 8.31, weight deacy loss is 8.07, training accuracy is 0.343750, time 56.754 samples/sec
+epoch 6, total_step 100460, total loss is 21.85 , inference loss is 13.78, weight deacy loss is 8.07, training accuracy is 0.250000, time 56.793 samples/sec
+epoch 6, total_step 100480, total loss is 19.11 , inference loss is 11.04, weight deacy loss is 8.07, training accuracy is 0.250000, time 56.228 samples/sec
+epoch 6, total_step 100500, total loss is 20.15 , inference loss is 12.08, weight deacy loss is 8.07, training accuracy is 0.343750, time 56.293 samples/sec
+epoch 6, total_step 100520, total loss is 22.31 , inference loss is 14.24, weight deacy loss is 8.07, training accuracy is 0.250000, time 55.625 samples/sec
+epoch 6, total_step 100540, total loss is 15.95 , inference loss is 7.88, weight deacy loss is 8.07, training accuracy is 0.375000, time 56.155 samples/sec
+epoch 6, total_step 100560, total loss is 20.02 , inference loss is 11.96, weight deacy loss is 8.07, training accuracy is 0.312500, time 56.813 samples/sec
+epoch 6, total_step 100580, total loss is 20.59 , inference loss is 12.53, weight deacy loss is 8.07, training accuracy is 0.375000, time 56.626 samples/sec
+epoch 6, total_step 100600, total loss is 21.04 , inference loss is 12.97, weight deacy loss is 8.07, training accuracy is 0.187500, time 56.236 samples/sec
+epoch 6, total_step 100620, total loss is 17.21 , inference loss is 9.14, weight deacy loss is 8.07, training accuracy is 0.406250, time 55.959 samples/sec
+epoch 6, total_step 100640, total loss is 17.90 , inference loss is 9.83, weight deacy loss is 8.06, training accuracy is 0.437500, time 56.905 samples/sec
+epoch 6, total_step 100660, total loss is 20.00 , inference loss is 11.94, weight deacy loss is 8.06, training accuracy is 0.187500, time 56.359 samples/sec
+epoch 6, total_step 100680, total loss is 15.67 , inference loss is 7.61, weight deacy loss is 8.06, training accuracy is 0.375000, time 56.961 samples/sec
+epoch 6, total_step 100700, total loss is 19.35 , inference loss is 11.29, weight deacy loss is 8.06, training accuracy is 0.406250, time 56.517 samples/sec
+epoch 6, total_step 100720, total loss is 13.27 , inference loss is 5.21, weight deacy loss is 8.06, training accuracy is 0.375000, time 55.491 samples/sec
+epoch 6, total_step 100740, total loss is 18.50 , inference loss is 10.44, weight deacy loss is 8.06, training accuracy is 0.312500, time 56.040 samples/sec
+epoch 6, total_step 100760, total loss is 19.72 , inference loss is 11.65, weight deacy loss is 8.06, training accuracy is 0.250000, time 56.882 samples/sec
+epoch 6, total_step 100780, total loss is 18.06 , inference loss is 10.00, weight deacy loss is 8.06, training accuracy is 0.281250, time 56.768 samples/sec
+epoch 6, total_step 100800, total loss is 18.56 , inference loss is 10.51, weight deacy loss is 8.06, training accuracy is 0.218750, time 56.413 samples/sec
+epoch 6, total_step 100820, total loss is 18.99 , inference loss is 10.94, weight deacy loss is 8.06, training accuracy is 0.312500, time 56.167 samples/sec
+epoch 6, total_step 100840, total loss is 20.14 , inference loss is 12.08, weight deacy loss is 8.06, training accuracy is 0.156250, time 56.684 samples/sec
+epoch 6, total_step 100860, total loss is 16.66 , inference loss is 8.60, weight deacy loss is 8.06, training accuracy is 0.437500, time 56.192 samples/sec
+epoch 6, total_step 100880, total loss is 18.23 , inference loss is 10.17, weight deacy loss is 8.06, training accuracy is 0.312500, time 57.263 samples/sec
+epoch 6, total_step 100900, total loss is 15.84 , inference loss is 7.79, weight deacy loss is 8.06, training accuracy is 0.343750, time 56.368 samples/sec
+epoch 6, total_step 100920, total loss is 17.51 , inference loss is 9.46, weight deacy loss is 8.05, training accuracy is 0.406250, time 57.007 samples/sec
+epoch 6, total_step 100940, total loss is 18.29 , inference loss is 10.24, weight deacy loss is 8.05, training accuracy is 0.281250, time 56.476 samples/sec
+epoch 6, total_step 100960, total loss is 18.92 , inference loss is 10.87, weight deacy loss is 8.05, training accuracy is 0.250000, time 57.131 samples/sec
+epoch 6, total_step 100980, total loss is 19.09 , inference loss is 11.03, weight deacy loss is 8.05, training accuracy is 0.312500, time 56.359 samples/sec
+epoch 6, total_step 101000, total loss is 21.52 , inference loss is 13.47, weight deacy loss is 8.05, training accuracy is 0.281250, time 57.159 samples/sec
+epoch 6, total_step 101020, total loss is 14.48 , inference loss is 6.43, weight deacy loss is 8.05, training accuracy is 0.375000, time 56.476 samples/sec
+epoch 6, total_step 101040, total loss is 17.11 , inference loss is 9.06, weight deacy loss is 8.05, training accuracy is 0.250000, time 56.236 samples/sec
+epoch 6, total_step 101060, total loss is 17.61 , inference loss is 9.56, weight deacy loss is 8.05, training accuracy is 0.375000, time 57.078 samples/sec
+epoch 6, total_step 101080, total loss is 18.31 , inference loss is 10.26, weight deacy loss is 8.05, training accuracy is 0.375000, time 56.398 samples/sec
+epoch 6, total_step 101100, total loss is 15.07 , inference loss is 7.02, weight deacy loss is 8.05, training accuracy is 0.437500, time 56.390 samples/sec
+epoch 6, total_step 101120, total loss is 15.83 , inference loss is 7.78, weight deacy loss is 8.05, training accuracy is 0.218750, time 56.216 samples/sec
+epoch 6, total_step 101140, total loss is 22.25 , inference loss is 14.20, weight deacy loss is 8.05, training accuracy is 0.218750, time 56.274 samples/sec
+epoch 6, total_step 101160, total loss is 16.95 , inference loss is 8.90, weight deacy loss is 8.05, training accuracy is 0.187500, time 56.597 samples/sec
+epoch 6, total_step 101180, total loss is 14.30 , inference loss is 6.25, weight deacy loss is 8.05, training accuracy is 0.468750, time 57.233 samples/sec
+epoch 6, total_step 101200, total loss is 20.63 , inference loss is 12.59, weight deacy loss is 8.04, training accuracy is 0.281250, time 56.363 samples/sec
+epoch 6, total_step 101220, total loss is 18.69 , inference loss is 10.65, weight deacy loss is 8.04, training accuracy is 0.250000, time 57.030 samples/sec
+epoch 6, total_step 101240, total loss is 18.56 , inference loss is 10.52, weight deacy loss is 8.04, training accuracy is 0.250000, time 56.843 samples/sec
+epoch 6, total_step 101260, total loss is 16.17 , inference loss is 8.13, weight deacy loss is 8.04, training accuracy is 0.343750, time 56.530 samples/sec
+epoch 6, total_step 101280, total loss is 16.90 , inference loss is 8.86, weight deacy loss is 8.04, training accuracy is 0.281250, time 56.936 samples/sec
+epoch 6, total_step 101300, total loss is 14.82 , inference loss is 6.78, weight deacy loss is 8.04, training accuracy is 0.343750, time 56.088 samples/sec
+epoch 6, total_step 101320, total loss is 16.26 , inference loss is 8.22, weight deacy loss is 8.04, training accuracy is 0.437500, time 56.698 samples/sec
+epoch 6, total_step 101340, total loss is 18.88 , inference loss is 10.84, weight deacy loss is 8.04, training accuracy is 0.531250, time 56.116 samples/sec
+epoch 6, total_step 101360, total loss is 17.72 , inference loss is 9.68, weight deacy loss is 8.04, training accuracy is 0.343750, time 56.320 samples/sec
+epoch 6, total_step 101380, total loss is 15.60 , inference loss is 7.56, weight deacy loss is 8.04, training accuracy is 0.375000, time 56.738 samples/sec
+epoch 6, total_step 101400, total loss is 17.96 , inference loss is 9.92, weight deacy loss is 8.04, training accuracy is 0.375000, time 55.578 samples/sec
+epoch 6, total_step 101420, total loss is 14.79 , inference loss is 6.76, weight deacy loss is 8.04, training accuracy is 0.312500, time 56.454 samples/sec
+epoch 6, total_step 101440, total loss is 18.08 , inference loss is 10.04, weight deacy loss is 8.04, training accuracy is 0.187500, time 56.756 samples/sec
+epoch 6, total_step 101460, total loss is 19.09 , inference loss is 11.05, weight deacy loss is 8.04, training accuracy is 0.375000, time 56.182 samples/sec
+epoch 6, total_step 101480, total loss is 16.48 , inference loss is 8.44, weight deacy loss is 8.04, training accuracy is 0.468750, time 56.388 samples/sec
+epoch 6, total_step 101500, total loss is 18.06 , inference loss is 10.02, weight deacy loss is 8.03, training accuracy is 0.343750, time 56.628 samples/sec
+epoch 6, total_step 101520, total loss is 19.90 , inference loss is 11.87, weight deacy loss is 8.03, training accuracy is 0.281250, time 56.456 samples/sec
+epoch 6, total_step 101540, total loss is 16.64 , inference loss is 8.60, weight deacy loss is 8.03, training accuracy is 0.437500, time 55.752 samples/sec
+epoch 6, total_step 101560, total loss is 19.90 , inference loss is 11.87, weight deacy loss is 8.03, training accuracy is 0.406250, time 56.088 samples/sec
+epoch 6, total_step 101580, total loss is 17.19 , inference loss is 9.16, weight deacy loss is 8.03, training accuracy is 0.281250, time 56.099 samples/sec
+epoch 6, total_step 101600, total loss is 15.32 , inference loss is 7.29, weight deacy loss is 8.03, training accuracy is 0.375000, time 56.490 samples/sec
+epoch 6, total_step 101620, total loss is 16.37 , inference loss is 8.34, weight deacy loss is 8.03, training accuracy is 0.312500, time 55.474 samples/sec
+epoch 6, total_step 101640, total loss is 18.64 , inference loss is 10.61, weight deacy loss is 8.03, training accuracy is 0.281250, time 56.410 samples/sec
+epoch 6, total_step 101660, total loss is 17.93 , inference loss is 9.90, weight deacy loss is 8.03, training accuracy is 0.312500, time 56.664 samples/sec
+epoch 6, total_step 101680, total loss is 16.41 , inference loss is 8.38, weight deacy loss is 8.03, training accuracy is 0.343750, time 56.427 samples/sec
+epoch 6, total_step 101700, total loss is 20.10 , inference loss is 12.07, weight deacy loss is 8.03, training accuracy is 0.343750, time 57.118 samples/sec
+epoch 6, total_step 101720, total loss is 16.86 , inference loss is 8.84, weight deacy loss is 8.03, training accuracy is 0.312500, time 56.444 samples/sec
+epoch 6, total_step 101740, total loss is 20.97 , inference loss is 12.95, weight deacy loss is 8.03, training accuracy is 0.312500, time 55.649 samples/sec
+epoch 6, total_step 101760, total loss is 15.91 , inference loss is 7.88, weight deacy loss is 8.03, training accuracy is 0.468750, time 56.741 samples/sec
+epoch 6, total_step 101780, total loss is 19.23 , inference loss is 11.21, weight deacy loss is 8.02, training accuracy is 0.281250, time 54.953 samples/sec
+epoch 6, total_step 101800, total loss is 16.74 , inference loss is 8.72, weight deacy loss is 8.02, training accuracy is 0.281250, time 56.227 samples/sec
+epoch 6, total_step 101820, total loss is 18.64 , inference loss is 10.62, weight deacy loss is 8.02, training accuracy is 0.281250, time 56.998 samples/sec
+epoch 6, total_step 101840, total loss is 15.54 , inference loss is 7.52, weight deacy loss is 8.02, training accuracy is 0.312500, time 56.680 samples/sec
+epoch 6, total_step 101860, total loss is 15.76 , inference loss is 7.74, weight deacy loss is 8.02, training accuracy is 0.468750, time 57.363 samples/sec
+epoch 6, total_step 101880, total loss is 20.38 , inference loss is 12.36, weight deacy loss is 8.02, training accuracy is 0.250000, time 56.475 samples/sec
+epoch 6, total_step 101900, total loss is 18.64 , inference loss is 10.62, weight deacy loss is 8.02, training accuracy is 0.218750, time 55.858 samples/sec
+epoch 6, total_step 101920, total loss is 19.47 , inference loss is 11.45, weight deacy loss is 8.02, training accuracy is 0.437500, time 56.835 samples/sec
+epoch 6, total_step 101940, total loss is 17.30 , inference loss is 9.28, weight deacy loss is 8.02, training accuracy is 0.437500, time 56.549 samples/sec
+epoch 6, total_step 101960, total loss is 15.79 , inference loss is 7.77, weight deacy loss is 8.02, training accuracy is 0.343750, time 56.307 samples/sec
+epoch 6, total_step 101980, total loss is 17.86 , inference loss is 9.84, weight deacy loss is 8.02, training accuracy is 0.375000, time 57.475 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.99203299999998
+best_threshold_index 111 0.9842592592592593
+best_threshold_index 111 0.9842592592592593
+best_threshold_index 111 0.9842592592592593
+best_threshold_index 104 0.9855555555555555
+best_threshold_index 103 0.9837037037037037
+best_threshold_index 111 0.9829629629629629
+best_threshold_index 111 0.9842592592592593
+best_threshold_index 111 0.9829629629629629
+best_threshold_index 111 0.9829629629629629
+best_threshold_index 111 0.9831481481481481
+[lfw][102000]XNorm: 22.394541
+[lfw][102000]Accuracy-Flip: 0.98217+-0.00934
+testing verification..
+(14000, 512)
+infer time 85.72441000000002
+best_threshold_index 112 0.9847619047619047
+best_threshold_index 112 0.9834920634920635
+best_threshold_index 112 0.9828571428571429
+best_threshold_index 112 0.9838095238095238
+best_threshold_index 112 0.9853968253968254
+best_threshold_index 112 0.9834920634920635
+best_threshold_index 112 0.9842857142857143
+best_threshold_index 112 0.9831746031746031
+best_threshold_index 112 0.9834920634920635
+best_threshold_index 112 0.9838095238095238
+[cfp_ff][102000]XNorm: 21.747552
+[cfp_ff][102000]Accuracy-Flip: 0.98386+-0.00652
+testing verification..
+(14000, 512)
+infer time 85.73957900000006
+best_threshold_index 135 0.910952380952381
+best_threshold_index 135 0.9123809523809524
+best_threshold_index 135 0.9098412698412699
+best_threshold_index 135 0.915079365079365
+best_threshold_index 135 0.9166666666666666
+best_threshold_index 135 0.9122222222222223
+best_threshold_index 135 0.9087301587301587
+best_threshold_index 135 0.9114285714285715
+best_threshold_index 135 0.9134920634920635
+best_threshold_index 135 0.912063492063492
+[cfp_fp][102000]XNorm: 21.360270
+[cfp_fp][102000]Accuracy-Flip: 0.91229+-0.02005
+testing verification..
+(12000, 512)
+infer time 73.53119499999997
+best_threshold_index 133 0.8992592592592593
+best_threshold_index 133 0.8962962962962963
+best_threshold_index 133 0.8988888888888888
+best_threshold_index 133 0.8998148148148148
+best_threshold_index 133 0.9037037037037037
+best_threshold_index 133 0.899074074074074
+best_threshold_index 135 0.8981481481481481
+best_threshold_index 133 0.9055555555555556
+best_threshold_index 131 0.9022222222222223
+best_threshold_index 133 0.9027777777777778
+[agedb_30][102000]XNorm: 22.398677
+[agedb_30][102000]Accuracy-Flip: 0.89850+-0.02440
+test accuracy is: 0.9821666666666667
+epoch 6, total_step 102000, total loss is 17.55 , inference loss is 9.53, weight deacy loss is 8.02, training accuracy is 0.218750, time 53.739 samples/sec
+epoch 6, total_step 102020, total loss is 17.53 , inference loss is 9.51, weight deacy loss is 8.02, training accuracy is 0.375000, time 57.506 samples/sec
+epoch 6, total_step 102040, total loss is 15.61 , inference loss is 7.60, weight deacy loss is 8.02, training accuracy is 0.312500, time 56.972 samples/sec
+epoch 6, total_step 102060, total loss is 15.38 , inference loss is 7.37, weight deacy loss is 8.01, training accuracy is 0.343750, time 56.722 samples/sec
+epoch 6, total_step 102080, total loss is 21.50 , inference loss is 13.48, weight deacy loss is 8.01, training accuracy is 0.312500, time 56.212 samples/sec
+epoch 6, total_step 102100, total loss is 17.35 , inference loss is 9.33, weight deacy loss is 8.01, training accuracy is 0.531250, time 57.212 samples/sec
+epoch 6, total_step 102120, total loss is 18.75 , inference loss is 10.74, weight deacy loss is 8.01, training accuracy is 0.281250, time 56.449 samples/sec
+epoch 6, total_step 102140, total loss is 18.29 , inference loss is 10.28, weight deacy loss is 8.01, training accuracy is 0.250000, time 57.138 samples/sec
+epoch 6, total_step 102160, total loss is 18.23 , inference loss is 10.22, weight deacy loss is 8.01, training accuracy is 0.343750, time 56.503 samples/sec
+epoch 6, total_step 102180, total loss is 18.93 , inference loss is 10.92, weight deacy loss is 8.01, training accuracy is 0.312500, time 56.682 samples/sec
+epoch 6, total_step 102200, total loss is 14.87 , inference loss is 6.86, weight deacy loss is 8.01, training accuracy is 0.468750, time 56.464 samples/sec
+epoch 6, total_step 102220, total loss is 18.86 , inference loss is 10.85, weight deacy loss is 8.01, training accuracy is 0.312500, time 56.503 samples/sec
+epoch 6, total_step 102240, total loss is 15.83 , inference loss is 7.83, weight deacy loss is 8.01, training accuracy is 0.437500, time 56.563 samples/sec
+epoch 6, total_step 102260, total loss is 17.71 , inference loss is 9.70, weight deacy loss is 8.01, training accuracy is 0.375000, time 56.521 samples/sec
+epoch 6, total_step 102280, total loss is 15.99 , inference loss is 7.98, weight deacy loss is 8.01, training accuracy is 0.343750, time 56.571 samples/sec
+epoch 6, total_step 102300, total loss is 15.77 , inference loss is 7.76, weight deacy loss is 8.01, training accuracy is 0.281250, time 56.653 samples/sec
+epoch 6, total_step 102320, total loss is 17.96 , inference loss is 9.95, weight deacy loss is 8.01, training accuracy is 0.406250, time 56.980 samples/sec
+epoch 6, total_step 102340, total loss is 19.31 , inference loss is 11.31, weight deacy loss is 8.00, training accuracy is 0.218750, time 56.486 samples/sec
+epoch 6, total_step 102360, total loss is 17.25 , inference loss is 9.24, weight deacy loss is 8.00, training accuracy is 0.500000, time 56.560 samples/sec
+epoch 6, total_step 102380, total loss is 15.93 , inference loss is 7.93, weight deacy loss is 8.00, training accuracy is 0.406250, time 56.567 samples/sec
+epoch 6, total_step 102400, total loss is 17.34 , inference loss is 9.33, weight deacy loss is 8.00, training accuracy is 0.281250, time 56.980 samples/sec
+epoch 6, total_step 102420, total loss is 17.16 , inference loss is 9.16, weight deacy loss is 8.00, training accuracy is 0.250000, time 56.996 samples/sec
+epoch 6, total_step 102440, total loss is 20.03 , inference loss is 12.03, weight deacy loss is 8.00, training accuracy is 0.281250, time 56.834 samples/sec
+epoch 6, total_step 102460, total loss is 17.29 , inference loss is 9.29, weight deacy loss is 8.00, training accuracy is 0.343750, time 55.891 samples/sec
+epoch 6, total_step 102480, total loss is 16.06 , inference loss is 8.07, weight deacy loss is 8.00, training accuracy is 0.406250, time 56.213 samples/sec
+epoch 6, total_step 102500, total loss is 18.93 , inference loss is 10.93, weight deacy loss is 8.00, training accuracy is 0.406250, time 56.582 samples/sec
+epoch 6, total_step 102520, total loss is 17.37 , inference loss is 9.37, weight deacy loss is 8.00, training accuracy is 0.406250, time 56.381 samples/sec
+epoch 6, total_step 102540, total loss is 14.54 , inference loss is 6.54, weight deacy loss is 8.00, training accuracy is 0.468750, time 56.528 samples/sec
+epoch 6, total_step 102560, total loss is 17.14 , inference loss is 9.15, weight deacy loss is 8.00, training accuracy is 0.343750, time 56.412 samples/sec
+epoch 6, total_step 102580, total loss is 18.31 , inference loss is 10.31, weight deacy loss is 8.00, training accuracy is 0.375000, time 56.851 samples/sec
+epoch 6, total_step 102600, total loss is 18.78 , inference loss is 10.78, weight deacy loss is 8.00, training accuracy is 0.343750, time 55.910 samples/sec
+epoch 6, total_step 102620, total loss is 16.64 , inference loss is 8.65, weight deacy loss is 7.99, training accuracy is 0.375000, time 56.880 samples/sec
+epoch 6, total_step 102640, total loss is 20.44 , inference loss is 12.44, weight deacy loss is 7.99, training accuracy is 0.281250, time 56.268 samples/sec
+epoch 6, total_step 102660, total loss is 23.07 , inference loss is 15.08, weight deacy loss is 7.99, training accuracy is 0.281250, time 57.079 samples/sec
+epoch 6, total_step 102680, total loss is 14.06 , inference loss is 6.06, weight deacy loss is 7.99, training accuracy is 0.437500, time 56.668 samples/sec
+epoch 6, total_step 102700, total loss is 19.23 , inference loss is 11.24, weight deacy loss is 7.99, training accuracy is 0.312500, time 56.046 samples/sec
+epoch 6, total_step 102720, total loss is 16.30 , inference loss is 8.31, weight deacy loss is 7.99, training accuracy is 0.406250, time 56.844 samples/sec
+epoch 6, total_step 102740, total loss is 16.44 , inference loss is 8.45, weight deacy loss is 7.99, training accuracy is 0.406250, time 56.931 samples/sec
+epoch 6, total_step 102760, total loss is 19.79 , inference loss is 11.80, weight deacy loss is 7.99, training accuracy is 0.281250, time 56.270 samples/sec
+epoch 6, total_step 102780, total loss is 22.65 , inference loss is 14.66, weight deacy loss is 7.99, training accuracy is 0.093750, time 55.912 samples/sec
+epoch 6, total_step 102800, total loss is 23.34 , inference loss is 15.35, weight deacy loss is 7.99, training accuracy is 0.156250, time 56.619 samples/sec
+epoch 6, total_step 102820, total loss is 16.12 , inference loss is 8.14, weight deacy loss is 7.99, training accuracy is 0.187500, time 57.000 samples/sec
+epoch 6, total_step 102840, total loss is 17.24 , inference loss is 9.25, weight deacy loss is 7.99, training accuracy is 0.187500, time 56.770 samples/sec
+epoch 6, total_step 102860, total loss is 17.22 , inference loss is 9.23, weight deacy loss is 7.99, training accuracy is 0.312500, time 56.323 samples/sec
+epoch 6, total_step 102880, total loss is 14.47 , inference loss is 6.48, weight deacy loss is 7.99, training accuracy is 0.375000, time 56.783 samples/sec
+epoch 6, total_step 102900, total loss is 16.09 , inference loss is 8.10, weight deacy loss is 7.98, training accuracy is 0.375000, time 54.439 samples/sec
+epoch 6, total_step 102920, total loss is 18.63 , inference loss is 10.65, weight deacy loss is 7.98, training accuracy is 0.281250, time 56.980 samples/sec
+epoch 6, total_step 102940, total loss is 17.50 , inference loss is 9.52, weight deacy loss is 7.98, training accuracy is 0.281250, time 56.339 samples/sec
+epoch 6, total_step 102960, total loss is 16.59 , inference loss is 8.61, weight deacy loss is 7.98, training accuracy is 0.281250, time 56.703 samples/sec
+epoch 6, total_step 102980, total loss is 21.36 , inference loss is 13.38, weight deacy loss is 7.98, training accuracy is 0.250000, time 56.552 samples/sec
+epoch 6, total_step 103000, total loss is 17.58 , inference loss is 9.60, weight deacy loss is 7.98, training accuracy is 0.343750, time 57.215 samples/sec
+epoch 6, total_step 103020, total loss is 19.09 , inference loss is 11.11, weight deacy loss is 7.98, training accuracy is 0.312500, time 57.230 samples/sec
+epoch 6, total_step 103040, total loss is 18.41 , inference loss is 10.43, weight deacy loss is 7.98, training accuracy is 0.406250, time 56.251 samples/sec
+epoch 6, total_step 103060, total loss is 19.97 , inference loss is 11.99, weight deacy loss is 7.98, training accuracy is 0.218750, time 56.427 samples/sec
+epoch 6, total_step 103080, total loss is 15.96 , inference loss is 7.98, weight deacy loss is 7.98, training accuracy is 0.312500, time 55.064 samples/sec
+epoch 6, total_step 103100, total loss is 17.04 , inference loss is 9.06, weight deacy loss is 7.98, training accuracy is 0.250000, time 56.422 samples/sec
+epoch 6, total_step 103120, total loss is 15.38 , inference loss is 7.41, weight deacy loss is 7.98, training accuracy is 0.375000, time 57.162 samples/sec
+epoch 6, total_step 103140, total loss is 20.25 , inference loss is 12.28, weight deacy loss is 7.98, training accuracy is 0.312500, time 56.479 samples/sec
+epoch 6, total_step 103160, total loss is 14.56 , inference loss is 6.58, weight deacy loss is 7.98, training accuracy is 0.468750, time 56.653 samples/sec
+epoch 6, total_step 103180, total loss is 18.93 , inference loss is 10.96, weight deacy loss is 7.97, training accuracy is 0.312500, time 57.072 samples/sec
+epoch 6, total_step 103200, total loss is 18.86 , inference loss is 10.89, weight deacy loss is 7.97, training accuracy is 0.343750, time 56.797 samples/sec
+epoch 6, total_step 103220, total loss is 18.60 , inference loss is 10.62, weight deacy loss is 7.97, training accuracy is 0.281250, time 55.803 samples/sec
+epoch 6, total_step 103240, total loss is 17.44 , inference loss is 9.46, weight deacy loss is 7.97, training accuracy is 0.312500, time 56.700 samples/sec
+epoch 6, total_step 103260, total loss is 16.98 , inference loss is 9.01, weight deacy loss is 7.97, training accuracy is 0.375000, time 56.976 samples/sec
+epoch 6, total_step 103280, total loss is 17.81 , inference loss is 9.84, weight deacy loss is 7.97, training accuracy is 0.281250, time 57.248 samples/sec
+epoch 6, total_step 103300, total loss is 18.01 , inference loss is 10.04, weight deacy loss is 7.97, training accuracy is 0.343750, time 57.071 samples/sec
+epoch 6, total_step 103320, total loss is 19.98 , inference loss is 12.01, weight deacy loss is 7.97, training accuracy is 0.312500, time 57.295 samples/sec
+epoch 6, total_step 103340, total loss is 18.62 , inference loss is 10.65, weight deacy loss is 7.97, training accuracy is 0.312500, time 56.860 samples/sec
+epoch 6, total_step 103360, total loss is 19.56 , inference loss is 11.60, weight deacy loss is 7.97, training accuracy is 0.250000, time 56.396 samples/sec
+epoch 6, total_step 103380, total loss is 18.45 , inference loss is 10.48, weight deacy loss is 7.97, training accuracy is 0.250000, time 56.920 samples/sec
+epoch 6, total_step 103400, total loss is 17.81 , inference loss is 9.84, weight deacy loss is 7.97, training accuracy is 0.281250, time 56.744 samples/sec
+epoch 6, total_step 103420, total loss is 15.54 , inference loss is 7.57, weight deacy loss is 7.97, training accuracy is 0.406250, time 56.671 samples/sec
+epoch 6, total_step 103440, total loss is 16.68 , inference loss is 8.71, weight deacy loss is 7.97, training accuracy is 0.343750, time 56.401 samples/sec
+epoch 6, total_step 103460, total loss is 17.81 , inference loss is 9.85, weight deacy loss is 7.97, training accuracy is 0.343750, time 57.320 samples/sec
+epoch 6, total_step 103480, total loss is 16.48 , inference loss is 8.52, weight deacy loss is 7.96, training accuracy is 0.281250, time 56.888 samples/sec
+epoch 6, total_step 103500, total loss is 17.11 , inference loss is 9.14, weight deacy loss is 7.96, training accuracy is 0.156250, time 56.211 samples/sec
+epoch 6, total_step 103520, total loss is 17.69 , inference loss is 9.72, weight deacy loss is 7.96, training accuracy is 0.343750, time 56.504 samples/sec
+epoch 6, total_step 103540, total loss is 17.25 , inference loss is 9.29, weight deacy loss is 7.96, training accuracy is 0.156250, time 56.211 samples/sec
+epoch 6, total_step 103560, total loss is 16.20 , inference loss is 8.24, weight deacy loss is 7.96, training accuracy is 0.312500, time 56.839 samples/sec
+epoch 6, total_step 103580, total loss is 21.86 , inference loss is 13.90, weight deacy loss is 7.96, training accuracy is 0.250000, time 57.062 samples/sec
+epoch 6, total_step 103600, total loss is 19.14 , inference loss is 11.18, weight deacy loss is 7.96, training accuracy is 0.343750, time 56.796 samples/sec
+epoch 6, total_step 103620, total loss is 18.75 , inference loss is 10.79, weight deacy loss is 7.96, training accuracy is 0.343750, time 56.925 samples/sec
+epoch 6, total_step 103640, total loss is 17.99 , inference loss is 10.03, weight deacy loss is 7.96, training accuracy is 0.375000, time 56.363 samples/sec
+epoch 6, total_step 103660, total loss is 20.64 , inference loss is 12.68, weight deacy loss is 7.96, training accuracy is 0.281250, time 56.161 samples/sec
+epoch 6, total_step 103680, total loss is 21.76 , inference loss is 13.80, weight deacy loss is 7.96, training accuracy is 0.218750, time 56.758 samples/sec
+epoch 6, total_step 103700, total loss is 16.72 , inference loss is 8.76, weight deacy loss is 7.96, training accuracy is 0.343750, time 56.956 samples/sec
+epoch 6, total_step 103720, total loss is 17.46 , inference loss is 9.50, weight deacy loss is 7.96, training accuracy is 0.375000, time 56.947 samples/sec
+epoch 6, total_step 103740, total loss is 18.80 , inference loss is 10.84, weight deacy loss is 7.96, training accuracy is 0.375000, time 56.338 samples/sec
+epoch 6, total_step 103760, total loss is 16.31 , inference loss is 8.36, weight deacy loss is 7.95, training accuracy is 0.281250, time 56.346 samples/sec
+epoch 6, total_step 103780, total loss is 19.24 , inference loss is 11.28, weight deacy loss is 7.95, training accuracy is 0.281250, time 56.348 samples/sec
+epoch 6, total_step 103800, total loss is 13.54 , inference loss is 5.59, weight deacy loss is 7.95, training accuracy is 0.531250, time 55.923 samples/sec
+epoch 6, total_step 103820, total loss is 24.70 , inference loss is 16.75, weight deacy loss is 7.95, training accuracy is 0.093750, time 55.341 samples/sec
+epoch 6, total_step 103840, total loss is 23.83 , inference loss is 15.88, weight deacy loss is 7.95, training accuracy is 0.281250, time 56.813 samples/sec
+epoch 6, total_step 103860, total loss is 16.08 , inference loss is 8.13, weight deacy loss is 7.95, training accuracy is 0.375000, time 56.555 samples/sec
+epoch 6, total_step 103880, total loss is 15.44 , inference loss is 7.49, weight deacy loss is 7.95, training accuracy is 0.312500, time 56.819 samples/sec
+epoch 6, total_step 103900, total loss is 17.41 , inference loss is 9.46, weight deacy loss is 7.95, training accuracy is 0.375000, time 55.927 samples/sec
+epoch 6, total_step 103920, total loss is 18.27 , inference loss is 10.32, weight deacy loss is 7.95, training accuracy is 0.218750, time 56.720 samples/sec
+epoch 6, total_step 103940, total loss is 20.20 , inference loss is 12.25, weight deacy loss is 7.95, training accuracy is 0.250000, time 56.477 samples/sec
+epoch 6, total_step 103960, total loss is 19.52 , inference loss is 11.57, weight deacy loss is 7.95, training accuracy is 0.250000, time 58.033 samples/sec
+epoch 6, total_step 103980, total loss is 21.42 , inference loss is 13.47, weight deacy loss is 7.95, training accuracy is 0.187500, time 56.343 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.42275599999999
+best_threshold_index 106 0.9844444444444445
+best_threshold_index 106 0.9835185185185186
+best_threshold_index 109 0.985
+best_threshold_index 106 0.9848148148148148
+best_threshold_index 106 0.9844444444444445
+best_threshold_index 109 0.9835185185185186
+best_threshold_index 106 0.9835185185185186
+best_threshold_index 106 0.9831481481481481
+best_threshold_index 106 0.9824074074074074
+best_threshold_index 106 0.9827777777777778
+[lfw][104000]XNorm: 21.847632
+[lfw][104000]Accuracy-Flip: 0.98283+-0.00749
+testing verification..
+(14000, 512)
+infer time 85.65434799999991
+best_threshold_index 115 0.9853968253968254
+best_threshold_index 115 0.9842857142857143
+best_threshold_index 115 0.9839682539682539
+best_threshold_index 115 0.9841269841269841
+best_threshold_index 115 0.986031746031746
+best_threshold_index 115 0.9846031746031746
+best_threshold_index 115 0.9855555555555555
+best_threshold_index 115 0.9844444444444445
+best_threshold_index 115 0.9844444444444445
+best_threshold_index 115 0.9857142857142858
+[cfp_ff][104000]XNorm: 21.575961
+[cfp_ff][104000]Accuracy-Flip: 0.98486+-0.00633
+testing verification..
+(14000, 512)
+infer time 85.92357499999999
+best_threshold_index 131 0.9152380952380952
+best_threshold_index 131 0.9163492063492064
+best_threshold_index 132 0.9130158730158731
+best_threshold_index 132 0.9157142857142857
+best_threshold_index 132 0.9185714285714286
+best_threshold_index 132 0.9131746031746032
+best_threshold_index 132 0.9136507936507936
+best_threshold_index 132 0.915079365079365
+best_threshold_index 132 0.9134920634920635
+best_threshold_index 132 0.9157142857142857
+[cfp_fp][104000]XNorm: 21.398488
+[cfp_fp][104000]Accuracy-Flip: 0.91443+-0.01511
+testing verification..
+(12000, 512)
+infer time 73.72945899999995
+best_threshold_index 139 0.8975925925925926
+best_threshold_index 138 0.894074074074074
+best_threshold_index 139 0.8953703703703704
+best_threshold_index 139 0.8972222222222223
+best_threshold_index 139 0.9
+best_threshold_index 138 0.8953703703703704
+best_threshold_index 137 0.897962962962963
+best_threshold_index 139 0.9027777777777778
+best_threshold_index 139 0.9018518518518519
+best_threshold_index 139 0.8998148148148148
+[agedb_30][104000]XNorm: 22.167424
+[agedb_30][104000]Accuracy-Flip: 0.89667+-0.02373
+test accuracy is: 0.9828333333333333
+epoch 6, total_step 104000, total loss is 16.83 , inference loss is 8.89, weight deacy loss is 7.95, training accuracy is 0.218750, time 52.975 samples/sec
+epoch 6, total_step 104020, total loss is 22.14 , inference loss is 14.19, weight deacy loss is 7.95, training accuracy is 0.187500, time 56.132 samples/sec
+epoch 6, total_step 104040, total loss is 18.87 , inference loss is 10.93, weight deacy loss is 7.95, training accuracy is 0.156250, time 56.767 samples/sec
+epoch 6, total_step 104060, total loss is 20.25 , inference loss is 12.31, weight deacy loss is 7.94, training accuracy is 0.250000, time 56.832 samples/sec
+epoch 6, total_step 104080, total loss is 18.08 , inference loss is 10.13, weight deacy loss is 7.94, training accuracy is 0.250000, time 56.762 samples/sec
+epoch 6, total_step 104100, total loss is 17.01 , inference loss is 9.07, weight deacy loss is 7.94, training accuracy is 0.281250, time 56.886 samples/sec
+epoch 6, total_step 104120, total loss is 16.89 , inference loss is 8.95, weight deacy loss is 7.94, training accuracy is 0.250000, time 56.945 samples/sec
+epoch 6, total_step 104140, total loss is 20.76 , inference loss is 12.82, weight deacy loss is 7.94, training accuracy is 0.281250, time 56.717 samples/sec
+epoch 6, total_step 104160, total loss is 20.67 , inference loss is 12.73, weight deacy loss is 7.94, training accuracy is 0.281250, time 56.228 samples/sec
+epoch 6, total_step 104180, total loss is 18.27 , inference loss is 10.33, weight deacy loss is 7.94, training accuracy is 0.375000, time 56.756 samples/sec
+epoch 6, total_step 104200, total loss is 17.12 , inference loss is 9.18, weight deacy loss is 7.94, training accuracy is 0.218750, time 55.409 samples/sec
+epoch 6, total_step 104220, total loss is 19.20 , inference loss is 11.26, weight deacy loss is 7.94, training accuracy is 0.312500, time 56.889 samples/sec
+epoch 6, total_step 104240, total loss is 18.75 , inference loss is 10.81, weight deacy loss is 7.94, training accuracy is 0.312500, time 55.901 samples/sec
+epoch 6, total_step 104260, total loss is 18.12 , inference loss is 10.19, weight deacy loss is 7.94, training accuracy is 0.312500, time 56.302 samples/sec
+epoch 6, total_step 104280, total loss is 17.01 , inference loss is 9.07, weight deacy loss is 7.94, training accuracy is 0.375000, time 56.507 samples/sec
+epoch 6, total_step 104300, total loss is 20.75 , inference loss is 12.81, weight deacy loss is 7.94, training accuracy is 0.281250, time 56.897 samples/sec
+epoch 6, total_step 104320, total loss is 19.05 , inference loss is 11.11, weight deacy loss is 7.94, training accuracy is 0.250000, time 56.767 samples/sec
+epoch 6, total_step 104340, total loss is 17.86 , inference loss is 9.92, weight deacy loss is 7.93, training accuracy is 0.312500, time 55.996 samples/sec
+epoch 6, total_step 104360, total loss is 21.40 , inference loss is 13.47, weight deacy loss is 7.93, training accuracy is 0.156250, time 57.203 samples/sec
+epoch 6, total_step 104380, total loss is 16.99 , inference loss is 9.05, weight deacy loss is 7.93, training accuracy is 0.218750, time 56.521 samples/sec
+epoch 6, total_step 104400, total loss is 19.55 , inference loss is 11.61, weight deacy loss is 7.93, training accuracy is 0.343750, time 55.870 samples/sec
+epoch 6, total_step 104420, total loss is 20.30 , inference loss is 12.36, weight deacy loss is 7.93, training accuracy is 0.125000, time 56.167 samples/sec
+epoch 6, total_step 104440, total loss is 18.36 , inference loss is 10.43, weight deacy loss is 7.93, training accuracy is 0.250000, time 55.998 samples/sec
+epoch 6, total_step 104460, total loss is 15.01 , inference loss is 7.08, weight deacy loss is 7.93, training accuracy is 0.343750, time 55.746 samples/sec
+epoch 6, total_step 104480, total loss is 16.77 , inference loss is 8.84, weight deacy loss is 7.93, training accuracy is 0.468750, time 56.627 samples/sec
+epoch 6, total_step 104500, total loss is 20.04 , inference loss is 12.11, weight deacy loss is 7.93, training accuracy is 0.250000, time 56.503 samples/sec
+epoch 6, total_step 104520, total loss is 19.27 , inference loss is 11.34, weight deacy loss is 7.93, training accuracy is 0.156250, time 56.155 samples/sec
+epoch 6, total_step 104540, total loss is 17.32 , inference loss is 9.39, weight deacy loss is 7.93, training accuracy is 0.312500, time 56.741 samples/sec
+epoch 6, total_step 104560, total loss is 22.28 , inference loss is 14.35, weight deacy loss is 7.93, training accuracy is 0.156250, time 56.198 samples/sec
+epoch 6, total_step 104580, total loss is 19.56 , inference loss is 11.63, weight deacy loss is 7.93, training accuracy is 0.375000, time 56.297 samples/sec
+epoch 6, total_step 104600, total loss is 14.92 , inference loss is 7.00, weight deacy loss is 7.93, training accuracy is 0.437500, time 54.633 samples/sec
+epoch 6, total_step 104620, total loss is 16.69 , inference loss is 8.76, weight deacy loss is 7.93, training accuracy is 0.312500, time 55.630 samples/sec
+epoch 6, total_step 104640, total loss is 17.05 , inference loss is 9.12, weight deacy loss is 7.92, training accuracy is 0.312500, time 56.522 samples/sec
+epoch 6, total_step 104660, total loss is 16.41 , inference loss is 8.49, weight deacy loss is 7.92, training accuracy is 0.187500, time 56.579 samples/sec
+epoch 6, total_step 104680, total loss is 22.36 , inference loss is 14.44, weight deacy loss is 7.92, training accuracy is 0.218750, time 56.279 samples/sec
+epoch 6, total_step 104700, total loss is 20.49 , inference loss is 12.57, weight deacy loss is 7.92, training accuracy is 0.281250, time 56.663 samples/sec
+epoch 6, total_step 104720, total loss is 20.03 , inference loss is 12.11, weight deacy loss is 7.92, training accuracy is 0.187500, time 56.121 samples/sec
+epoch 6, total_step 104740, total loss is 19.25 , inference loss is 11.33, weight deacy loss is 7.92, training accuracy is 0.156250, time 57.507 samples/sec
+epoch 6, total_step 104760, total loss is 23.28 , inference loss is 15.36, weight deacy loss is 7.92, training accuracy is 0.218750, time 56.282 samples/sec
+epoch 6, total_step 104780, total loss is 19.32 , inference loss is 11.40, weight deacy loss is 7.92, training accuracy is 0.281250, time 56.111 samples/sec
+epoch 6, total_step 104800, total loss is 19.56 , inference loss is 11.64, weight deacy loss is 7.92, training accuracy is 0.156250, time 56.765 samples/sec
+epoch 6, total_step 104820, total loss is 17.17 , inference loss is 9.25, weight deacy loss is 7.92, training accuracy is 0.312500, time 56.122 samples/sec
+epoch 6, total_step 104840, total loss is 15.15 , inference loss is 7.23, weight deacy loss is 7.92, training accuracy is 0.312500, time 56.934 samples/sec
+epoch 6, total_step 104860, total loss is 18.77 , inference loss is 10.85, weight deacy loss is 7.92, training accuracy is 0.218750, time 56.567 samples/sec
+epoch 6, total_step 104880, total loss is 16.22 , inference loss is 8.31, weight deacy loss is 7.92, training accuracy is 0.218750, time 56.586 samples/sec
+epoch 6, total_step 104900, total loss is 18.86 , inference loss is 10.95, weight deacy loss is 7.92, training accuracy is 0.281250, time 55.364 samples/sec
+epoch 6, total_step 104920, total loss is 15.18 , inference loss is 7.27, weight deacy loss is 7.91, training accuracy is 0.375000, time 57.118 samples/sec
+epoch 6, total_step 104940, total loss is 18.38 , inference loss is 10.47, weight deacy loss is 7.91, training accuracy is 0.281250, time 56.390 samples/sec
+epoch 6, total_step 104960, total loss is 19.78 , inference loss is 11.87, weight deacy loss is 7.91, training accuracy is 0.375000, time 56.041 samples/sec
+epoch 6, total_step 104980, total loss is 14.89 , inference loss is 6.98, weight deacy loss is 7.91, training accuracy is 0.312500, time 56.794 samples/sec
+epoch 6, total_step 105000, total loss is 21.15 , inference loss is 13.24, weight deacy loss is 7.91, training accuracy is 0.218750, time 56.116 samples/sec
+epoch 6, total_step 105020, total loss is 19.05 , inference loss is 11.14, weight deacy loss is 7.91, training accuracy is 0.281250, time 56.572 samples/sec
+epoch 6, total_step 105040, total loss is 19.65 , inference loss is 11.74, weight deacy loss is 7.91, training accuracy is 0.218750, time 56.062 samples/sec
+epoch 6, total_step 105060, total loss is 19.86 , inference loss is 11.95, weight deacy loss is 7.91, training accuracy is 0.187500, time 56.802 samples/sec
+epoch 6, total_step 105080, total loss is 22.86 , inference loss is 14.95, weight deacy loss is 7.91, training accuracy is 0.218750, time 57.688 samples/sec
+epoch 6, total_step 105100, total loss is 18.71 , inference loss is 10.80, weight deacy loss is 7.91, training accuracy is 0.187500, time 57.320 samples/sec
+epoch 6, total_step 105120, total loss is 19.05 , inference loss is 11.14, weight deacy loss is 7.91, training accuracy is 0.312500, time 56.932 samples/sec
+epoch 6, total_step 105140, total loss is 16.59 , inference loss is 8.68, weight deacy loss is 7.91, training accuracy is 0.343750, time 55.504 samples/sec
+epoch 6, total_step 105160, total loss is 20.43 , inference loss is 12.52, weight deacy loss is 7.91, training accuracy is 0.218750, time 56.495 samples/sec
+epoch 6, total_step 105180, total loss is 17.05 , inference loss is 9.14, weight deacy loss is 7.91, training accuracy is 0.250000, time 56.446 samples/sec
+epoch 6, total_step 105200, total loss is 19.25 , inference loss is 11.35, weight deacy loss is 7.91, training accuracy is 0.156250, time 56.663 samples/sec
+epoch 6, total_step 105220, total loss is 15.54 , inference loss is 7.64, weight deacy loss is 7.90, training accuracy is 0.343750, time 56.601 samples/sec
+epoch 6, total_step 105240, total loss is 20.51 , inference loss is 12.61, weight deacy loss is 7.90, training accuracy is 0.250000, time 55.979 samples/sec
+epoch 6, total_step 105260, total loss is 20.14 , inference loss is 12.24, weight deacy loss is 7.90, training accuracy is 0.187500, time 56.671 samples/sec
+epoch 6, total_step 105280, total loss is 23.45 , inference loss is 15.55, weight deacy loss is 7.90, training accuracy is 0.156250, time 56.783 samples/sec
+epoch 6, total_step 105300, total loss is 18.53 , inference loss is 10.63, weight deacy loss is 7.90, training accuracy is 0.156250, time 56.750 samples/sec
+epoch 6, total_step 105320, total loss is 21.92 , inference loss is 14.01, weight deacy loss is 7.90, training accuracy is 0.156250, time 57.168 samples/sec
+epoch 6, total_step 105340, total loss is 17.89 , inference loss is 9.99, weight deacy loss is 7.90, training accuracy is 0.187500, time 56.348 samples/sec
+epoch 6, total_step 105360, total loss is 17.62 , inference loss is 9.72, weight deacy loss is 7.90, training accuracy is 0.343750, time 56.298 samples/sec
+epoch 6, total_step 105380, total loss is 19.29 , inference loss is 11.39, weight deacy loss is 7.90, training accuracy is 0.250000, time 56.757 samples/sec
+epoch 6, total_step 105400, total loss is 16.61 , inference loss is 8.71, weight deacy loss is 7.90, training accuracy is 0.343750, time 56.781 samples/sec
+epoch 6, total_step 105420, total loss is 22.40 , inference loss is 14.51, weight deacy loss is 7.90, training accuracy is 0.250000, time 56.291 samples/sec
+epoch 6, total_step 105440, total loss is 21.08 , inference loss is 13.18, weight deacy loss is 7.90, training accuracy is 0.281250, time 56.285 samples/sec
+epoch 6, total_step 105460, total loss is 17.78 , inference loss is 9.88, weight deacy loss is 7.90, training accuracy is 0.281250, time 56.316 samples/sec
+epoch 6, total_step 105480, total loss is 16.87 , inference loss is 8.97, weight deacy loss is 7.90, training accuracy is 0.250000, time 55.840 samples/sec
+epoch 6, total_step 105500, total loss is 19.59 , inference loss is 11.69, weight deacy loss is 7.90, training accuracy is 0.218750, time 56.663 samples/sec
+epoch 6, total_step 105520, total loss is 24.97 , inference loss is 17.08, weight deacy loss is 7.89, training accuracy is 0.218750, time 56.710 samples/sec
+epoch 6, total_step 105540, total loss is 16.81 , inference loss is 8.92, weight deacy loss is 7.89, training accuracy is 0.375000, time 56.937 samples/sec
+epoch 6, total_step 105560, total loss is 21.58 , inference loss is 13.69, weight deacy loss is 7.89, training accuracy is 0.125000, time 56.206 samples/sec
+epoch 6, total_step 105580, total loss is 19.31 , inference loss is 11.42, weight deacy loss is 7.89, training accuracy is 0.187500, time 56.862 samples/sec
+epoch 6, total_step 105600, total loss is 20.94 , inference loss is 13.05, weight deacy loss is 7.89, training accuracy is 0.281250, time 54.469 samples/sec
+epoch 6, total_step 105620, total loss is 17.61 , inference loss is 9.72, weight deacy loss is 7.89, training accuracy is 0.187500, time 56.382 samples/sec
+epoch 6, total_step 105640, total loss is 24.17 , inference loss is 16.28, weight deacy loss is 7.89, training accuracy is 0.125000, time 56.208 samples/sec
+epoch 6, total_step 105660, total loss is 22.05 , inference loss is 14.16, weight deacy loss is 7.89, training accuracy is 0.125000, time 56.469 samples/sec
+epoch 6, total_step 105680, total loss is 18.98 , inference loss is 11.09, weight deacy loss is 7.89, training accuracy is 0.281250, time 56.643 samples/sec
+epoch 6, total_step 105700, total loss is 16.58 , inference loss is 8.69, weight deacy loss is 7.89, training accuracy is 0.281250, time 55.859 samples/sec
+epoch 6, total_step 105720, total loss is 20.22 , inference loss is 12.33, weight deacy loss is 7.89, training accuracy is 0.218750, time 56.636 samples/sec
+epoch 6, total_step 105740, total loss is 19.37 , inference loss is 11.48, weight deacy loss is 7.89, training accuracy is 0.343750, time 56.297 samples/sec
+epoch 6, total_step 105760, total loss is 16.01 , inference loss is 8.13, weight deacy loss is 7.89, training accuracy is 0.312500, time 55.794 samples/sec
+epoch 6, total_step 105780, total loss is 19.62 , inference loss is 11.73, weight deacy loss is 7.89, training accuracy is 0.156250, time 56.842 samples/sec
+epoch 6, total_step 105800, total loss is 18.22 , inference loss is 10.33, weight deacy loss is 7.88, training accuracy is 0.187500, time 56.580 samples/sec
+epoch 6, total_step 105820, total loss is 16.05 , inference loss is 8.16, weight deacy loss is 7.88, training accuracy is 0.343750, time 56.790 samples/sec
+epoch 6, total_step 105840, total loss is 14.99 , inference loss is 7.11, weight deacy loss is 7.88, training accuracy is 0.406250, time 56.486 samples/sec
+epoch 6, total_step 105860, total loss is 21.62 , inference loss is 13.73, weight deacy loss is 7.88, training accuracy is 0.125000, time 56.800 samples/sec
+epoch 6, total_step 105880, total loss is 19.53 , inference loss is 11.65, weight deacy loss is 7.88, training accuracy is 0.312500, time 56.593 samples/sec
+epoch 6, total_step 105900, total loss is 16.17 , inference loss is 8.29, weight deacy loss is 7.88, training accuracy is 0.187500, time 56.294 samples/sec
+epoch 6, total_step 105920, total loss is 16.94 , inference loss is 9.06, weight deacy loss is 7.88, training accuracy is 0.281250, time 56.076 samples/sec
+epoch 6, total_step 105940, total loss is 17.10 , inference loss is 9.22, weight deacy loss is 7.88, training accuracy is 0.187500, time 56.508 samples/sec
+epoch 6, total_step 105960, total loss is 20.11 , inference loss is 12.23, weight deacy loss is 7.88, training accuracy is 0.250000, time 57.072 samples/sec
+epoch 6, total_step 105980, total loss is 19.72 , inference loss is 11.84, weight deacy loss is 7.88, training accuracy is 0.312500, time 57.062 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.89231399999997
+best_threshold_index 106 0.9842592592592593
+best_threshold_index 110 0.9825925925925926
+best_threshold_index 105 0.9837037037037037
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 110 0.9837037037037037
+best_threshold_index 110 0.9825925925925926
+best_threshold_index 105 0.9831481481481481
+best_threshold_index 105 0.9824074074074074
+best_threshold_index 110 0.9822222222222222
+best_threshold_index 110 0.9824074074074074
+[lfw][106000]XNorm: 23.306064
+[lfw][106000]Accuracy-Flip: 0.98183+-0.00880
+testing verification..
+(14000, 512)
+infer time 85.47297800000004
+best_threshold_index 113 0.9866666666666667
+best_threshold_index 116 0.9863492063492063
+best_threshold_index 116 0.9858730158730159
+best_threshold_index 113 0.9853968253968254
+best_threshold_index 117 0.9873015873015873
+best_threshold_index 113 0.9855555555555555
+best_threshold_index 114 0.9863492063492063
+best_threshold_index 114 0.9853968253968254
+best_threshold_index 114 0.9855555555555555
+best_threshold_index 114 0.9861904761904762
+[cfp_ff][106000]XNorm: 23.098120
+[cfp_ff][106000]Accuracy-Flip: 0.98500+-0.00524
+testing verification..
+(14000, 512)
+infer time 85.78680199999992
+best_threshold_index 136 0.9188888888888889
+best_threshold_index 136 0.9188888888888889
+best_threshold_index 136 0.9177777777777778
+best_threshold_index 136 0.9211111111111111
+best_threshold_index 136 0.9223809523809524
+best_threshold_index 136 0.9171428571428571
+best_threshold_index 136 0.9160317460317461
+best_threshold_index 136 0.9184126984126985
+best_threshold_index 136 0.9180952380952381
+best_threshold_index 136 0.9198412698412698
+[cfp_fp][106000]XNorm: 22.412987
+[cfp_fp][106000]Accuracy-Flip: 0.91886+-0.01593
+testing verification..
+(12000, 512)
+infer time 73.56906700000002
+best_threshold_index 139 0.9003703703703704
+best_threshold_index 139 0.8977777777777778
+best_threshold_index 139 0.9001851851851852
+best_threshold_index 139 0.9014814814814814
+best_threshold_index 139 0.9064814814814814
+best_threshold_index 141 0.8992592592592593
+best_threshold_index 139 0.9009259259259259
+best_threshold_index 139 0.9068518518518518
+best_threshold_index 139 0.9042592592592592
+best_threshold_index 139 0.9025925925925926
+[agedb_30][106000]XNorm: 23.379545
+[agedb_30][106000]Accuracy-Flip: 0.90133+-0.02530
+test accuracy is: 0.9818333333333333
+epoch 6, total_step 106000, total loss is 21.03 , inference loss is 13.15, weight deacy loss is 7.88, training accuracy is 0.218750, time 53.643 samples/sec
+epoch 6, total_step 106020, total loss is 23.11 , inference loss is 15.24, weight deacy loss is 7.88, training accuracy is 0.218750, time 56.531 samples/sec
+epoch 6, total_step 106040, total loss is 21.23 , inference loss is 13.36, weight deacy loss is 7.88, training accuracy is 0.125000, time 57.050 samples/sec
+epoch 6, total_step 106060, total loss is 19.94 , inference loss is 12.07, weight deacy loss is 7.88, training accuracy is 0.281250, time 56.755 samples/sec
+epoch 6, total_step 106080, total loss is 16.84 , inference loss is 8.96, weight deacy loss is 7.87, training accuracy is 0.187500, time 56.321 samples/sec
+epoch 6, total_step 106100, total loss is 16.18 , inference loss is 8.31, weight deacy loss is 7.87, training accuracy is 0.343750, time 57.710 samples/sec
+epoch 6, total_step 106120, total loss is 20.13 , inference loss is 12.25, weight deacy loss is 7.87, training accuracy is 0.250000, time 56.937 samples/sec
+epoch 6, total_step 106140, total loss is 19.26 , inference loss is 11.39, weight deacy loss is 7.87, training accuracy is 0.218750, time 56.161 samples/sec
+epoch 6, total_step 106160, total loss is 16.18 , inference loss is 8.31, weight deacy loss is 7.87, training accuracy is 0.312500, time 56.751 samples/sec
+epoch 6, total_step 106180, total loss is 19.69 , inference loss is 11.82, weight deacy loss is 7.87, training accuracy is 0.156250, time 55.428 samples/sec
+epoch 6, total_step 106200, total loss is 18.43 , inference loss is 10.56, weight deacy loss is 7.87, training accuracy is 0.187500, time 56.984 samples/sec
+epoch 6, total_step 106220, total loss is 18.02 , inference loss is 10.15, weight deacy loss is 7.87, training accuracy is 0.218750, time 56.892 samples/sec
+epoch 6, total_step 106240, total loss is 19.12 , inference loss is 11.26, weight deacy loss is 7.87, training accuracy is 0.125000, time 56.687 samples/sec
+epoch 6, total_step 106260, total loss is 17.52 , inference loss is 9.65, weight deacy loss is 7.87, training accuracy is 0.281250, time 56.795 samples/sec
+epoch 6, total_step 106280, total loss is 18.08 , inference loss is 10.21, weight deacy loss is 7.87, training accuracy is 0.250000, time 56.892 samples/sec
+epoch 6, total_step 106300, total loss is 18.82 , inference loss is 10.96, weight deacy loss is 7.87, training accuracy is 0.156250, time 56.318 samples/sec
+epoch 6, total_step 106320, total loss is 21.49 , inference loss is 13.62, weight deacy loss is 7.87, training accuracy is 0.218750, time 57.190 samples/sec
+epoch 6, total_step 106340, total loss is 19.28 , inference loss is 11.42, weight deacy loss is 7.87, training accuracy is 0.281250, time 56.376 samples/sec
+epoch 6, total_step 106360, total loss is 21.84 , inference loss is 13.97, weight deacy loss is 7.87, training accuracy is 0.156250, time 57.047 samples/sec
+epoch 6, total_step 106380, total loss is 20.07 , inference loss is 12.21, weight deacy loss is 7.86, training accuracy is 0.187500, time 55.922 samples/sec
+epoch 6, total_step 106400, total loss is 20.15 , inference loss is 12.28, weight deacy loss is 7.86, training accuracy is 0.312500, time 55.791 samples/sec
+epoch 6, total_step 106420, total loss is 20.73 , inference loss is 12.87, weight deacy loss is 7.86, training accuracy is 0.125000, time 56.007 samples/sec
+epoch 6, total_step 106440, total loss is 17.04 , inference loss is 9.18, weight deacy loss is 7.86, training accuracy is 0.250000, time 56.513 samples/sec
+epoch 6, total_step 106460, total loss is 19.62 , inference loss is 11.76, weight deacy loss is 7.86, training accuracy is 0.312500, time 56.294 samples/sec
+epoch 6, total_step 106480, total loss is 18.29 , inference loss is 10.42, weight deacy loss is 7.86, training accuracy is 0.312500, time 56.047 samples/sec
+epoch 6, total_step 106500, total loss is 20.83 , inference loss is 12.97, weight deacy loss is 7.86, training accuracy is 0.250000, time 56.402 samples/sec2022-12-01 21:02:29.972881: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:111] Filling up shuffle buffer (this may take a while): 7423 of 10000
+2022-12-01 21:02:33.237230: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:162] Shuffle buffer filled.
+
+epoch 6, total_step 106520, total loss is 15.63 , inference loss is 7.77, weight deacy loss is 7.86, training accuracy is 0.406250, time 55.917 samples/sec
+epoch 6, total_step 106540, total loss is 21.38 , inference loss is 13.52, weight deacy loss is 7.86, training accuracy is 0.250000, time 56.398 samples/sec
+epoch 6, total_step 106560, total loss is 15.54 , inference loss is 7.68, weight deacy loss is 7.86, training accuracy is 0.437500, time 57.036 samples/sec
+epoch 6, total_step 106580, total loss is 20.02 , inference loss is 12.16, weight deacy loss is 7.86, training accuracy is 0.187500, time 56.152 samples/sec
+epoch 6, total_step 106600, total loss is 21.86 , inference loss is 14.00, weight deacy loss is 7.86, training accuracy is 0.187500, time 57.092 samples/sec
+epoch 6, total_step 106620, total loss is 18.79 , inference loss is 10.93, weight deacy loss is 7.86, training accuracy is 0.375000, time 56.608 samples/sec
+epoch 6, total_step 106640, total loss is 21.34 , inference loss is 13.49, weight deacy loss is 7.86, training accuracy is 0.156250, time 55.898 samples/sec
+epoch 6, total_step 106660, total loss is 24.16 , inference loss is 16.31, weight deacy loss is 7.86, training accuracy is 0.156250, time 56.846 samples/sec
+epoch 6, total_step 106680, total loss is 20.30 , inference loss is 12.44, weight deacy loss is 7.85, training accuracy is 0.187500, time 56.757 samples/sec
+epoch 6, total_step 106700, total loss is 20.16 , inference loss is 12.30, weight deacy loss is 7.85, training accuracy is 0.125000, time 55.997 samples/sec
+epoch 6, total_step 106720, total loss is 19.42 , inference loss is 11.57, weight deacy loss is 7.85, training accuracy is 0.281250, time 56.142 samples/sec
+epoch 6, total_step 106740, total loss is 16.46 , inference loss is 8.60, weight deacy loss is 7.85, training accuracy is 0.250000, time 56.854 samples/sec
+epoch 6, total_step 106760, total loss is 21.06 , inference loss is 13.21, weight deacy loss is 7.85, training accuracy is 0.281250, time 56.283 samples/sec
+epoch 6, total_step 106780, total loss is 23.27 , inference loss is 15.42, weight deacy loss is 7.85, training accuracy is 0.281250, time 56.555 samples/sec
+epoch 6, total_step 106800, total loss is 21.86 , inference loss is 14.01, weight deacy loss is 7.85, training accuracy is 0.187500, time 56.578 samples/sec
+epoch 6, total_step 106820, total loss is 18.10 , inference loss is 10.25, weight deacy loss is 7.85, training accuracy is 0.250000, time 56.544 samples/sec
+epoch 6, total_step 106840, total loss is 19.75 , inference loss is 11.90, weight deacy loss is 7.85, training accuracy is 0.125000, time 56.283 samples/sec
+epoch 6, total_step 106860, total loss is 15.52 , inference loss is 7.67, weight deacy loss is 7.85, training accuracy is 0.218750, time 56.425 samples/sec
+epoch 6, total_step 106880, total loss is 24.02 , inference loss is 16.18, weight deacy loss is 7.85, training accuracy is 0.187500, time 57.169 samples/sec
+epoch 6, total_step 106900, total loss is 16.43 , inference loss is 8.58, weight deacy loss is 7.85, training accuracy is 0.281250, time 56.516 samples/sec
+epoch 6, total_step 106920, total loss is 20.59 , inference loss is 12.74, weight deacy loss is 7.85, training accuracy is 0.281250, time 56.355 samples/sec
+epoch 6, total_step 106940, total loss is 21.38 , inference loss is 13.54, weight deacy loss is 7.85, training accuracy is 0.218750, time 56.520 samples/sec
+epoch 6, total_step 106960, total loss is 16.39 , inference loss is 8.54, weight deacy loss is 7.85, training accuracy is 0.281250, time 56.428 samples/sec
+epoch 6, total_step 106980, total loss is 19.12 , inference loss is 11.28, weight deacy loss is 7.84, training accuracy is 0.312500, time 56.254 samples/sec
+epoch 6, total_step 107000, total loss is 15.93 , inference loss is 8.08, weight deacy loss is 7.84, training accuracy is 0.156250, time 55.905 samples/sec
+epoch 6, total_step 107020, total loss is 17.86 , inference loss is 10.02, weight deacy loss is 7.84, training accuracy is 0.187500, time 56.737 samples/sec
+epoch 6, total_step 107040, total loss is 22.72 , inference loss is 14.88, weight deacy loss is 7.84, training accuracy is 0.125000, time 56.621 samples/sec
+epoch 6, total_step 107060, total loss is 19.53 , inference loss is 11.69, weight deacy loss is 7.84, training accuracy is 0.250000, time 57.170 samples/sec
+epoch 6, total_step 107080, total loss is 18.68 , inference loss is 10.84, weight deacy loss is 7.84, training accuracy is 0.156250, time 56.935 samples/sec
+epoch 6, total_step 107100, total loss is 16.79 , inference loss is 8.94, weight deacy loss is 7.84, training accuracy is 0.343750, time 56.081 samples/sec
+epoch 6, total_step 107120, total loss is 21.14 , inference loss is 13.30, weight deacy loss is 7.84, training accuracy is 0.156250, time 56.423 samples/sec
+epoch 6, total_step 107140, total loss is 17.74 , inference loss is 9.90, weight deacy loss is 7.84, training accuracy is 0.375000, time 56.501 samples/sec
+epoch 6, total_step 107160, total loss is 17.06 , inference loss is 9.23, weight deacy loss is 7.84, training accuracy is 0.250000, time 56.808 samples/sec
+epoch 6, total_step 107180, total loss is 20.00 , inference loss is 12.16, weight deacy loss is 7.84, training accuracy is 0.312500, time 54.953 samples/sec
+epoch 6, total_step 107200, total loss is 17.82 , inference loss is 9.98, weight deacy loss is 7.84, training accuracy is 0.312500, time 56.829 samples/sec
+epoch 6, total_step 107220, total loss is 20.31 , inference loss is 12.47, weight deacy loss is 7.84, training accuracy is 0.281250, time 56.608 samples/sec
+epoch 6, total_step 107240, total loss is 20.06 , inference loss is 12.22, weight deacy loss is 7.84, training accuracy is 0.281250, time 56.479 samples/sec
+epoch 6, total_step 107260, total loss is 18.50 , inference loss is 10.66, weight deacy loss is 7.84, training accuracy is 0.312500, time 56.989 samples/sec
+epoch 6, total_step 107280, total loss is 17.66 , inference loss is 9.83, weight deacy loss is 7.83, training accuracy is 0.312500, time 52.644 samples/sec
+epoch 6, total_step 107300, total loss is 15.93 , inference loss is 8.09, weight deacy loss is 7.83, training accuracy is 0.406250, time 56.590 samples/sec
+epoch 6, total_step 107320, total loss is 17.10 , inference loss is 9.26, weight deacy loss is 7.83, training accuracy is 0.250000, time 57.139 samples/sec
+
+
+
+
+End of epoch 6
+epoch 7, total_step 107340, total loss is 18.51 , inference loss is 10.68, weight deacy loss is 7.83, training accuracy is 0.250000, time 56.314 samples/sec
+epoch 7, total_step 107360, total loss is 13.86 , inference loss is 6.03, weight deacy loss is 7.83, training accuracy is 0.656250, time 56.406 samples/sec
+epoch 7, total_step 107380, total loss is 17.18 , inference loss is 9.34, weight deacy loss is 7.83, training accuracy is 0.406250, time 56.988 samples/sec
+epoch 7, total_step 107400, total loss is 15.32 , inference loss is 7.48, weight deacy loss is 7.83, training accuracy is 0.562500, time 56.166 samples/sec
+epoch 7, total_step 107420, total loss is 15.15 , inference loss is 7.32, weight deacy loss is 7.83, training accuracy is 0.562500, time 56.489 samples/sec
+epoch 7, total_step 107440, total loss is 13.29 , inference loss is 5.46, weight deacy loss is 7.83, training accuracy is 0.562500, time 56.188 samples/sec
+epoch 7, total_step 107460, total loss is 14.67 , inference loss is 6.84, weight deacy loss is 7.83, training accuracy is 0.625000, time 55.926 samples/sec
+epoch 7, total_step 107480, total loss is 12.71 , inference loss is 4.89, weight deacy loss is 7.83, training accuracy is 0.593750, time 55.902 samples/sec
+epoch 7, total_step 107500, total loss is 14.81 , inference loss is 6.98, weight deacy loss is 7.83, training accuracy is 0.531250, time 56.215 samples/sec
+epoch 7, total_step 107520, total loss is 14.38 , inference loss is 6.56, weight deacy loss is 7.83, training accuracy is 0.531250, time 55.989 samples/sec
+epoch 7, total_step 107540, total loss is 15.81 , inference loss is 7.98, weight deacy loss is 7.83, training accuracy is 0.468750, time 56.026 samples/sec
+epoch 7, total_step 107560, total loss is 12.66 , inference loss is 4.84, weight deacy loss is 7.82, training accuracy is 0.625000, time 56.626 samples/sec
+epoch 7, total_step 107580, total loss is 17.23 , inference loss is 9.41, weight deacy loss is 7.82, training accuracy is 0.500000, time 56.939 samples/sec
+epoch 7, total_step 107600, total loss is 11.56 , inference loss is 3.74, weight deacy loss is 7.82, training accuracy is 0.718750, time 55.898 samples/sec
+epoch 7, total_step 107620, total loss is 13.41 , inference loss is 5.59, weight deacy loss is 7.82, training accuracy is 0.593750, time 56.387 samples/sec
+epoch 7, total_step 107640, total loss is 16.68 , inference loss is 8.85, weight deacy loss is 7.82, training accuracy is 0.562500, time 56.703 samples/sec
+epoch 7, total_step 107660, total loss is 15.58 , inference loss is 7.76, weight deacy loss is 7.82, training accuracy is 0.437500, time 57.283 samples/sec
+epoch 7, total_step 107680, total loss is 11.03 , inference loss is 3.21, weight deacy loss is 7.82, training accuracy is 0.593750, time 55.890 samples/sec
+epoch 7, total_step 107700, total loss is 16.63 , inference loss is 8.81, weight deacy loss is 7.82, training accuracy is 0.375000, time 55.668 samples/sec
+epoch 7, total_step 107720, total loss is 12.38 , inference loss is 4.56, weight deacy loss is 7.82, training accuracy is 0.687500, time 54.432 samples/sec
+epoch 7, total_step 107740, total loss is 15.30 , inference loss is 7.48, weight deacy loss is 7.82, training accuracy is 0.468750, time 55.668 samples/sec
+epoch 7, total_step 107760, total loss is 13.67 , inference loss is 5.85, weight deacy loss is 7.82, training accuracy is 0.625000, time 56.111 samples/sec
+epoch 7, total_step 107780, total loss is 14.77 , inference loss is 6.95, weight deacy loss is 7.82, training accuracy is 0.562500, time 56.793 samples/sec
+epoch 7, total_step 107800, total loss is 14.51 , inference loss is 6.69, weight deacy loss is 7.82, training accuracy is 0.593750, time 56.296 samples/sec
+epoch 7, total_step 107820, total loss is 16.30 , inference loss is 8.48, weight deacy loss is 7.82, training accuracy is 0.500000, time 56.893 samples/sec
+epoch 7, total_step 107840, total loss is 21.06 , inference loss is 13.24, weight deacy loss is 7.82, training accuracy is 0.312500, time 56.406 samples/sec
+epoch 7, total_step 107860, total loss is 16.86 , inference loss is 9.04, weight deacy loss is 7.82, training accuracy is 0.531250, time 56.440 samples/sec
+epoch 7, total_step 107880, total loss is 11.51 , inference loss is 3.70, weight deacy loss is 7.82, training accuracy is 0.656250, time 56.424 samples/sec
+epoch 7, total_step 107900, total loss is 14.47 , inference loss is 6.65, weight deacy loss is 7.81, training accuracy is 0.531250, time 56.951 samples/sec
+epoch 7, total_step 107920, total loss is 20.05 , inference loss is 12.24, weight deacy loss is 7.81, training accuracy is 0.468750, time 56.842 samples/sec
+epoch 7, total_step 107940, total loss is 18.50 , inference loss is 10.69, weight deacy loss is 7.81, training accuracy is 0.500000, time 56.205 samples/sec
+epoch 7, total_step 107960, total loss is 18.08 , inference loss is 10.27, weight deacy loss is 7.81, training accuracy is 0.375000, time 56.131 samples/sec
+epoch 7, total_step 107980, total loss is 17.50 , inference loss is 9.69, weight deacy loss is 7.81, training accuracy is 0.500000, time 56.155 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.33802599999997
+best_threshold_index 112 0.9894444444444445
+best_threshold_index 112 0.9883333333333333
+best_threshold_index 112 0.9887037037037038
+best_threshold_index 112 0.9892592592592593
+best_threshold_index 112 0.9888888888888889
+best_threshold_index 112 0.9877777777777778
+best_threshold_index 112 0.9885185185185185
+best_threshold_index 112 0.9885185185185185
+best_threshold_index 112 0.9875925925925926
+best_threshold_index 112 0.9879629629629629
+[lfw][108000]XNorm: 20.099139
+[lfw][108000]Accuracy-Flip: 0.98850+-0.00519
+testing verification..
+(14000, 512)
+infer time 85.60477900000018
+best_threshold_index 111 0.9841269841269841
+best_threshold_index 111 0.9825396825396825
+best_threshold_index 107 0.9826984126984127
+best_threshold_index 111 0.9820634920634921
+best_threshold_index 107 0.9820634920634921
+best_threshold_index 111 0.9812698412698413
+best_threshold_index 107 0.983015873015873
+best_threshold_index 111 0.9823809523809524
+best_threshold_index 107 0.9817460317460317
+best_threshold_index 111 0.983015873015873
+[cfp_ff][108000]XNorm: 20.134995
+[cfp_ff][108000]Accuracy-Flip: 0.98129+-0.00674
+testing verification..
+(14000, 512)
+infer time 85.32609899999996
+best_threshold_index 133 0.8996825396825396
+best_threshold_index 133 0.9001587301587302
+best_threshold_index 133 0.8974603174603175
+best_threshold_index 133 0.9014285714285715
+best_threshold_index 133 0.9006349206349207
+best_threshold_index 133 0.895079365079365
+best_threshold_index 133 0.8971428571428571
+best_threshold_index 133 0.8961904761904762
+best_threshold_index 133 0.8985714285714286
+best_threshold_index 133 0.8993650793650794
+[cfp_fp][108000]XNorm: 20.876824
+[cfp_fp][108000]Accuracy-Flip: 0.89857+-0.01753
+testing verification..
+(12000, 512)
+infer time 72.92658100000001
+best_threshold_index 134 0.9016666666666666
+best_threshold_index 134 0.899074074074074
+best_threshold_index 134 0.9009259259259259
+best_threshold_index 134 0.9038888888888889
+best_threshold_index 134 0.9072222222222223
+best_threshold_index 134 0.9025925925925926
+best_threshold_index 134 0.9057407407407407
+best_threshold_index 134 0.9072222222222223
+best_threshold_index 134 0.9055555555555556
+best_threshold_index 134 0.9061111111111111
+[agedb_30][108000]XNorm: 21.024401
+[agedb_30][108000]Accuracy-Flip: 0.90400+-0.02412
+test accuracy is: 0.9884999999999999
+epoch 7, total_step 108000, total loss is 15.89 , inference loss is 8.08, weight deacy loss is 7.81, training accuracy is 0.437500, time 52.913 samples/sec
+epoch 7, total_step 108020, total loss is 16.40 , inference loss is 8.59, weight deacy loss is 7.81, training accuracy is 0.531250, time 56.614 samples/sec
+epoch 7, total_step 108040, total loss is 13.98 , inference loss is 6.17, weight deacy loss is 7.81, training accuracy is 0.500000, time 56.578 samples/sec
+epoch 7, total_step 108060, total loss is 14.78 , inference loss is 6.97, weight deacy loss is 7.81, training accuracy is 0.343750, time 56.954 samples/sec
+epoch 7, total_step 108080, total loss is 12.72 , inference loss is 4.91, weight deacy loss is 7.81, training accuracy is 0.656250, time 56.891 samples/sec
+epoch 7, total_step 108100, total loss is 13.78 , inference loss is 5.97, weight deacy loss is 7.81, training accuracy is 0.531250, time 56.012 samples/sec
+epoch 7, total_step 108120, total loss is 14.14 , inference loss is 6.33, weight deacy loss is 7.81, training accuracy is 0.562500, time 56.619 samples/sec
+epoch 7, total_step 108140, total loss is 13.79 , inference loss is 5.98, weight deacy loss is 7.81, training accuracy is 0.593750, time 56.413 samples/sec
+epoch 7, total_step 108160, total loss is 15.41 , inference loss is 7.60, weight deacy loss is 7.81, training accuracy is 0.562500, time 56.394 samples/sec
+epoch 7, total_step 108180, total loss is 13.17 , inference loss is 5.36, weight deacy loss is 7.81, training accuracy is 0.531250, time 56.053 samples/sec
+epoch 7, total_step 108200, total loss is 17.13 , inference loss is 9.32, weight deacy loss is 7.81, training accuracy is 0.406250, time 56.904 samples/sec
+epoch 7, total_step 108220, total loss is 17.96 , inference loss is 10.16, weight deacy loss is 7.81, training accuracy is 0.375000, time 56.245 samples/sec
+epoch 7, total_step 108240, total loss is 16.61 , inference loss is 8.80, weight deacy loss is 7.81, training accuracy is 0.531250, time 55.542 samples/sec
+epoch 7, total_step 108260, total loss is 16.78 , inference loss is 8.97, weight deacy loss is 7.81, training accuracy is 0.281250, time 56.623 samples/sec
+epoch 7, total_step 108280, total loss is 16.31 , inference loss is 8.50, weight deacy loss is 7.81, training accuracy is 0.250000, time 57.210 samples/sec
+epoch 7, total_step 108300, total loss is 22.68 , inference loss is 14.87, weight deacy loss is 7.81, training accuracy is 0.312500, time 56.678 samples/sec
+epoch 7, total_step 108320, total loss is 18.72 , inference loss is 10.92, weight deacy loss is 7.80, training accuracy is 0.343750, time 56.074 samples/sec
+epoch 7, total_step 108340, total loss is 17.80 , inference loss is 10.00, weight deacy loss is 7.80, training accuracy is 0.281250, time 57.567 samples/sec
+epoch 7, total_step 108360, total loss is 15.35 , inference loss is 7.54, weight deacy loss is 7.80, training accuracy is 0.531250, time 55.996 samples/sec
+epoch 7, total_step 108380, total loss is 14.96 , inference loss is 7.16, weight deacy loss is 7.80, training accuracy is 0.500000, time 56.152 samples/sec
+epoch 7, total_step 108400, total loss is 19.34 , inference loss is 11.53, weight deacy loss is 7.80, training accuracy is 0.375000, time 56.700 samples/sec
+epoch 7, total_step 108420, total loss is 14.67 , inference loss is 6.87, weight deacy loss is 7.80, training accuracy is 0.406250, time 57.386 samples/sec
+epoch 7, total_step 108440, total loss is 15.50 , inference loss is 7.70, weight deacy loss is 7.80, training accuracy is 0.343750, time 56.150 samples/sec
+epoch 7, total_step 108460, total loss is 13.62 , inference loss is 5.82, weight deacy loss is 7.80, training accuracy is 0.437500, time 56.228 samples/sec
+epoch 7, total_step 108480, total loss is 11.31 , inference loss is 3.51, weight deacy loss is 7.80, training accuracy is 0.625000, time 56.713 samples/sec
+epoch 7, total_step 108500, total loss is 15.62 , inference loss is 7.82, weight deacy loss is 7.80, training accuracy is 0.375000, time 56.426 samples/sec
+epoch 7, total_step 108520, total loss is 14.47 , inference loss is 6.67, weight deacy loss is 7.80, training accuracy is 0.500000, time 55.980 samples/sec
+epoch 7, total_step 108540, total loss is 18.96 , inference loss is 11.16, weight deacy loss is 7.80, training accuracy is 0.343750, time 56.660 samples/sec
+epoch 7, total_step 108560, total loss is 15.67 , inference loss is 7.87, weight deacy loss is 7.80, training accuracy is 0.500000, time 57.162 samples/sec
+epoch 7, total_step 108580, total loss is 19.94 , inference loss is 12.14, weight deacy loss is 7.80, training accuracy is 0.406250, time 57.065 samples/sec
+epoch 7, total_step 108600, total loss is 18.06 , inference loss is 10.26, weight deacy loss is 7.80, training accuracy is 0.468750, time 56.075 samples/sec
+epoch 7, total_step 108620, total loss is 13.94 , inference loss is 6.14, weight deacy loss is 7.80, training accuracy is 0.531250, time 56.183 samples/sec
+epoch 7, total_step 108640, total loss is 22.53 , inference loss is 14.73, weight deacy loss is 7.80, training accuracy is 0.437500, time 55.953 samples/sec
+epoch 7, total_step 108660, total loss is 19.80 , inference loss is 12.00, weight deacy loss is 7.80, training accuracy is 0.406250, time 56.556 samples/sec
+epoch 7, total_step 108680, total loss is 18.28 , inference loss is 10.49, weight deacy loss is 7.80, training accuracy is 0.375000, time 56.214 samples/sec
+epoch 7, total_step 108700, total loss is 14.42 , inference loss is 6.62, weight deacy loss is 7.80, training accuracy is 0.437500, time 56.106 samples/sec
+epoch 7, total_step 108720, total loss is 18.86 , inference loss is 11.07, weight deacy loss is 7.80, training accuracy is 0.406250, time 56.028 samples/sec
+epoch 7, total_step 108740, total loss is 15.68 , inference loss is 7.89, weight deacy loss is 7.79, training accuracy is 0.531250, time 56.837 samples/sec
+epoch 7, total_step 108760, total loss is 18.82 , inference loss is 11.03, weight deacy loss is 7.79, training accuracy is 0.375000, time 56.363 samples/sec
+epoch 7, total_step 108780, total loss is 15.71 , inference loss is 7.92, weight deacy loss is 7.79, training accuracy is 0.656250, time 56.589 samples/sec
+epoch 7, total_step 108800, total loss is 20.32 , inference loss is 12.53, weight deacy loss is 7.79, training accuracy is 0.250000, time 56.736 samples/sec
+epoch 7, total_step 108820, total loss is 17.98 , inference loss is 10.19, weight deacy loss is 7.79, training accuracy is 0.437500, time 56.777 samples/sec
+epoch 7, total_step 108840, total loss is 19.02 , inference loss is 11.23, weight deacy loss is 7.79, training accuracy is 0.312500, time 56.908 samples/sec
+epoch 7, total_step 108860, total loss is 16.84 , inference loss is 9.05, weight deacy loss is 7.79, training accuracy is 0.468750, time 56.411 samples/sec
+epoch 7, total_step 108880, total loss is 14.88 , inference loss is 7.09, weight deacy loss is 7.79, training accuracy is 0.312500, time 56.292 samples/sec
+epoch 7, total_step 108900, total loss is 19.64 , inference loss is 11.85, weight deacy loss is 7.79, training accuracy is 0.312500, time 55.831 samples/sec
+epoch 7, total_step 108920, total loss is 20.44 , inference loss is 12.65, weight deacy loss is 7.79, training accuracy is 0.375000, time 56.487 samples/sec
+epoch 7, total_step 108940, total loss is 17.01 , inference loss is 9.22, weight deacy loss is 7.79, training accuracy is 0.500000, time 58.092 samples/sec
+epoch 7, total_step 108960, total loss is 13.59 , inference loss is 5.80, weight deacy loss is 7.79, training accuracy is 0.562500, time 56.835 samples/sec
+epoch 7, total_step 108980, total loss is 15.84 , inference loss is 8.05, weight deacy loss is 7.79, training accuracy is 0.406250, time 56.051 samples/sec
+epoch 7, total_step 109000, total loss is 13.42 , inference loss is 5.63, weight deacy loss is 7.79, training accuracy is 0.468750, time 56.494 samples/sec
+epoch 7, total_step 109020, total loss is 15.94 , inference loss is 8.15, weight deacy loss is 7.79, training accuracy is 0.281250, time 57.006 samples/sec
+epoch 7, total_step 109040, total loss is 19.76 , inference loss is 11.97, weight deacy loss is 7.79, training accuracy is 0.218750, time 56.669 samples/sec
+epoch 7, total_step 109060, total loss is 17.18 , inference loss is 9.39, weight deacy loss is 7.79, training accuracy is 0.468750, time 56.518 samples/sec
+epoch 7, total_step 109080, total loss is 17.27 , inference loss is 9.49, weight deacy loss is 7.79, training accuracy is 0.281250, time 56.836 samples/sec
+epoch 7, total_step 109100, total loss is 18.77 , inference loss is 10.99, weight deacy loss is 7.79, training accuracy is 0.406250, time 56.194 samples/sec
+epoch 7, total_step 109120, total loss is 18.05 , inference loss is 10.26, weight deacy loss is 7.78, training accuracy is 0.437500, time 55.993 samples/sec
+epoch 7, total_step 109140, total loss is 15.72 , inference loss is 7.94, weight deacy loss is 7.78, training accuracy is 0.375000, time 56.088 samples/sec
+epoch 7, total_step 109160, total loss is 16.85 , inference loss is 9.06, weight deacy loss is 7.78, training accuracy is 0.406250, time 56.011 samples/sec
+epoch 7, total_step 109180, total loss is 17.09 , inference loss is 9.31, weight deacy loss is 7.78, training accuracy is 0.437500, time 56.317 samples/sec
+epoch 7, total_step 109200, total loss is 18.29 , inference loss is 10.50, weight deacy loss is 7.78, training accuracy is 0.312500, time 56.141 samples/sec
+epoch 7, total_step 109220, total loss is 14.86 , inference loss is 7.07, weight deacy loss is 7.78, training accuracy is 0.625000, time 56.401 samples/sec
+epoch 7, total_step 109240, total loss is 17.39 , inference loss is 9.61, weight deacy loss is 7.78, training accuracy is 0.375000, time 56.727 samples/sec
+epoch 7, total_step 109260, total loss is 14.58 , inference loss is 6.80, weight deacy loss is 7.78, training accuracy is 0.343750, time 55.823 samples/sec
+epoch 7, total_step 109280, total loss is 15.89 , inference loss is 8.11, weight deacy loss is 7.78, training accuracy is 0.437500, time 56.231 samples/sec
+epoch 7, total_step 109300, total loss is 12.25 , inference loss is 4.47, weight deacy loss is 7.78, training accuracy is 0.562500, time 56.637 samples/sec
+epoch 7, total_step 109320, total loss is 13.01 , inference loss is 5.23, weight deacy loss is 7.78, training accuracy is 0.562500, time 57.098 samples/sec
+epoch 7, total_step 109340, total loss is 16.60 , inference loss is 8.83, weight deacy loss is 7.78, training accuracy is 0.437500, time 56.399 samples/sec
+epoch 7, total_step 109360, total loss is 20.61 , inference loss is 12.83, weight deacy loss is 7.78, training accuracy is 0.343750, time 56.406 samples/sec
+epoch 7, total_step 109380, total loss is 12.38 , inference loss is 4.61, weight deacy loss is 7.78, training accuracy is 0.531250, time 57.542 samples/sec
+epoch 7, total_step 109400, total loss is 11.92 , inference loss is 4.15, weight deacy loss is 7.78, training accuracy is 0.593750, time 57.146 samples/sec
+epoch 7, total_step 109420, total loss is 16.77 , inference loss is 8.99, weight deacy loss is 7.78, training accuracy is 0.500000, time 56.119 samples/sec
+epoch 7, total_step 109440, total loss is 16.11 , inference loss is 8.33, weight deacy loss is 7.77, training accuracy is 0.531250, time 56.300 samples/sec
+epoch 7, total_step 109460, total loss is 17.09 , inference loss is 9.32, weight deacy loss is 7.77, training accuracy is 0.468750, time 56.545 samples/sec
+epoch 7, total_step 109480, total loss is 15.81 , inference loss is 8.04, weight deacy loss is 7.77, training accuracy is 0.468750, time 55.296 samples/sec
+epoch 7, total_step 109500, total loss is 14.45 , inference loss is 6.67, weight deacy loss is 7.77, training accuracy is 0.531250, time 56.470 samples/sec
+epoch 7, total_step 109520, total loss is 16.42 , inference loss is 8.65, weight deacy loss is 7.77, training accuracy is 0.468750, time 56.349 samples/sec
+epoch 7, total_step 109540, total loss is 16.42 , inference loss is 8.65, weight deacy loss is 7.77, training accuracy is 0.406250, time 55.564 samples/sec
+epoch 7, total_step 109560, total loss is 14.05 , inference loss is 6.28, weight deacy loss is 7.77, training accuracy is 0.562500, time 56.906 samples/sec
+epoch 7, total_step 109580, total loss is 12.82 , inference loss is 5.05, weight deacy loss is 7.77, training accuracy is 0.562500, time 56.406 samples/sec
+epoch 7, total_step 109600, total loss is 15.13 , inference loss is 7.36, weight deacy loss is 7.77, training accuracy is 0.531250, time 56.980 samples/sec
+epoch 7, total_step 109620, total loss is 16.11 , inference loss is 8.34, weight deacy loss is 7.77, training accuracy is 0.468750, time 56.623 samples/sec
+epoch 7, total_step 109640, total loss is 13.51 , inference loss is 5.74, weight deacy loss is 7.77, training accuracy is 0.468750, time 55.586 samples/sec
+epoch 7, total_step 109660, total loss is 12.41 , inference loss is 4.64, weight deacy loss is 7.77, training accuracy is 0.468750, time 56.067 samples/sec
+epoch 7, total_step 109680, total loss is 11.22 , inference loss is 3.45, weight deacy loss is 7.77, training accuracy is 0.625000, time 55.473 samples/sec
+epoch 7, total_step 109700, total loss is 16.96 , inference loss is 9.19, weight deacy loss is 7.77, training accuracy is 0.500000, time 56.436 samples/sec
+epoch 7, total_step 109720, total loss is 13.39 , inference loss is 5.62, weight deacy loss is 7.77, training accuracy is 0.531250, time 56.531 samples/sec
+epoch 7, total_step 109740, total loss is 12.84 , inference loss is 5.08, weight deacy loss is 7.77, training accuracy is 0.562500, time 56.114 samples/sec
+epoch 7, total_step 109760, total loss is 13.19 , inference loss is 5.42, weight deacy loss is 7.76, training accuracy is 0.406250, time 56.001 samples/sec
+epoch 7, total_step 109780, total loss is 16.82 , inference loss is 9.06, weight deacy loss is 7.76, training accuracy is 0.406250, time 56.757 samples/sec
+epoch 7, total_step 109800, total loss is 16.41 , inference loss is 8.65, weight deacy loss is 7.76, training accuracy is 0.437500, time 56.598 samples/sec
+epoch 7, total_step 109820, total loss is 12.03 , inference loss is 4.26, weight deacy loss is 7.76, training accuracy is 0.500000, time 56.898 samples/sec
+epoch 7, total_step 109840, total loss is 15.85 , inference loss is 8.09, weight deacy loss is 7.76, training accuracy is 0.468750, time 57.005 samples/sec
+epoch 7, total_step 109860, total loss is 18.36 , inference loss is 10.60, weight deacy loss is 7.76, training accuracy is 0.406250, time 56.813 samples/sec
+epoch 7, total_step 109880, total loss is 14.00 , inference loss is 6.24, weight deacy loss is 7.76, training accuracy is 0.437500, time 56.536 samples/sec
+epoch 7, total_step 109900, total loss is 14.68 , inference loss is 6.92, weight deacy loss is 7.76, training accuracy is 0.468750, time 56.462 samples/sec
+epoch 7, total_step 109920, total loss is 16.20 , inference loss is 8.44, weight deacy loss is 7.76, training accuracy is 0.406250, time 56.868 samples/sec
+epoch 7, total_step 109940, total loss is 14.42 , inference loss is 6.66, weight deacy loss is 7.76, training accuracy is 0.406250, time 56.491 samples/sec
+epoch 7, total_step 109960, total loss is 17.06 , inference loss is 9.30, weight deacy loss is 7.76, training accuracy is 0.468750, time 57.925 samples/sec
+epoch 7, total_step 109980, total loss is 11.05 , inference loss is 3.29, weight deacy loss is 7.76, training accuracy is 0.593750, time 56.796 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.21520599999991
+best_threshold_index 118 0.9879629629629629
+best_threshold_index 118 0.9877777777777778
+best_threshold_index 118 0.9872222222222222
+best_threshold_index 119 0.9883333333333333
+best_threshold_index 118 0.9879629629629629
+best_threshold_index 118 0.9868518518518519
+best_threshold_index 112 0.9874074074074074
+best_threshold_index 118 0.9868518518518519
+best_threshold_index 118 0.9866666666666667
+best_threshold_index 118 0.9864814814814815
+[lfw][110000]XNorm: 23.682693
+[lfw][110000]Accuracy-Flip: 0.98667+-0.00582
+testing verification..
+(14000, 512)
+infer time 85.468106
+best_threshold_index 115 0.9850793650793651
+best_threshold_index 120 0.9844444444444445
+best_threshold_index 116 0.9844444444444445
+best_threshold_index 120 0.9839682539682539
+best_threshold_index 120 0.9846031746031746
+best_threshold_index 116 0.9836507936507937
+best_threshold_index 120 0.9846031746031746
+best_threshold_index 120 0.9846031746031746
+best_threshold_index 120 0.9852380952380952
+best_threshold_index 120 0.9855555555555555
+[cfp_ff][110000]XNorm: 22.829026
+[cfp_ff][110000]Accuracy-Flip: 0.98314+-0.00486
+testing verification..
+(14000, 512)
+infer time 85.31471700000006
+best_threshold_index 137 0.9130158730158731
+best_threshold_index 136 0.91
+best_threshold_index 138 0.9093650793650794
+best_threshold_index 136 0.9117460317460317
+best_threshold_index 136 0.9126984126984127
+best_threshold_index 136 0.9079365079365079
+best_threshold_index 136 0.9077777777777778
+best_threshold_index 136 0.909047619047619
+best_threshold_index 137 0.9104761904761904
+best_threshold_index 136 0.9098412698412699
+[cfp_fp][110000]XNorm: 23.574401
+[cfp_fp][110000]Accuracy-Flip: 0.90900+-0.01565
+testing verification..
+(12000, 512)
+infer time 73.07699199999998
+best_threshold_index 134 0.8955555555555555
+best_threshold_index 132 0.8953703703703704
+best_threshold_index 132 0.8977777777777778
+best_threshold_index 132 0.8988888888888888
+best_threshold_index 132 0.9024074074074074
+best_threshold_index 132 0.8983333333333333
+best_threshold_index 132 0.9001851851851852
+best_threshold_index 132 0.9042592592592592
+best_threshold_index 134 0.9031481481481481
+best_threshold_index 134 0.9005555555555556
+[agedb_30][110000]XNorm: 24.020487
+[agedb_30][110000]Accuracy-Flip: 0.89817+-0.02594
+test accuracy is: 0.9866666666666666
+epoch 7, total_step 110000, total loss is 13.32 , inference loss is 5.56, weight deacy loss is 7.76, training accuracy is 0.656250, time 54.484 samples/sec
+epoch 7, total_step 110020, total loss is 16.03 , inference loss is 8.27, weight deacy loss is 7.76, training accuracy is 0.531250, time 56.737 samples/sec
+epoch 7, total_step 110040, total loss is 13.44 , inference loss is 5.68, weight deacy loss is 7.76, training accuracy is 0.531250, time 56.588 samples/sec
+epoch 7, total_step 110060, total loss is 15.22 , inference loss is 7.46, weight deacy loss is 7.75, training accuracy is 0.406250, time 56.578 samples/sec
+epoch 7, total_step 110080, total loss is 16.52 , inference loss is 8.76, weight deacy loss is 7.75, training accuracy is 0.468750, time 56.598 samples/sec
+epoch 7, total_step 110100, total loss is 13.92 , inference loss is 6.16, weight deacy loss is 7.75, training accuracy is 0.531250, time 57.267 samples/sec
+epoch 7, total_step 110120, total loss is 15.37 , inference loss is 7.62, weight deacy loss is 7.75, training accuracy is 0.406250, time 56.218 samples/sec
+epoch 7, total_step 110140, total loss is 15.48 , inference loss is 7.73, weight deacy loss is 7.75, training accuracy is 0.468750, time 56.221 samples/sec
+epoch 7, total_step 110160, total loss is 15.20 , inference loss is 7.45, weight deacy loss is 7.75, training accuracy is 0.343750, time 56.941 samples/sec
+epoch 7, total_step 110180, total loss is 15.26 , inference loss is 7.51, weight deacy loss is 7.75, training accuracy is 0.468750, time 56.507 samples/sec
+epoch 7, total_step 110200, total loss is 13.72 , inference loss is 5.97, weight deacy loss is 7.75, training accuracy is 0.593750, time 56.699 samples/sec
+epoch 7, total_step 110220, total loss is 12.73 , inference loss is 4.98, weight deacy loss is 7.75, training accuracy is 0.562500, time 56.465 samples/sec
+epoch 7, total_step 110240, total loss is 13.95 , inference loss is 6.20, weight deacy loss is 7.75, training accuracy is 0.406250, time 56.437 samples/sec
+epoch 7, total_step 110260, total loss is 15.65 , inference loss is 7.90, weight deacy loss is 7.75, training accuracy is 0.625000, time 56.977 samples/sec
+epoch 7, total_step 110280, total loss is 14.58 , inference loss is 6.83, weight deacy loss is 7.75, training accuracy is 0.500000, time 56.302 samples/sec
+epoch 7, total_step 110300, total loss is 13.30 , inference loss is 5.55, weight deacy loss is 7.75, training accuracy is 0.531250, time 56.624 samples/sec
+epoch 7, total_step 110320, total loss is 16.13 , inference loss is 8.38, weight deacy loss is 7.75, training accuracy is 0.406250, time 56.038 samples/sec
+epoch 7, total_step 110340, total loss is 14.76 , inference loss is 7.01, weight deacy loss is 7.75, training accuracy is 0.437500, time 56.184 samples/sec
+epoch 7, total_step 110360, total loss is 12.93 , inference loss is 5.18, weight deacy loss is 7.75, training accuracy is 0.593750, time 56.357 samples/sec
+epoch 7, total_step 110380, total loss is 13.89 , inference loss is 6.15, weight deacy loss is 7.74, training accuracy is 0.468750, time 56.691 samples/sec
+epoch 7, total_step 110400, total loss is 16.01 , inference loss is 8.27, weight deacy loss is 7.74, training accuracy is 0.531250, time 56.667 samples/sec
+epoch 7, total_step 110420, total loss is 15.38 , inference loss is 7.64, weight deacy loss is 7.74, training accuracy is 0.593750, time 56.821 samples/sec
+epoch 7, total_step 110440, total loss is 11.82 , inference loss is 4.08, weight deacy loss is 7.74, training accuracy is 0.500000, time 56.904 samples/sec
+epoch 7, total_step 110460, total loss is 18.63 , inference loss is 10.89, weight deacy loss is 7.74, training accuracy is 0.406250, time 56.522 samples/sec
+epoch 7, total_step 110480, total loss is 17.84 , inference loss is 10.10, weight deacy loss is 7.74, training accuracy is 0.437500, time 56.293 samples/sec
+epoch 7, total_step 110500, total loss is 14.57 , inference loss is 6.83, weight deacy loss is 7.74, training accuracy is 0.437500, time 56.858 samples/sec
+epoch 7, total_step 110520, total loss is 16.97 , inference loss is 9.23, weight deacy loss is 7.74, training accuracy is 0.312500, time 56.730 samples/sec
+epoch 7, total_step 110540, total loss is 15.19 , inference loss is 7.45, weight deacy loss is 7.74, training accuracy is 0.406250, time 56.235 samples/sec
+epoch 7, total_step 110560, total loss is 14.11 , inference loss is 6.37, weight deacy loss is 7.74, training accuracy is 0.500000, time 56.478 samples/sec
+epoch 7, total_step 110580, total loss is 14.80 , inference loss is 7.06, weight deacy loss is 7.74, training accuracy is 0.437500, time 57.155 samples/sec
+epoch 7, total_step 110600, total loss is 18.40 , inference loss is 10.67, weight deacy loss is 7.74, training accuracy is 0.375000, time 56.837 samples/sec
+epoch 7, total_step 110620, total loss is 16.49 , inference loss is 8.75, weight deacy loss is 7.74, training accuracy is 0.500000, time 56.895 samples/sec
+epoch 7, total_step 110640, total loss is 15.01 , inference loss is 7.27, weight deacy loss is 7.74, training accuracy is 0.375000, time 57.077 samples/sec
+epoch 7, total_step 110660, total loss is 17.09 , inference loss is 9.35, weight deacy loss is 7.74, training accuracy is 0.500000, time 56.110 samples/sec
+epoch 7, total_step 110680, total loss is 16.45 , inference loss is 8.72, weight deacy loss is 7.74, training accuracy is 0.375000, time 56.904 samples/sec
+epoch 7, total_step 110700, total loss is 13.44 , inference loss is 5.71, weight deacy loss is 7.74, training accuracy is 0.375000, time 56.142 samples/sec
+epoch 7, total_step 110720, total loss is 16.44 , inference loss is 8.70, weight deacy loss is 7.74, training accuracy is 0.375000, time 55.569 samples/sec
+epoch 7, total_step 110740, total loss is 16.72 , inference loss is 8.99, weight deacy loss is 7.73, training accuracy is 0.500000, time 55.643 samples/sec
+epoch 7, total_step 110760, total loss is 19.38 , inference loss is 11.65, weight deacy loss is 7.73, training accuracy is 0.437500, time 55.841 samples/sec
+epoch 7, total_step 110780, total loss is 13.86 , inference loss is 6.12, weight deacy loss is 7.73, training accuracy is 0.531250, time 57.003 samples/sec
+epoch 7, total_step 110800, total loss is 16.82 , inference loss is 9.09, weight deacy loss is 7.73, training accuracy is 0.500000, time 56.686 samples/sec
+epoch 7, total_step 110820, total loss is 16.47 , inference loss is 8.74, weight deacy loss is 7.73, training accuracy is 0.406250, time 57.091 samples/sec
+epoch 7, total_step 110840, total loss is 16.83 , inference loss is 9.09, weight deacy loss is 7.73, training accuracy is 0.250000, time 56.773 samples/sec
+epoch 7, total_step 110860, total loss is 15.11 , inference loss is 7.38, weight deacy loss is 7.73, training accuracy is 0.375000, time 57.027 samples/sec
+epoch 7, total_step 110880, total loss is 16.84 , inference loss is 9.11, weight deacy loss is 7.73, training accuracy is 0.375000, time 56.168 samples/sec
+epoch 7, total_step 110900, total loss is 19.01 , inference loss is 11.28, weight deacy loss is 7.73, training accuracy is 0.500000, time 55.587 samples/sec
+epoch 7, total_step 110920, total loss is 13.95 , inference loss is 6.22, weight deacy loss is 7.73, training accuracy is 0.406250, time 56.408 samples/sec
+epoch 7, total_step 110940, total loss is 15.96 , inference loss is 8.23, weight deacy loss is 7.73, training accuracy is 0.437500, time 56.592 samples/sec
+epoch 7, total_step 110960, total loss is 14.19 , inference loss is 6.46, weight deacy loss is 7.73, training accuracy is 0.343750, time 56.175 samples/sec
+epoch 7, total_step 110980, total loss is 18.34 , inference loss is 10.61, weight deacy loss is 7.73, training accuracy is 0.343750, time 56.014 samples/sec
+epoch 7, total_step 111000, total loss is 14.41 , inference loss is 6.68, weight deacy loss is 7.73, training accuracy is 0.406250, time 56.357 samples/sec
+epoch 7, total_step 111020, total loss is 13.15 , inference loss is 5.43, weight deacy loss is 7.73, training accuracy is 0.468750, time 56.164 samples/sec
+epoch 7, total_step 111040, total loss is 18.93 , inference loss is 11.21, weight deacy loss is 7.73, training accuracy is 0.375000, time 56.866 samples/sec
+epoch 7, total_step 111060, total loss is 16.51 , inference loss is 8.79, weight deacy loss is 7.73, training accuracy is 0.375000, time 56.413 samples/sec
+epoch 7, total_step 111080, total loss is 21.50 , inference loss is 13.78, weight deacy loss is 7.72, training accuracy is 0.218750, time 56.594 samples/sec
+epoch 7, total_step 111100, total loss is 20.17 , inference loss is 12.44, weight deacy loss is 7.72, training accuracy is 0.250000, time 56.696 samples/sec
+epoch 7, total_step 111120, total loss is 15.32 , inference loss is 7.60, weight deacy loss is 7.72, training accuracy is 0.406250, time 56.794 samples/sec
+epoch 7, total_step 111140, total loss is 20.50 , inference loss is 12.78, weight deacy loss is 7.72, training accuracy is 0.250000, time 55.933 samples/sec
+epoch 7, total_step 111160, total loss is 15.57 , inference loss is 7.84, weight deacy loss is 7.72, training accuracy is 0.375000, time 57.348 samples/sec
+epoch 7, total_step 111180, total loss is 16.83 , inference loss is 9.11, weight deacy loss is 7.72, training accuracy is 0.375000, time 55.413 samples/sec
+epoch 7, total_step 111200, total loss is 18.13 , inference loss is 10.40, weight deacy loss is 7.72, training accuracy is 0.437500, time 55.865 samples/sec
+epoch 7, total_step 111220, total loss is 16.10 , inference loss is 8.38, weight deacy loss is 7.72, training accuracy is 0.468750, time 56.240 samples/sec
+epoch 7, total_step 111240, total loss is 18.82 , inference loss is 11.10, weight deacy loss is 7.72, training accuracy is 0.343750, time 56.010 samples/sec
+epoch 7, total_step 111260, total loss is 22.48 , inference loss is 14.76, weight deacy loss is 7.72, training accuracy is 0.250000, time 55.921 samples/sec
+epoch 7, total_step 111280, total loss is 17.94 , inference loss is 10.22, weight deacy loss is 7.72, training accuracy is 0.343750, time 56.703 samples/sec
+epoch 7, total_step 111300, total loss is 19.13 , inference loss is 11.42, weight deacy loss is 7.72, training accuracy is 0.281250, time 56.310 samples/sec
+epoch 7, total_step 111320, total loss is 20.66 , inference loss is 12.94, weight deacy loss is 7.72, training accuracy is 0.312500, time 56.953 samples/sec
+epoch 7, total_step 111340, total loss is 20.62 , inference loss is 12.91, weight deacy loss is 7.72, training accuracy is 0.375000, time 56.864 samples/sec
+epoch 7, total_step 111360, total loss is 18.84 , inference loss is 11.13, weight deacy loss is 7.72, training accuracy is 0.375000, time 57.536 samples/sec
+epoch 7, total_step 111380, total loss is 21.82 , inference loss is 14.11, weight deacy loss is 7.72, training accuracy is 0.156250, time 56.995 samples/sec
+epoch 7, total_step 111400, total loss is 18.49 , inference loss is 10.78, weight deacy loss is 7.72, training accuracy is 0.312500, time 56.354 samples/sec
+epoch 7, total_step 111420, total loss is 16.13 , inference loss is 8.41, weight deacy loss is 7.72, training accuracy is 0.468750, time 56.866 samples/sec
+epoch 7, total_step 111440, total loss is 15.41 , inference loss is 7.69, weight deacy loss is 7.71, training accuracy is 0.468750, time 56.055 samples/sec
+epoch 7, total_step 111460, total loss is 14.23 , inference loss is 6.51, weight deacy loss is 7.71, training accuracy is 0.500000, time 57.063 samples/sec
+epoch 7, total_step 111480, total loss is 15.65 , inference loss is 7.94, weight deacy loss is 7.71, training accuracy is 0.343750, time 56.546 samples/sec
+epoch 7, total_step 111500, total loss is 17.55 , inference loss is 9.84, weight deacy loss is 7.71, training accuracy is 0.250000, time 56.577 samples/sec
+epoch 7, total_step 111520, total loss is 14.99 , inference loss is 7.28, weight deacy loss is 7.71, training accuracy is 0.437500, time 55.270 samples/sec
+epoch 7, total_step 111540, total loss is 15.47 , inference loss is 7.76, weight deacy loss is 7.71, training accuracy is 0.437500, time 56.230 samples/sec
+epoch 7, total_step 111560, total loss is 19.49 , inference loss is 11.77, weight deacy loss is 7.71, training accuracy is 0.281250, time 56.654 samples/sec
+epoch 7, total_step 111580, total loss is 14.07 , inference loss is 6.36, weight deacy loss is 7.71, training accuracy is 0.375000, time 56.714 samples/sec
+epoch 7, total_step 111600, total loss is 21.61 , inference loss is 13.90, weight deacy loss is 7.71, training accuracy is 0.312500, time 56.311 samples/sec
+epoch 7, total_step 111620, total loss is 17.59 , inference loss is 9.88, weight deacy loss is 7.71, training accuracy is 0.312500, time 57.329 samples/sec
+epoch 7, total_step 111640, total loss is 13.40 , inference loss is 5.69, weight deacy loss is 7.71, training accuracy is 0.500000, time 56.288 samples/sec
+epoch 7, total_step 111660, total loss is 17.50 , inference loss is 9.79, weight deacy loss is 7.71, training accuracy is 0.312500, time 55.706 samples/sec
+epoch 7, total_step 111680, total loss is 15.18 , inference loss is 7.48, weight deacy loss is 7.71, training accuracy is 0.437500, time 56.955 samples/sec
+epoch 7, total_step 111700, total loss is 14.50 , inference loss is 6.80, weight deacy loss is 7.71, training accuracy is 0.437500, time 57.019 samples/sec
+epoch 7, total_step 111720, total loss is 17.28 , inference loss is 9.57, weight deacy loss is 7.71, training accuracy is 0.218750, time 57.098 samples/sec
+epoch 7, total_step 111740, total loss is 20.26 , inference loss is 12.55, weight deacy loss is 7.71, training accuracy is 0.218750, time 56.391 samples/sec
+epoch 7, total_step 111760, total loss is 14.19 , inference loss is 6.48, weight deacy loss is 7.71, training accuracy is 0.437500, time 57.638 samples/sec
+epoch 7, total_step 111780, total loss is 18.41 , inference loss is 10.71, weight deacy loss is 7.70, training accuracy is 0.375000, time 56.598 samples/sec
+epoch 7, total_step 111800, total loss is 20.35 , inference loss is 12.64, weight deacy loss is 7.70, training accuracy is 0.375000, time 57.200 samples/sec
+epoch 7, total_step 111820, total loss is 18.89 , inference loss is 11.19, weight deacy loss is 7.70, training accuracy is 0.343750, time 56.370 samples/sec
+epoch 7, total_step 111840, total loss is 16.07 , inference loss is 8.37, weight deacy loss is 7.70, training accuracy is 0.343750, time 56.801 samples/sec
+epoch 7, total_step 111860, total loss is 18.73 , inference loss is 11.03, weight deacy loss is 7.70, training accuracy is 0.437500, time 56.833 samples/sec
+epoch 7, total_step 111880, total loss is 16.80 , inference loss is 9.10, weight deacy loss is 7.70, training accuracy is 0.437500, time 56.559 samples/sec
+epoch 7, total_step 111900, total loss is 21.34 , inference loss is 13.63, weight deacy loss is 7.70, training accuracy is 0.312500, time 56.913 samples/sec
+epoch 7, total_step 111920, total loss is 19.21 , inference loss is 11.51, weight deacy loss is 7.70, training accuracy is 0.375000, time 56.757 samples/sec
+epoch 7, total_step 111940, total loss is 18.09 , inference loss is 10.39, weight deacy loss is 7.70, training accuracy is 0.281250, time 56.545 samples/sec
+epoch 7, total_step 111960, total loss is 14.14 , inference loss is 6.44, weight deacy loss is 7.70, training accuracy is 0.531250, time 56.713 samples/sec
+epoch 7, total_step 111980, total loss is 14.95 , inference loss is 7.25, weight deacy loss is 7.70, training accuracy is 0.500000, time 56.154 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.12867199999995
+best_threshold_index 117 0.9887037037037038
+best_threshold_index 117 0.987037037037037
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 117 0.9877777777777778
+best_threshold_index 118 0.9881481481481481
+best_threshold_index 117 0.987037037037037
+best_threshold_index 117 0.9885185185185185
+best_threshold_index 117 0.9868518518518519
+best_threshold_index 118 0.9868518518518519
+best_threshold_index 117 0.9868518518518519
+[lfw][112000]XNorm: 22.286510
+[lfw][112000]Accuracy-Flip: 0.98717+-0.00615
+testing verification..
+(14000, 512)
+infer time 85.48390799999989
+best_threshold_index 122 0.9877777777777778
+best_threshold_index 122 0.9863492063492063
+best_threshold_index 122 0.986031746031746
+best_threshold_index 122 0.9868253968253968
+best_threshold_index 122 0.9868253968253968
+best_threshold_index 123 0.9863492063492063
+best_threshold_index 122 0.9868253968253968
+best_threshold_index 122 0.9866666666666667
+best_threshold_index 123 0.9853968253968254
+best_threshold_index 124 0.9874603174603175
+[cfp_ff][112000]XNorm: 22.304243
+[cfp_ff][112000]Accuracy-Flip: 0.98586+-0.00580
+testing verification..
+(14000, 512)
+infer time 85.99702499999995
+best_threshold_index 142 0.912063492063492
+best_threshold_index 142 0.909047619047619
+best_threshold_index 142 0.909047619047619
+best_threshold_index 141 0.9117460317460317
+best_threshold_index 142 0.9126984126984127
+best_threshold_index 142 0.9096825396825396
+best_threshold_index 142 0.9092063492063492
+best_threshold_index 142 0.909047619047619
+best_threshold_index 140 0.9106349206349207
+best_threshold_index 142 0.9125396825396825
+[cfp_fp][112000]XNorm: 23.125312
+[cfp_fp][112000]Accuracy-Flip: 0.90971+-0.01361
+testing verification..
+(12000, 512)
+infer time 73.21428299999998
+best_threshold_index 140 0.8918518518518519
+best_threshold_index 138 0.8912962962962963
+best_threshold_index 138 0.8935185185185185
+best_threshold_index 138 0.8914814814814814
+best_threshold_index 136 0.8983333333333333
+best_threshold_index 140 0.8922222222222222
+best_threshold_index 138 0.8938888888888888
+best_threshold_index 140 0.8983333333333333
+best_threshold_index 138 0.897962962962963
+best_threshold_index 138 0.8975925925925926
+[agedb_30][112000]XNorm: 23.399000
+[agedb_30][112000]Accuracy-Flip: 0.89150+-0.02760
+test accuracy is: 0.9871666666666667
+epoch 7, total_step 112000, total loss is 14.00 , inference loss is 6.30, weight deacy loss is 7.70, training accuracy is 0.312500, time 54.136 samples/sec
+epoch 7, total_step 112020, total loss is 16.15 , inference loss is 8.46, weight deacy loss is 7.70, training accuracy is 0.375000, time 56.746 samples/sec
+epoch 7, total_step 112040, total loss is 18.40 , inference loss is 10.70, weight deacy loss is 7.70, training accuracy is 0.437500, time 56.666 samples/sec
+epoch 7, total_step 112060, total loss is 15.42 , inference loss is 7.72, weight deacy loss is 7.70, training accuracy is 0.343750, time 56.966 samples/sec
+epoch 7, total_step 112080, total loss is 18.79 , inference loss is 11.09, weight deacy loss is 7.70, training accuracy is 0.281250, time 56.607 samples/sec
+epoch 7, total_step 112100, total loss is 13.40 , inference loss is 5.70, weight deacy loss is 7.70, training accuracy is 0.531250, time 56.314 samples/sec
+epoch 7, total_step 112120, total loss is 20.20 , inference loss is 12.51, weight deacy loss is 7.69, training accuracy is 0.343750, time 56.991 samples/sec
+epoch 7, total_step 112140, total loss is 20.50 , inference loss is 12.81, weight deacy loss is 7.69, training accuracy is 0.312500, time 56.194 samples/sec
+epoch 7, total_step 112160, total loss is 19.43 , inference loss is 11.74, weight deacy loss is 7.69, training accuracy is 0.312500, time 56.510 samples/sec
+epoch 7, total_step 112180, total loss is 14.51 , inference loss is 6.82, weight deacy loss is 7.69, training accuracy is 0.562500, time 56.289 samples/sec
+epoch 7, total_step 112200, total loss is 13.03 , inference loss is 5.33, weight deacy loss is 7.69, training accuracy is 0.468750, time 56.702 samples/sec
+epoch 7, total_step 112220, total loss is 19.05 , inference loss is 11.36, weight deacy loss is 7.69, training accuracy is 0.468750, time 56.903 samples/sec
+epoch 7, total_step 112240, total loss is 17.29 , inference loss is 9.60, weight deacy loss is 7.69, training accuracy is 0.375000, time 56.938 samples/sec
+epoch 7, total_step 112260, total loss is 14.15 , inference loss is 6.46, weight deacy loss is 7.69, training accuracy is 0.437500, time 56.322 samples/sec
+epoch 7, total_step 112280, total loss is 18.10 , inference loss is 10.41, weight deacy loss is 7.69, training accuracy is 0.406250, time 56.808 samples/sec
+epoch 7, total_step 112300, total loss is 14.74 , inference loss is 7.05, weight deacy loss is 7.69, training accuracy is 0.343750, time 56.336 samples/sec
+epoch 7, total_step 112320, total loss is 15.57 , inference loss is 7.88, weight deacy loss is 7.69, training accuracy is 0.343750, time 56.237 samples/sec
+epoch 7, total_step 112340, total loss is 14.94 , inference loss is 7.25, weight deacy loss is 7.69, training accuracy is 0.500000, time 56.698 samples/sec
+epoch 7, total_step 112360, total loss is 18.74 , inference loss is 11.05, weight deacy loss is 7.69, training accuracy is 0.281250, time 56.233 samples/sec
+epoch 7, total_step 112380, total loss is 16.74 , inference loss is 9.05, weight deacy loss is 7.69, training accuracy is 0.281250, time 54.212 samples/sec
+epoch 7, total_step 112400, total loss is 18.54 , inference loss is 10.85, weight deacy loss is 7.69, training accuracy is 0.250000, time 56.477 samples/sec
+epoch 7, total_step 112420, total loss is 14.09 , inference loss is 6.41, weight deacy loss is 7.69, training accuracy is 0.500000, time 55.608 samples/sec
+epoch 7, total_step 112440, total loss is 15.06 , inference loss is 7.38, weight deacy loss is 7.69, training accuracy is 0.437500, time 56.010 samples/sec
+epoch 7, total_step 112460, total loss is 13.17 , inference loss is 5.49, weight deacy loss is 7.68, training accuracy is 0.468750, time 55.864 samples/sec
+epoch 7, total_step 112480, total loss is 19.42 , inference loss is 11.74, weight deacy loss is 7.68, training accuracy is 0.343750, time 56.390 samples/sec
+epoch 7, total_step 112500, total loss is 20.27 , inference loss is 12.58, weight deacy loss is 7.68, training accuracy is 0.250000, time 56.883 samples/sec
+epoch 7, total_step 112520, total loss is 16.19 , inference loss is 8.50, weight deacy loss is 7.68, training accuracy is 0.343750, time 55.862 samples/sec
+epoch 7, total_step 112540, total loss is 17.98 , inference loss is 10.30, weight deacy loss is 7.68, training accuracy is 0.375000, time 56.003 samples/sec
+epoch 7, total_step 112560, total loss is 15.96 , inference loss is 8.28, weight deacy loss is 7.68, training accuracy is 0.437500, time 57.337 samples/sec
+epoch 7, total_step 112580, total loss is 17.06 , inference loss is 9.37, weight deacy loss is 7.68, training accuracy is 0.531250, time 56.548 samples/sec
+epoch 7, total_step 112600, total loss is 16.82 , inference loss is 9.14, weight deacy loss is 7.68, training accuracy is 0.218750, time 56.786 samples/sec
+epoch 7, total_step 112620, total loss is 19.88 , inference loss is 12.20, weight deacy loss is 7.68, training accuracy is 0.375000, time 56.349 samples/sec
+epoch 7, total_step 112640, total loss is 18.61 , inference loss is 10.93, weight deacy loss is 7.68, training accuracy is 0.468750, time 56.611 samples/sec
+epoch 7, total_step 112660, total loss is 19.57 , inference loss is 11.89, weight deacy loss is 7.68, training accuracy is 0.375000, time 56.851 samples/sec
+epoch 7, total_step 112680, total loss is 20.80 , inference loss is 13.12, weight deacy loss is 7.68, training accuracy is 0.250000, time 57.019 samples/sec
+epoch 7, total_step 112700, total loss is 15.74 , inference loss is 8.06, weight deacy loss is 7.68, training accuracy is 0.437500, time 56.509 samples/sec
+epoch 7, total_step 112720, total loss is 18.08 , inference loss is 10.40, weight deacy loss is 7.68, training accuracy is 0.281250, time 56.909 samples/sec
+epoch 7, total_step 112740, total loss is 17.68 , inference loss is 10.00, weight deacy loss is 7.68, training accuracy is 0.437500, time 56.155 samples/sec
+epoch 7, total_step 112760, total loss is 13.95 , inference loss is 6.27, weight deacy loss is 7.68, training accuracy is 0.468750, time 56.753 samples/sec
+epoch 7, total_step 112780, total loss is 20.56 , inference loss is 12.89, weight deacy loss is 7.68, training accuracy is 0.218750, time 57.055 samples/sec
+epoch 7, total_step 112800, total loss is 14.83 , inference loss is 7.15, weight deacy loss is 7.68, training accuracy is 0.375000, time 56.931 samples/sec
+epoch 7, total_step 112820, total loss is 17.34 , inference loss is 9.66, weight deacy loss is 7.67, training accuracy is 0.375000, time 56.193 samples/sec
+epoch 7, total_step 112840, total loss is 16.76 , inference loss is 9.08, weight deacy loss is 7.67, training accuracy is 0.343750, time 55.980 samples/sec
+epoch 7, total_step 112860, total loss is 15.99 , inference loss is 8.32, weight deacy loss is 7.67, training accuracy is 0.375000, time 55.551 samples/sec
+epoch 7, total_step 112880, total loss is 18.56 , inference loss is 10.89, weight deacy loss is 7.67, training accuracy is 0.437500, time 56.778 samples/sec
+epoch 7, total_step 112900, total loss is 17.86 , inference loss is 10.18, weight deacy loss is 7.67, training accuracy is 0.531250, time 55.885 samples/sec
+epoch 7, total_step 112920, total loss is 15.13 , inference loss is 7.46, weight deacy loss is 7.67, training accuracy is 0.343750, time 56.428 samples/sec
+epoch 7, total_step 112940, total loss is 17.12 , inference loss is 9.45, weight deacy loss is 7.67, training accuracy is 0.500000, time 56.006 samples/sec
+epoch 7, total_step 112960, total loss is 15.99 , inference loss is 8.32, weight deacy loss is 7.67, training accuracy is 0.312500, time 56.026 samples/sec
+epoch 7, total_step 112980, total loss is 14.66 , inference loss is 6.99, weight deacy loss is 7.67, training accuracy is 0.500000, time 56.334 samples/sec
+epoch 7, total_step 113000, total loss is 14.71 , inference loss is 7.04, weight deacy loss is 7.67, training accuracy is 0.468750, time 57.161 samples/sec
+epoch 7, total_step 113020, total loss is 14.70 , inference loss is 7.03, weight deacy loss is 7.67, training accuracy is 0.437500, time 56.012 samples/sec
+epoch 7, total_step 113040, total loss is 13.49 , inference loss is 5.83, weight deacy loss is 7.67, training accuracy is 0.593750, time 57.032 samples/sec
+epoch 7, total_step 113060, total loss is 15.55 , inference loss is 7.89, weight deacy loss is 7.67, training accuracy is 0.343750, time 56.672 samples/sec
+epoch 7, total_step 113080, total loss is 16.99 , inference loss is 9.32, weight deacy loss is 7.67, training accuracy is 0.437500, time 54.843 samples/sec
+epoch 7, total_step 113100, total loss is 18.65 , inference loss is 10.99, weight deacy loss is 7.67, training accuracy is 0.250000, time 57.075 samples/sec
+epoch 7, total_step 113120, total loss is 17.20 , inference loss is 9.53, weight deacy loss is 7.67, training accuracy is 0.406250, time 56.092 samples/sec
+epoch 7, total_step 113140, total loss is 17.55 , inference loss is 9.88, weight deacy loss is 7.67, training accuracy is 0.406250, time 56.013 samples/sec
+epoch 7, total_step 113160, total loss is 15.95 , inference loss is 8.29, weight deacy loss is 7.66, training accuracy is 0.312500, time 56.982 samples/sec
+epoch 7, total_step 113180, total loss is 16.28 , inference loss is 8.61, weight deacy loss is 7.66, training accuracy is 0.468750, time 56.500 samples/sec
+epoch 7, total_step 113200, total loss is 16.34 , inference loss is 8.68, weight deacy loss is 7.66, training accuracy is 0.250000, time 55.626 samples/sec
+epoch 7, total_step 113220, total loss is 15.59 , inference loss is 7.93, weight deacy loss is 7.66, training accuracy is 0.343750, time 56.446 samples/sec
+epoch 7, total_step 113240, total loss is 20.04 , inference loss is 12.38, weight deacy loss is 7.66, training accuracy is 0.375000, time 56.264 samples/sec
+epoch 7, total_step 113260, total loss is 18.16 , inference loss is 10.50, weight deacy loss is 7.66, training accuracy is 0.312500, time 55.907 samples/sec
+epoch 7, total_step 113280, total loss is 19.96 , inference loss is 12.30, weight deacy loss is 7.66, training accuracy is 0.375000, time 56.919 samples/sec
+epoch 7, total_step 113300, total loss is 13.12 , inference loss is 5.46, weight deacy loss is 7.66, training accuracy is 0.625000, time 56.797 samples/sec
+epoch 7, total_step 113320, total loss is 20.61 , inference loss is 12.95, weight deacy loss is 7.66, training accuracy is 0.312500, time 56.810 samples/sec
+epoch 7, total_step 113340, total loss is 20.14 , inference loss is 12.48, weight deacy loss is 7.66, training accuracy is 0.156250, time 56.876 samples/sec
+epoch 7, total_step 113360, total loss is 15.26 , inference loss is 7.60, weight deacy loss is 7.66, training accuracy is 0.406250, time 56.975 samples/sec
+epoch 7, total_step 113380, total loss is 14.84 , inference loss is 7.18, weight deacy loss is 7.66, training accuracy is 0.500000, time 56.470 samples/sec
+epoch 7, total_step 113400, total loss is 16.79 , inference loss is 9.13, weight deacy loss is 7.66, training accuracy is 0.343750, time 56.791 samples/sec
+epoch 7, total_step 113420, total loss is 20.38 , inference loss is 12.72, weight deacy loss is 7.66, training accuracy is 0.406250, time 56.526 samples/sec
+epoch 7, total_step 113440, total loss is 17.67 , inference loss is 10.02, weight deacy loss is 7.66, training accuracy is 0.281250, time 55.997 samples/sec
+epoch 7, total_step 113460, total loss is 14.32 , inference loss is 6.66, weight deacy loss is 7.66, training accuracy is 0.468750, time 57.028 samples/sec
+epoch 7, total_step 113480, total loss is 15.76 , inference loss is 8.10, weight deacy loss is 7.66, training accuracy is 0.281250, time 56.862 samples/sec
+epoch 7, total_step 113500, total loss is 14.99 , inference loss is 7.33, weight deacy loss is 7.65, training accuracy is 0.312500, time 55.858 samples/sec
+epoch 7, total_step 113520, total loss is 13.64 , inference loss is 5.99, weight deacy loss is 7.65, training accuracy is 0.531250, time 56.421 samples/sec
+epoch 7, total_step 113540, total loss is 17.44 , inference loss is 9.79, weight deacy loss is 7.65, training accuracy is 0.500000, time 56.332 samples/sec
+epoch 7, total_step 113560, total loss is 15.28 , inference loss is 7.63, weight deacy loss is 7.65, training accuracy is 0.343750, time 56.750 samples/sec
+epoch 7, total_step 113580, total loss is 17.76 , inference loss is 10.11, weight deacy loss is 7.65, training accuracy is 0.343750, time 56.873 samples/sec
+epoch 7, total_step 113600, total loss is 15.09 , inference loss is 7.44, weight deacy loss is 7.65, training accuracy is 0.437500, time 56.542 samples/sec
+epoch 7, total_step 113620, total loss is 17.01 , inference loss is 9.36, weight deacy loss is 7.65, training accuracy is 0.375000, time 56.465 samples/sec
+epoch 7, total_step 113640, total loss is 16.35 , inference loss is 8.70, weight deacy loss is 7.65, training accuracy is 0.406250, time 56.882 samples/sec
+epoch 7, total_step 113660, total loss is 18.26 , inference loss is 10.61, weight deacy loss is 7.65, training accuracy is 0.281250, time 56.212 samples/sec
+epoch 7, total_step 113680, total loss is 16.16 , inference loss is 8.51, weight deacy loss is 7.65, training accuracy is 0.375000, time 56.480 samples/sec
+epoch 7, total_step 113700, total loss is 21.31 , inference loss is 13.66, weight deacy loss is 7.65, training accuracy is 0.343750, time 55.265 samples/sec
+epoch 7, total_step 113720, total loss is 18.40 , inference loss is 10.76, weight deacy loss is 7.65, training accuracy is 0.375000, time 57.031 samples/sec
+epoch 7, total_step 113740, total loss is 15.83 , inference loss is 8.18, weight deacy loss is 7.65, training accuracy is 0.312500, time 56.508 samples/sec
+epoch 7, total_step 113760, total loss is 15.73 , inference loss is 8.08, weight deacy loss is 7.65, training accuracy is 0.500000, time 56.852 samples/sec
+epoch 7, total_step 113780, total loss is 17.34 , inference loss is 9.69, weight deacy loss is 7.65, training accuracy is 0.375000, time 57.138 samples/sec
+epoch 7, total_step 113800, total loss is 15.47 , inference loss is 7.82, weight deacy loss is 7.65, training accuracy is 0.281250, time 56.868 samples/sec
+epoch 7, total_step 113820, total loss is 14.50 , inference loss is 6.86, weight deacy loss is 7.65, training accuracy is 0.468750, time 56.626 samples/sec
+epoch 7, total_step 113840, total loss is 12.76 , inference loss is 5.12, weight deacy loss is 7.64, training accuracy is 0.468750, time 56.498 samples/sec
+epoch 7, total_step 113860, total loss is 17.11 , inference loss is 9.47, weight deacy loss is 7.64, training accuracy is 0.437500, time 56.932 samples/sec
+epoch 7, total_step 113880, total loss is 15.74 , inference loss is 8.10, weight deacy loss is 7.64, training accuracy is 0.375000, time 55.927 samples/sec
+epoch 7, total_step 113900, total loss is 14.18 , inference loss is 6.54, weight deacy loss is 7.64, training accuracy is 0.343750, time 55.978 samples/sec
+epoch 7, total_step 113920, total loss is 19.34 , inference loss is 11.69, weight deacy loss is 7.64, training accuracy is 0.312500, time 56.458 samples/sec
+epoch 7, total_step 113940, total loss is 12.79 , inference loss is 5.15, weight deacy loss is 7.64, training accuracy is 0.500000, time 56.101 samples/sec
+epoch 7, total_step 113960, total loss is 18.36 , inference loss is 10.72, weight deacy loss is 7.64, training accuracy is 0.312500, time 56.339 samples/sec
+epoch 7, total_step 113980, total loss is 17.12 , inference loss is 9.47, weight deacy loss is 7.64, training accuracy is 0.406250, time 56.297 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.48386000000002
+best_threshold_index 114 0.9864814814814815
+best_threshold_index 114 0.9864814814814815
+best_threshold_index 114 0.9861111111111112
+best_threshold_index 114 0.9868518518518519
+best_threshold_index 114 0.9866666666666667
+best_threshold_index 111 0.9851851851851852
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 111 0.9855555555555555
+best_threshold_index 114 0.9846296296296296
+best_threshold_index 114 0.9853703703703703
+[lfw][114000]XNorm: 22.669949
+[lfw][114000]Accuracy-Flip: 0.98500+-0.00641
+testing verification..
+(14000, 512)
+infer time 85.31706800000006
+best_threshold_index 116 0.9857142857142858
+best_threshold_index 114 0.9846031746031746
+best_threshold_index 115 0.9846031746031746
+best_threshold_index 116 0.9846031746031746
+best_threshold_index 116 0.986031746031746
+best_threshold_index 116 0.9842857142857143
+best_threshold_index 116 0.9852380952380952
+best_threshold_index 116 0.9841269841269841
+best_threshold_index 116 0.9846031746031746
+best_threshold_index 115 0.984920634920635
+[cfp_ff][114000]XNorm: 22.253392
+[cfp_ff][114000]Accuracy-Flip: 0.98400+-0.00514
+testing verification..
+(14000, 512)
+infer time 85.18311399999997
+best_threshold_index 135 0.9138095238095238
+best_threshold_index 134 0.9115873015873016
+best_threshold_index 139 0.9117460317460317
+best_threshold_index 135 0.9130158730158731
+best_threshold_index 135 0.9165079365079365
+best_threshold_index 135 0.91
+best_threshold_index 139 0.9092063492063492
+best_threshold_index 139 0.910952380952381
+best_threshold_index 135 0.9112698412698412
+best_threshold_index 135 0.9123809523809524
+[cfp_fp][114000]XNorm: 22.178458
+[cfp_fp][114000]Accuracy-Flip: 0.90971+-0.01759
+testing verification..
+(12000, 512)
+infer time 73.198036
+best_threshold_index 139 0.9087037037037037
+best_threshold_index 135 0.9068518518518518
+best_threshold_index 135 0.9077777777777778
+best_threshold_index 138 0.9103703703703704
+best_threshold_index 138 0.912962962962963
+best_threshold_index 134 0.9085185185185185
+best_threshold_index 135 0.909074074074074
+best_threshold_index 139 0.9151851851851852
+best_threshold_index 139 0.9131481481481482
+best_threshold_index 139 0.9118518518518518
+[agedb_30][114000]XNorm: 23.273659
+[agedb_30][114000]Accuracy-Flip: 0.90733+-0.02199
+test accuracy is: 0.9850000000000001
+epoch 7, total_step 114000, total loss is 11.64 , inference loss is 4.00, weight deacy loss is 7.64, training accuracy is 0.500000, time 53.823 samples/sec
+epoch 7, total_step 114020, total loss is 15.05 , inference loss is 7.41, weight deacy loss is 7.64, training accuracy is 0.281250, time 57.263 samples/sec
+epoch 7, total_step 114040, total loss is 15.36 , inference loss is 7.72, weight deacy loss is 7.64, training accuracy is 0.500000, time 56.290 samples/sec
+epoch 7, total_step 114060, total loss is 17.03 , inference loss is 9.39, weight deacy loss is 7.64, training accuracy is 0.375000, time 56.916 samples/sec
+epoch 7, total_step 114080, total loss is 18.62 , inference loss is 10.98, weight deacy loss is 7.64, training accuracy is 0.375000, time 56.684 samples/sec
+epoch 7, total_step 114100, total loss is 17.77 , inference loss is 10.13, weight deacy loss is 7.64, training accuracy is 0.406250, time 57.272 samples/sec
+epoch 7, total_step 114120, total loss is 18.55 , inference loss is 10.91, weight deacy loss is 7.64, training accuracy is 0.343750, time 56.224 samples/sec
+epoch 7, total_step 114140, total loss is 16.79 , inference loss is 9.15, weight deacy loss is 7.64, training accuracy is 0.312500, time 56.665 samples/sec
+epoch 7, total_step 114160, total loss is 17.20 , inference loss is 9.57, weight deacy loss is 7.64, training accuracy is 0.375000, time 55.831 samples/sec
+epoch 7, total_step 114180, total loss is 19.88 , inference loss is 12.24, weight deacy loss is 7.64, training accuracy is 0.281250, time 56.280 samples/sec
+epoch 7, total_step 114200, total loss is 15.90 , inference loss is 8.27, weight deacy loss is 7.63, training accuracy is 0.312500, time 56.766 samples/sec
+epoch 7, total_step 114220, total loss is 15.54 , inference loss is 7.90, weight deacy loss is 7.63, training accuracy is 0.375000, time 56.479 samples/sec
+epoch 7, total_step 114240, total loss is 17.12 , inference loss is 9.49, weight deacy loss is 7.63, training accuracy is 0.343750, time 56.886 samples/sec
+epoch 7, total_step 114260, total loss is 15.99 , inference loss is 8.35, weight deacy loss is 7.63, training accuracy is 0.343750, time 56.252 samples/sec
+epoch 7, total_step 114280, total loss is 15.65 , inference loss is 8.02, weight deacy loss is 7.63, training accuracy is 0.312500, time 57.070 samples/sec
+epoch 7, total_step 114300, total loss is 19.27 , inference loss is 11.64, weight deacy loss is 7.63, training accuracy is 0.281250, time 55.452 samples/sec
+epoch 7, total_step 114320, total loss is 16.84 , inference loss is 9.21, weight deacy loss is 7.63, training accuracy is 0.375000, time 55.916 samples/sec
+epoch 7, total_step 114340, total loss is 17.51 , inference loss is 9.88, weight deacy loss is 7.63, training accuracy is 0.312500, time 56.626 samples/sec
+epoch 7, total_step 114360, total loss is 13.89 , inference loss is 6.26, weight deacy loss is 7.63, training accuracy is 0.437500, time 55.352 samples/sec
+epoch 7, total_step 114380, total loss is 17.81 , inference loss is 10.18, weight deacy loss is 7.63, training accuracy is 0.218750, time 55.867 samples/sec
+epoch 7, total_step 114400, total loss is 16.39 , inference loss is 8.76, weight deacy loss is 7.63, training accuracy is 0.406250, time 56.279 samples/sec
+epoch 7, total_step 114420, total loss is 18.52 , inference loss is 10.89, weight deacy loss is 7.63, training accuracy is 0.343750, time 56.271 samples/sec
+epoch 7, total_step 114440, total loss is 17.34 , inference loss is 9.71, weight deacy loss is 7.63, training accuracy is 0.437500, time 56.820 samples/sec
+epoch 7, total_step 114460, total loss is 19.06 , inference loss is 11.43, weight deacy loss is 7.63, training accuracy is 0.343750, time 56.751 samples/sec
+epoch 7, total_step 114480, total loss is 16.28 , inference loss is 8.65, weight deacy loss is 7.63, training accuracy is 0.468750, time 57.176 samples/sec
+epoch 7, total_step 114500, total loss is 16.50 , inference loss is 8.87, weight deacy loss is 7.63, training accuracy is 0.250000, time 56.992 samples/sec
+epoch 7, total_step 114520, total loss is 17.73 , inference loss is 10.10, weight deacy loss is 7.63, training accuracy is 0.406250, time 56.544 samples/sec
+epoch 7, total_step 114540, total loss is 18.86 , inference loss is 11.24, weight deacy loss is 7.63, training accuracy is 0.406250, time 56.299 samples/sec
+epoch 7, total_step 114560, total loss is 18.15 , inference loss is 10.52, weight deacy loss is 7.62, training accuracy is 0.343750, time 56.407 samples/sec
+epoch 7, total_step 114580, total loss is 19.32 , inference loss is 11.69, weight deacy loss is 7.62, training accuracy is 0.375000, time 55.482 samples/sec
+epoch 7, total_step 114600, total loss is 16.83 , inference loss is 9.21, weight deacy loss is 7.62, training accuracy is 0.437500, time 55.552 samples/sec
+epoch 7, total_step 114620, total loss is 18.37 , inference loss is 10.75, weight deacy loss is 7.62, training accuracy is 0.281250, time 56.868 samples/sec
+epoch 7, total_step 114640, total loss is 16.05 , inference loss is 8.42, weight deacy loss is 7.62, training accuracy is 0.218750, time 56.171 samples/sec
+epoch 7, total_step 114660, total loss is 19.66 , inference loss is 12.04, weight deacy loss is 7.62, training accuracy is 0.250000, time 56.975 samples/sec
+epoch 7, total_step 114680, total loss is 14.34 , inference loss is 6.72, weight deacy loss is 7.62, training accuracy is 0.500000, time 57.587 samples/sec
+epoch 7, total_step 114700, total loss is 19.63 , inference loss is 12.01, weight deacy loss is 7.62, training accuracy is 0.250000, time 56.112 samples/sec
+epoch 7, total_step 114720, total loss is 14.44 , inference loss is 6.82, weight deacy loss is 7.62, training accuracy is 0.343750, time 57.019 samples/sec
+epoch 7, total_step 114740, total loss is 18.39 , inference loss is 10.77, weight deacy loss is 7.62, training accuracy is 0.343750, time 56.392 samples/sec
+epoch 7, total_step 114760, total loss is 15.52 , inference loss is 7.91, weight deacy loss is 7.62, training accuracy is 0.375000, time 56.362 samples/sec
+epoch 7, total_step 114780, total loss is 14.59 , inference loss is 6.98, weight deacy loss is 7.62, training accuracy is 0.562500, time 56.167 samples/sec
+epoch 7, total_step 114800, total loss is 14.85 , inference loss is 7.24, weight deacy loss is 7.62, training accuracy is 0.343750, time 57.522 samples/sec
+epoch 7, total_step 114820, total loss is 17.29 , inference loss is 9.67, weight deacy loss is 7.62, training accuracy is 0.343750, time 56.688 samples/sec
+epoch 7, total_step 114840, total loss is 15.67 , inference loss is 8.05, weight deacy loss is 7.62, training accuracy is 0.375000, time 55.789 samples/sec
+epoch 7, total_step 114860, total loss is 15.79 , inference loss is 8.18, weight deacy loss is 7.62, training accuracy is 0.437500, time 56.755 samples/sec
+epoch 7, total_step 114880, total loss is 17.27 , inference loss is 9.66, weight deacy loss is 7.62, training accuracy is 0.312500, time 57.150 samples/sec
+epoch 7, total_step 114900, total loss is 17.72 , inference loss is 10.11, weight deacy loss is 7.61, training accuracy is 0.468750, time 56.051 samples/sec
+epoch 7, total_step 114920, total loss is 18.26 , inference loss is 10.64, weight deacy loss is 7.61, training accuracy is 0.406250, time 56.927 samples/sec
+epoch 7, total_step 114940, total loss is 21.74 , inference loss is 14.12, weight deacy loss is 7.61, training accuracy is 0.375000, time 56.716 samples/sec
+epoch 7, total_step 114960, total loss is 19.28 , inference loss is 11.67, weight deacy loss is 7.61, training accuracy is 0.281250, time 56.119 samples/sec
+epoch 7, total_step 114980, total loss is 18.05 , inference loss is 10.44, weight deacy loss is 7.61, training accuracy is 0.406250, time 56.340 samples/sec
+epoch 7, total_step 115000, total loss is 16.05 , inference loss is 8.44, weight deacy loss is 7.61, training accuracy is 0.531250, time 56.835 samples/sec
+epoch 7, total_step 115020, total loss is 14.06 , inference loss is 6.45, weight deacy loss is 7.61, training accuracy is 0.375000, time 55.486 samples/sec
+epoch 7, total_step 115040, total loss is 17.41 , inference loss is 9.80, weight deacy loss is 7.61, training accuracy is 0.343750, time 56.531 samples/sec
+epoch 7, total_step 115060, total loss is 18.55 , inference loss is 10.94, weight deacy loss is 7.61, training accuracy is 0.375000, time 57.164 samples/sec
+epoch 7, total_step 115080, total loss is 18.94 , inference loss is 11.33, weight deacy loss is 7.61, training accuracy is 0.312500, time 56.814 samples/sec
+epoch 7, total_step 115100, total loss is 15.86 , inference loss is 8.25, weight deacy loss is 7.61, training accuracy is 0.406250, time 55.828 samples/sec
+epoch 7, total_step 115120, total loss is 15.56 , inference loss is 7.95, weight deacy loss is 7.61, training accuracy is 0.437500, time 55.835 samples/sec
+epoch 7, total_step 115140, total loss is 14.96 , inference loss is 7.35, weight deacy loss is 7.61, training accuracy is 0.375000, time 56.322 samples/sec
+epoch 7, total_step 115160, total loss is 19.22 , inference loss is 11.61, weight deacy loss is 7.61, training accuracy is 0.218750, time 56.144 samples/sec
+epoch 7, total_step 115180, total loss is 17.75 , inference loss is 10.14, weight deacy loss is 7.61, training accuracy is 0.406250, time 57.842 samples/sec
+epoch 7, total_step 115200, total loss is 20.28 , inference loss is 12.67, weight deacy loss is 7.61, training accuracy is 0.312500, time 55.133 samples/sec
+epoch 7, total_step 115220, total loss is 19.95 , inference loss is 12.34, weight deacy loss is 7.60, training accuracy is 0.281250, time 56.658 samples/sec
+epoch 7, total_step 115240, total loss is 16.24 , inference loss is 8.63, weight deacy loss is 7.60, training accuracy is 0.375000, time 55.447 samples/sec
+epoch 7, total_step 115260, total loss is 14.68 , inference loss is 7.08, weight deacy loss is 7.60, training accuracy is 0.468750, time 56.892 samples/sec
+epoch 7, total_step 115280, total loss is 17.67 , inference loss is 10.07, weight deacy loss is 7.60, training accuracy is 0.281250, time 55.062 samples/sec
+epoch 7, total_step 115300, total loss is 13.89 , inference loss is 6.29, weight deacy loss is 7.60, training accuracy is 0.343750, time 57.313 samples/sec
+epoch 7, total_step 115320, total loss is 17.58 , inference loss is 9.98, weight deacy loss is 7.60, training accuracy is 0.281250, time 56.185 samples/sec
+epoch 7, total_step 115340, total loss is 20.69 , inference loss is 13.09, weight deacy loss is 7.60, training accuracy is 0.312500, time 56.674 samples/sec
+epoch 7, total_step 115360, total loss is 19.41 , inference loss is 11.81, weight deacy loss is 7.60, training accuracy is 0.156250, time 57.165 samples/sec
+epoch 7, total_step 115380, total loss is 17.80 , inference loss is 10.20, weight deacy loss is 7.60, training accuracy is 0.250000, time 56.964 samples/sec
+epoch 7, total_step 115400, total loss is 16.93 , inference loss is 9.33, weight deacy loss is 7.60, training accuracy is 0.281250, time 56.242 samples/sec
+epoch 7, total_step 115420, total loss is 17.21 , inference loss is 9.61, weight deacy loss is 7.60, training accuracy is 0.312500, time 55.662 samples/sec
+epoch 7, total_step 115440, total loss is 18.91 , inference loss is 11.31, weight deacy loss is 7.60, training accuracy is 0.375000, time 56.825 samples/sec
+epoch 7, total_step 115460, total loss is 13.68 , inference loss is 6.08, weight deacy loss is 7.60, training accuracy is 0.437500, time 56.815 samples/sec
+epoch 7, total_step 115480, total loss is 15.78 , inference loss is 8.18, weight deacy loss is 7.60, training accuracy is 0.437500, time 56.213 samples/sec
+epoch 7, total_step 115500, total loss is 19.35 , inference loss is 11.76, weight deacy loss is 7.60, training accuracy is 0.250000, time 55.958 samples/sec
+epoch 7, total_step 115520, total loss is 15.59 , inference loss is 7.99, weight deacy loss is 7.60, training accuracy is 0.375000, time 57.012 samples/sec
+epoch 7, total_step 115540, total loss is 17.06 , inference loss is 9.46, weight deacy loss is 7.60, training accuracy is 0.375000, time 56.909 samples/sec
+epoch 7, total_step 115560, total loss is 18.49 , inference loss is 10.89, weight deacy loss is 7.60, training accuracy is 0.250000, time 56.437 samples/sec
+epoch 7, total_step 115580, total loss is 15.87 , inference loss is 8.28, weight deacy loss is 7.59, training accuracy is 0.343750, time 56.421 samples/sec
+epoch 7, total_step 115600, total loss is 15.09 , inference loss is 7.50, weight deacy loss is 7.59, training accuracy is 0.343750, time 56.741 samples/sec
+epoch 7, total_step 115620, total loss is 17.61 , inference loss is 10.02, weight deacy loss is 7.59, training accuracy is 0.406250, time 55.677 samples/sec
+epoch 7, total_step 115640, total loss is 21.27 , inference loss is 13.68, weight deacy loss is 7.59, training accuracy is 0.250000, time 56.745 samples/sec
+epoch 7, total_step 115660, total loss is 15.29 , inference loss is 7.70, weight deacy loss is 7.59, training accuracy is 0.312500, time 56.720 samples/sec
+epoch 7, total_step 115680, total loss is 21.31 , inference loss is 13.72, weight deacy loss is 7.59, training accuracy is 0.218750, time 56.655 samples/sec
+epoch 7, total_step 115700, total loss is 21.65 , inference loss is 14.06, weight deacy loss is 7.59, training accuracy is 0.125000, time 56.663 samples/sec
+epoch 7, total_step 115720, total loss is 16.20 , inference loss is 8.61, weight deacy loss is 7.59, training accuracy is 0.312500, time 56.688 samples/sec
+epoch 7, total_step 115740, total loss is 15.39 , inference loss is 7.80, weight deacy loss is 7.59, training accuracy is 0.343750, time 56.517 samples/sec
+epoch 7, total_step 115760, total loss is 14.83 , inference loss is 7.24, weight deacy loss is 7.59, training accuracy is 0.406250, time 56.556 samples/sec
+epoch 7, total_step 115780, total loss is 14.45 , inference loss is 6.86, weight deacy loss is 7.59, training accuracy is 0.312500, time 55.846 samples/sec
+epoch 7, total_step 115800, total loss is 15.47 , inference loss is 7.88, weight deacy loss is 7.59, training accuracy is 0.562500, time 54.185 samples/sec
+epoch 7, total_step 115820, total loss is 16.68 , inference loss is 9.09, weight deacy loss is 7.59, training accuracy is 0.375000, time 55.869 samples/sec
+epoch 7, total_step 115840, total loss is 18.97 , inference loss is 11.38, weight deacy loss is 7.59, training accuracy is 0.375000, time 56.428 samples/sec
+epoch 7, total_step 115860, total loss is 15.90 , inference loss is 8.32, weight deacy loss is 7.59, training accuracy is 0.406250, time 56.581 samples/sec
+epoch 7, total_step 115880, total loss is 19.14 , inference loss is 11.56, weight deacy loss is 7.59, training accuracy is 0.218750, time 57.572 samples/sec
+epoch 7, total_step 115900, total loss is 15.75 , inference loss is 8.16, weight deacy loss is 7.59, training accuracy is 0.343750, time 56.957 samples/sec
+epoch 7, total_step 115920, total loss is 14.07 , inference loss is 6.49, weight deacy loss is 7.58, training accuracy is 0.343750, time 56.394 samples/sec
+epoch 7, total_step 115940, total loss is 17.67 , inference loss is 10.08, weight deacy loss is 7.58, training accuracy is 0.406250, time 56.170 samples/sec
+epoch 7, total_step 115960, total loss is 18.02 , inference loss is 10.43, weight deacy loss is 7.58, training accuracy is 0.375000, time 56.242 samples/sec
+epoch 7, total_step 115980, total loss is 14.03 , inference loss is 6.44, weight deacy loss is 7.58, training accuracy is 0.468750, time 56.289 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.03847700000003
+best_threshold_index 115 0.9864814814814815
+best_threshold_index 116 0.9853703703703703
+best_threshold_index 108 0.9868518518518519
+best_threshold_index 108 0.987037037037037
+best_threshold_index 110 0.9861111111111112
+best_threshold_index 112 0.9853703703703703
+best_threshold_index 111 0.9864814814814815
+best_threshold_index 116 0.985
+best_threshold_index 116 0.9851851851851852
+best_threshold_index 112 0.9846296296296296
+[lfw][116000]XNorm: 21.860361
+[lfw][116000]Accuracy-Flip: 0.98300+-0.00806
+testing verification..
+(14000, 512)
+infer time 85.27343499999994
+best_threshold_index 113 0.9873015873015873
+best_threshold_index 113 0.9863492063492063
+best_threshold_index 113 0.9865079365079366
+best_threshold_index 113 0.986031746031746
+best_threshold_index 118 0.9863492063492063
+best_threshold_index 113 0.986031746031746
+best_threshold_index 113 0.9866666666666667
+best_threshold_index 113 0.986031746031746
+best_threshold_index 118 0.9866666666666667
+best_threshold_index 113 0.9868253968253968
+[cfp_ff][116000]XNorm: 21.569554
+[cfp_ff][116000]Accuracy-Flip: 0.98571+-0.00356
+testing verification..
+(14000, 512)
+infer time 85.42846199999997
+best_threshold_index 131 0.9206349206349206
+best_threshold_index 131 0.9204761904761904
+best_threshold_index 131 0.9196825396825397
+best_threshold_index 131 0.920952380952381
+best_threshold_index 131 0.9234920634920635
+best_threshold_index 131 0.9188888888888889
+best_threshold_index 130 0.9184126984126985
+best_threshold_index 131 0.919047619047619
+best_threshold_index 135 0.9198412698412698
+best_threshold_index 131 0.919047619047619
+[cfp_fp][116000]XNorm: 21.636699
+[cfp_fp][116000]Accuracy-Flip: 0.91871+-0.01225
+testing verification..
+(12000, 512)
+infer time 72.72354199999998
+best_threshold_index 131 0.9007407407407407
+best_threshold_index 130 0.9007407407407407
+best_threshold_index 131 0.9025925925925926
+best_threshold_index 131 0.9037037037037037
+best_threshold_index 131 0.9059259259259259
+best_threshold_index 131 0.9016666666666666
+best_threshold_index 131 0.9037037037037037
+best_threshold_index 131 0.9068518518518518
+best_threshold_index 131 0.9077777777777778
+best_threshold_index 131 0.9046296296296297
+[agedb_30][116000]XNorm: 22.364655
+[agedb_30][116000]Accuracy-Flip: 0.90300+-0.02015
+test accuracy is: 0.983
+epoch 7, total_step 116000, total loss is 18.19 , inference loss is 10.61, weight deacy loss is 7.58, training accuracy is 0.250000, time 53.590 samples/sec
+epoch 7, total_step 116020, total loss is 16.97 , inference loss is 9.39, weight deacy loss is 7.58, training accuracy is 0.343750, time 56.340 samples/sec
+epoch 7, total_step 116040, total loss is 18.69 , inference loss is 11.11, weight deacy loss is 7.58, training accuracy is 0.218750, time 56.686 samples/sec
+epoch 7, total_step 116060, total loss is 18.59 , inference loss is 11.01, weight deacy loss is 7.58, training accuracy is 0.375000, time 57.153 samples/sec
+epoch 7, total_step 116080, total loss is 19.06 , inference loss is 11.48, weight deacy loss is 7.58, training accuracy is 0.250000, time 56.407 samples/sec
+epoch 7, total_step 116100, total loss is 15.98 , inference loss is 8.40, weight deacy loss is 7.58, training accuracy is 0.437500, time 56.455 samples/sec
+epoch 7, total_step 116120, total loss is 16.10 , inference loss is 8.52, weight deacy loss is 7.58, training accuracy is 0.375000, time 56.991 samples/sec
+epoch 7, total_step 116140, total loss is 17.87 , inference loss is 10.29, weight deacy loss is 7.58, training accuracy is 0.312500, time 56.628 samples/sec
+epoch 7, total_step 116160, total loss is 18.01 , inference loss is 10.43, weight deacy loss is 7.58, training accuracy is 0.156250, time 56.463 samples/sec
+epoch 7, total_step 116180, total loss is 12.18 , inference loss is 4.60, weight deacy loss is 7.58, training accuracy is 0.562500, time 57.238 samples/sec
+epoch 7, total_step 116200, total loss is 15.35 , inference loss is 7.77, weight deacy loss is 7.58, training accuracy is 0.375000, time 56.883 samples/sec
+epoch 7, total_step 116220, total loss is 15.78 , inference loss is 8.20, weight deacy loss is 7.58, training accuracy is 0.437500, time 56.580 samples/sec
+epoch 7, total_step 116240, total loss is 20.12 , inference loss is 12.55, weight deacy loss is 7.58, training accuracy is 0.250000, time 56.944 samples/sec
+epoch 7, total_step 116260, total loss is 14.40 , inference loss is 6.83, weight deacy loss is 7.58, training accuracy is 0.531250, time 56.810 samples/sec
+epoch 7, total_step 116280, total loss is 17.35 , inference loss is 9.78, weight deacy loss is 7.58, training accuracy is 0.187500, time 57.420 samples/sec
+epoch 7, total_step 116300, total loss is 19.28 , inference loss is 11.70, weight deacy loss is 7.57, training accuracy is 0.281250, time 57.279 samples/sec
+epoch 7, total_step 116320, total loss is 16.93 , inference loss is 9.36, weight deacy loss is 7.57, training accuracy is 0.406250, time 56.895 samples/sec
+epoch 7, total_step 116340, total loss is 13.20 , inference loss is 5.63, weight deacy loss is 7.57, training accuracy is 0.468750, time 56.711 samples/sec
+epoch 7, total_step 116360, total loss is 15.70 , inference loss is 8.13, weight deacy loss is 7.57, training accuracy is 0.406250, time 56.219 samples/sec
+epoch 7, total_step 116380, total loss is 16.66 , inference loss is 9.08, weight deacy loss is 7.57, training accuracy is 0.406250, time 55.936 samples/sec
+epoch 7, total_step 116400, total loss is 19.52 , inference loss is 11.95, weight deacy loss is 7.57, training accuracy is 0.343750, time 56.627 samples/sec
+epoch 7, total_step 116420, total loss is 17.06 , inference loss is 9.49, weight deacy loss is 7.57, training accuracy is 0.281250, time 56.851 samples/sec
+epoch 7, total_step 116440, total loss is 17.06 , inference loss is 9.49, weight deacy loss is 7.57, training accuracy is 0.343750, time 56.551 samples/sec
+epoch 7, total_step 116460, total loss is 15.23 , inference loss is 7.66, weight deacy loss is 7.57, training accuracy is 0.343750, time 56.315 samples/sec
+epoch 7, total_step 116480, total loss is 17.57 , inference loss is 10.00, weight deacy loss is 7.57, training accuracy is 0.281250, time 57.154 samples/sec
+epoch 7, total_step 116500, total loss is 16.12 , inference loss is 8.55, weight deacy loss is 7.57, training accuracy is 0.437500, time 56.851 samples/sec
+epoch 7, total_step 116520, total loss is 17.32 , inference loss is 9.75, weight deacy loss is 7.57, training accuracy is 0.312500, time 56.462 samples/sec
+epoch 7, total_step 116540, total loss is 18.25 , inference loss is 10.68, weight deacy loss is 7.57, training accuracy is 0.281250, time 56.831 samples/sec
+epoch 7, total_step 116560, total loss is 16.36 , inference loss is 8.79, weight deacy loss is 7.57, training accuracy is 0.312500, time 55.867 samples/sec
+epoch 7, total_step 116580, total loss is 17.96 , inference loss is 10.40, weight deacy loss is 7.57, training accuracy is 0.312500, time 55.605 samples/sec
+epoch 7, total_step 116600, total loss is 13.34 , inference loss is 5.78, weight deacy loss is 7.57, training accuracy is 0.468750, time 56.128 samples/sec
+epoch 7, total_step 116620, total loss is 12.48 , inference loss is 4.92, weight deacy loss is 7.57, training accuracy is 0.500000, time 56.489 samples/sec
+epoch 7, total_step 116640, total loss is 18.42 , inference loss is 10.85, weight deacy loss is 7.57, training accuracy is 0.343750, time 56.594 samples/sec
+epoch 7, total_step 116660, total loss is 15.57 , inference loss is 8.01, weight deacy loss is 7.56, training accuracy is 0.531250, time 55.997 samples/sec
+epoch 7, total_step 116680, total loss is 16.74 , inference loss is 9.18, weight deacy loss is 7.56, training accuracy is 0.375000, time 56.077 samples/sec
+epoch 7, total_step 116700, total loss is 20.93 , inference loss is 13.36, weight deacy loss is 7.56, training accuracy is 0.343750, time 56.731 samples/sec
+epoch 7, total_step 116720, total loss is 16.18 , inference loss is 8.62, weight deacy loss is 7.56, training accuracy is 0.312500, time 57.317 samples/sec
+epoch 7, total_step 116740, total loss is 17.52 , inference loss is 9.96, weight deacy loss is 7.56, training accuracy is 0.500000, time 57.263 samples/sec
+epoch 7, total_step 116760, total loss is 14.00 , inference loss is 6.44, weight deacy loss is 7.56, training accuracy is 0.468750, time 56.065 samples/sec
+epoch 7, total_step 116780, total loss is 15.06 , inference loss is 7.50, weight deacy loss is 7.56, training accuracy is 0.343750, time 56.367 samples/sec
+epoch 7, total_step 116800, total loss is 17.93 , inference loss is 10.37, weight deacy loss is 7.56, training accuracy is 0.218750, time 56.954 samples/sec
+epoch 7, total_step 116820, total loss is 20.51 , inference loss is 12.95, weight deacy loss is 7.56, training accuracy is 0.250000, time 56.136 samples/sec
+epoch 7, total_step 116840, total loss is 17.62 , inference loss is 10.06, weight deacy loss is 7.56, training accuracy is 0.437500, time 56.970 samples/sec
+epoch 7, total_step 116860, total loss is 14.68 , inference loss is 7.12, weight deacy loss is 7.56, training accuracy is 0.406250, time 56.840 samples/sec
+epoch 7, total_step 116880, total loss is 21.82 , inference loss is 14.26, weight deacy loss is 7.56, training accuracy is 0.343750, time 56.324 samples/sec
+epoch 7, total_step 116900, total loss is 16.79 , inference loss is 9.23, weight deacy loss is 7.56, training accuracy is 0.343750, time 55.614 samples/sec
+epoch 7, total_step 116920, total loss is 16.92 , inference loss is 9.36, weight deacy loss is 7.56, training accuracy is 0.312500, time 56.437 samples/sec
+epoch 7, total_step 116940, total loss is 17.61 , inference loss is 10.06, weight deacy loss is 7.56, training accuracy is 0.343750, time 56.771 samples/sec
+epoch 7, total_step 116960, total loss is 22.41 , inference loss is 14.85, weight deacy loss is 7.56, training accuracy is 0.187500, time 56.602 samples/sec
+epoch 7, total_step 116980, total loss is 17.25 , inference loss is 9.70, weight deacy loss is 7.56, training accuracy is 0.500000, time 56.779 samples/sec
+epoch 7, total_step 117000, total loss is 17.24 , inference loss is 9.69, weight deacy loss is 7.56, training accuracy is 0.406250, time 56.037 samples/sec
+epoch 7, total_step 117020, total loss is 16.27 , inference loss is 8.72, weight deacy loss is 7.55, training accuracy is 0.375000, time 56.270 samples/sec
+epoch 7, total_step 117040, total loss is 15.91 , inference loss is 8.35, weight deacy loss is 7.55, training accuracy is 0.406250, time 56.074 samples/sec
+epoch 7, total_step 117060, total loss is 16.82 , inference loss is 9.27, weight deacy loss is 7.55, training accuracy is 0.218750, time 56.671 samples/sec
+epoch 7, total_step 117080, total loss is 14.97 , inference loss is 7.42, weight deacy loss is 7.55, training accuracy is 0.343750, time 56.558 samples/sec
+epoch 7, total_step 117100, total loss is 18.54 , inference loss is 10.98, weight deacy loss is 7.55, training accuracy is 0.250000, time 56.870 samples/sec
+epoch 7, total_step 117120, total loss is 17.48 , inference loss is 9.93, weight deacy loss is 7.55, training accuracy is 0.312500, time 57.351 samples/sec
+epoch 7, total_step 117140, total loss is 18.94 , inference loss is 11.39, weight deacy loss is 7.55, training accuracy is 0.343750, time 56.266 samples/sec
+epoch 7, total_step 117160, total loss is 14.76 , inference loss is 7.21, weight deacy loss is 7.55, training accuracy is 0.531250, time 56.702 samples/sec
+epoch 7, total_step 117180, total loss is 16.08 , inference loss is 8.53, weight deacy loss is 7.55, training accuracy is 0.281250, time 56.881 samples/sec
+epoch 7, total_step 117200, total loss is 17.95 , inference loss is 10.40, weight deacy loss is 7.55, training accuracy is 0.343750, time 57.499 samples/sec
+epoch 7, total_step 117220, total loss is 14.50 , inference loss is 6.96, weight deacy loss is 7.55, training accuracy is 0.406250, time 56.697 samples/sec
+epoch 7, total_step 117240, total loss is 20.39 , inference loss is 12.84, weight deacy loss is 7.55, training accuracy is 0.218750, time 55.970 samples/sec
+epoch 7, total_step 117260, total loss is 12.31 , inference loss is 4.76, weight deacy loss is 7.55, training accuracy is 0.468750, time 55.867 samples/sec
+epoch 7, total_step 117280, total loss is 18.92 , inference loss is 11.38, weight deacy loss is 7.55, training accuracy is 0.281250, time 55.655 samples/sec
+epoch 7, total_step 117300, total loss is 17.96 , inference loss is 10.41, weight deacy loss is 7.55, training accuracy is 0.437500, time 56.773 samples/sec
+epoch 7, total_step 117320, total loss is 14.85 , inference loss is 7.31, weight deacy loss is 7.55, training accuracy is 0.375000, time 56.249 samples/sec
+epoch 7, total_step 117340, total loss is 19.43 , inference loss is 11.88, weight deacy loss is 7.55, training accuracy is 0.250000, time 56.419 samples/sec
+epoch 7, total_step 117360, total loss is 12.73 , inference loss is 5.19, weight deacy loss is 7.54, training accuracy is 0.375000, time 56.607 samples/sec
+epoch 7, total_step 117380, total loss is 17.18 , inference loss is 9.63, weight deacy loss is 7.54, training accuracy is 0.437500, time 56.329 samples/sec
+epoch 7, total_step 117400, total loss is 18.15 , inference loss is 10.60, weight deacy loss is 7.54, training accuracy is 0.281250, time 55.003 samples/sec
+epoch 7, total_step 117420, total loss is 17.17 , inference loss is 9.63, weight deacy loss is 7.54, training accuracy is 0.250000, time 56.233 samples/sec
+epoch 7, total_step 117440, total loss is 16.24 , inference loss is 8.69, weight deacy loss is 7.54, training accuracy is 0.406250, time 56.633 samples/sec
+epoch 7, total_step 117460, total loss is 14.70 , inference loss is 7.16, weight deacy loss is 7.54, training accuracy is 0.375000, time 57.138 samples/sec
+epoch 7, total_step 117480, total loss is 14.30 , inference loss is 6.75, weight deacy loss is 7.54, training accuracy is 0.406250, time 55.608 samples/sec
+epoch 7, total_step 117500, total loss is 16.02 , inference loss is 8.48, weight deacy loss is 7.54, training accuracy is 0.375000, time 56.329 samples/sec
+epoch 7, total_step 117520, total loss is 15.24 , inference loss is 7.70, weight deacy loss is 7.54, training accuracy is 0.250000, time 56.030 samples/sec
+epoch 7, total_step 117540, total loss is 16.90 , inference loss is 9.36, weight deacy loss is 7.54, training accuracy is 0.250000, time 56.762 samples/sec
+epoch 7, total_step 117560, total loss is 19.64 , inference loss is 12.10, weight deacy loss is 7.54, training accuracy is 0.375000, time 55.555 samples/sec
+epoch 7, total_step 117580, total loss is 20.90 , inference loss is 13.36, weight deacy loss is 7.54, training accuracy is 0.312500, time 55.730 samples/sec
+epoch 7, total_step 117600, total loss is 14.51 , inference loss is 6.97, weight deacy loss is 7.54, training accuracy is 0.406250, time 55.831 samples/sec
+epoch 7, total_step 117620, total loss is 17.15 , inference loss is 9.61, weight deacy loss is 7.54, training accuracy is 0.375000, time 56.812 samples/sec
+epoch 7, total_step 117640, total loss is 17.63 , inference loss is 10.09, weight deacy loss is 7.54, training accuracy is 0.375000, time 56.178 samples/sec
+epoch 7, total_step 117660, total loss is 13.54 , inference loss is 6.01, weight deacy loss is 7.54, training accuracy is 0.406250, time 56.055 samples/sec
+epoch 7, total_step 117680, total loss is 19.00 , inference loss is 11.46, weight deacy loss is 7.54, training accuracy is 0.281250, time 56.364 samples/sec
+epoch 7, total_step 117700, total loss is 15.40 , inference loss is 7.87, weight deacy loss is 7.54, training accuracy is 0.437500, time 56.434 samples/sec
+epoch 7, total_step 117720, total loss is 16.90 , inference loss is 9.36, weight deacy loss is 7.53, training accuracy is 0.343750, time 56.464 samples/sec
+epoch 7, total_step 117740, total loss is 18.00 , inference loss is 10.46, weight deacy loss is 7.53, training accuracy is 0.437500, time 55.817 samples/sec
+epoch 7, total_step 117760, total loss is 21.30 , inference loss is 13.76, weight deacy loss is 7.53, training accuracy is 0.281250, time 55.668 samples/sec
+epoch 7, total_step 117780, total loss is 16.21 , inference loss is 8.67, weight deacy loss is 7.53, training accuracy is 0.281250, time 56.339 samples/sec
+epoch 7, total_step 117800, total loss is 15.26 , inference loss is 7.72, weight deacy loss is 7.53, training accuracy is 0.312500, time 56.292 samples/sec
+epoch 7, total_step 117820, total loss is 18.54 , inference loss is 11.01, weight deacy loss is 7.53, training accuracy is 0.281250, time 54.564 samples/sec
+epoch 7, total_step 117840, total loss is 13.79 , inference loss is 6.26, weight deacy loss is 7.53, training accuracy is 0.312500, time 56.451 samples/sec
+epoch 7, total_step 117860, total loss is 14.90 , inference loss is 7.37, weight deacy loss is 7.53, training accuracy is 0.406250, time 55.679 samples/sec
+epoch 7, total_step 117880, total loss is 17.84 , inference loss is 10.31, weight deacy loss is 7.53, training accuracy is 0.437500, time 56.510 samples/sec
+epoch 7, total_step 117900, total loss is 18.18 , inference loss is 10.65, weight deacy loss is 7.53, training accuracy is 0.281250, time 55.836 samples/sec
+epoch 7, total_step 117920, total loss is 17.69 , inference loss is 10.16, weight deacy loss is 7.53, training accuracy is 0.312500, time 56.734 samples/sec
+epoch 7, total_step 117940, total loss is 20.27 , inference loss is 12.74, weight deacy loss is 7.53, training accuracy is 0.281250, time 55.619 samples/sec
+epoch 7, total_step 117960, total loss is 19.51 , inference loss is 11.98, weight deacy loss is 7.53, training accuracy is 0.281250, time 56.612 samples/sec
+epoch 7, total_step 117980, total loss is 15.24 , inference loss is 7.72, weight deacy loss is 7.53, training accuracy is 0.500000, time 54.097 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.23511900000003
+best_threshold_index 108 0.9842592592592593
+best_threshold_index 108 0.9833333333333333
+best_threshold_index 108 0.9838888888888889
+best_threshold_index 108 0.9842592592592593
+best_threshold_index 108 0.9831481481481481
+best_threshold_index 110 0.9833333333333333
+best_threshold_index 108 0.9837037037037037
+best_threshold_index 113 0.9831481481481481
+best_threshold_index 110 0.9822222222222222
+best_threshold_index 108 0.9827777777777778
+[lfw][118000]XNorm: 21.312240
+[lfw][118000]Accuracy-Flip: 0.98200+-0.00510
+testing verification..
+(14000, 512)
+infer time 85.76414499999998
+best_threshold_index 111 0.9873015873015873
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 111 0.9855555555555555
+best_threshold_index 111 0.986031746031746
+best_threshold_index 111 0.9873015873015873
+best_threshold_index 111 0.9857142857142858
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 111 0.9861904761904762
+best_threshold_index 111 0.9857142857142858
+best_threshold_index 111 0.9857142857142858
+[cfp_ff][118000]XNorm: 21.673582
+[cfp_ff][118000]Accuracy-Flip: 0.98629+-0.00565
+testing verification..
+(14000, 512)
+infer time 85.330546
+best_threshold_index 133 0.9249206349206349
+best_threshold_index 133 0.926984126984127
+best_threshold_index 133 0.9233333333333333
+best_threshold_index 133 0.9263492063492064
+best_threshold_index 130 0.9277777777777778
+best_threshold_index 133 0.9246031746031746
+best_threshold_index 133 0.9230158730158731
+best_threshold_index 133 0.9244444444444444
+best_threshold_index 133 0.9247619047619048
+best_threshold_index 133 0.9242857142857143
+[cfp_fp][118000]XNorm: 21.193070
+[cfp_fp][118000]Accuracy-Flip: 0.92386+-0.01454
+testing verification..
+(12000, 512)
+infer time 73.15120399999994
+best_threshold_index 133 0.902962962962963
+best_threshold_index 133 0.9037037037037037
+best_threshold_index 133 0.9033333333333333
+best_threshold_index 133 0.9053703703703704
+best_threshold_index 133 0.9096296296296297
+best_threshold_index 133 0.9020370370370371
+best_threshold_index 133 0.9044444444444445
+best_threshold_index 133 0.91
+best_threshold_index 133 0.9116666666666666
+best_threshold_index 133 0.9051851851851852
+[agedb_30][118000]XNorm: 22.072490
+[agedb_30][118000]Accuracy-Flip: 0.90583+-0.02872
+test accuracy is: 0.982
+epoch 7, total_step 118000, total loss is 16.63 , inference loss is 9.10, weight deacy loss is 7.53, training accuracy is 0.312500, time 54.840 samples/sec
+epoch 7, total_step 118020, total loss is 15.64 , inference loss is 8.12, weight deacy loss is 7.53, training accuracy is 0.375000, time 56.791 samples/sec
+epoch 7, total_step 118040, total loss is 19.96 , inference loss is 12.44, weight deacy loss is 7.53, training accuracy is 0.312500, time 57.099 samples/sec
+epoch 7, total_step 118060, total loss is 14.67 , inference loss is 7.14, weight deacy loss is 7.53, training accuracy is 0.468750, time 56.475 samples/sec
+epoch 7, total_step 118080, total loss is 20.48 , inference loss is 12.95, weight deacy loss is 7.52, training accuracy is 0.312500, time 56.294 samples/sec
+epoch 7, total_step 118100, total loss is 17.60 , inference loss is 10.08, weight deacy loss is 7.52, training accuracy is 0.343750, time 57.440 samples/sec
+epoch 7, total_step 118120, total loss is 15.67 , inference loss is 8.15, weight deacy loss is 7.52, training accuracy is 0.281250, time 56.349 samples/sec
+epoch 7, total_step 118140, total loss is 16.03 , inference loss is 8.51, weight deacy loss is 7.52, training accuracy is 0.375000, time 56.618 samples/sec
+epoch 7, total_step 118160, total loss is 13.52 , inference loss is 6.00, weight deacy loss is 7.52, training accuracy is 0.406250, time 56.302 samples/sec
+epoch 7, total_step 118180, total loss is 17.07 , inference loss is 9.55, weight deacy loss is 7.52, training accuracy is 0.281250, time 55.807 samples/sec
+epoch 7, total_step 118200, total loss is 14.47 , inference loss is 6.94, weight deacy loss is 7.52, training accuracy is 0.312500, time 54.395 samples/sec
+epoch 7, total_step 118220, total loss is 13.55 , inference loss is 6.03, weight deacy loss is 7.52, training accuracy is 0.406250, time 56.664 samples/sec
+epoch 7, total_step 118240, total loss is 15.77 , inference loss is 8.25, weight deacy loss is 7.52, training accuracy is 0.312500, time 55.958 samples/sec
+epoch 7, total_step 118260, total loss is 14.22 , inference loss is 6.70, weight deacy loss is 7.52, training accuracy is 0.437500, time 56.297 samples/sec
+epoch 7, total_step 118280, total loss is 14.63 , inference loss is 7.11, weight deacy loss is 7.52, training accuracy is 0.437500, time 56.235 samples/sec
+epoch 7, total_step 118300, total loss is 17.97 , inference loss is 10.45, weight deacy loss is 7.52, training accuracy is 0.281250, time 56.333 samples/sec
+epoch 7, total_step 118320, total loss is 16.49 , inference loss is 8.97, weight deacy loss is 7.52, training accuracy is 0.281250, time 56.119 samples/sec
+epoch 7, total_step 118340, total loss is 15.64 , inference loss is 8.13, weight deacy loss is 7.52, training accuracy is 0.468750, time 55.968 samples/sec
+epoch 7, total_step 118360, total loss is 18.50 , inference loss is 10.98, weight deacy loss is 7.52, training accuracy is 0.375000, time 57.008 samples/sec
+epoch 7, total_step 118380, total loss is 16.96 , inference loss is 9.45, weight deacy loss is 7.52, training accuracy is 0.500000, time 56.207 samples/sec
+epoch 7, total_step 118400, total loss is 14.45 , inference loss is 6.93, weight deacy loss is 7.52, training accuracy is 0.437500, time 56.571 samples/sec
+epoch 7, total_step 118420, total loss is 13.70 , inference loss is 6.18, weight deacy loss is 7.52, training accuracy is 0.375000, time 56.381 samples/sec
+epoch 7, total_step 118440, total loss is 16.16 , inference loss is 8.65, weight deacy loss is 7.51, training accuracy is 0.312500, time 56.304 samples/sec
+epoch 7, total_step 118460, total loss is 15.42 , inference loss is 7.91, weight deacy loss is 7.51, training accuracy is 0.468750, time 56.868 samples/sec
+epoch 7, total_step 118480, total loss is 18.93 , inference loss is 11.42, weight deacy loss is 7.51, training accuracy is 0.156250, time 56.649 samples/sec
+epoch 7, total_step 118500, total loss is 13.85 , inference loss is 6.34, weight deacy loss is 7.51, training accuracy is 0.500000, time 56.108 samples/sec
+epoch 7, total_step 118520, total loss is 15.00 , inference loss is 7.49, weight deacy loss is 7.51, training accuracy is 0.312500, time 56.141 samples/sec
+epoch 7, total_step 118540, total loss is 17.70 , inference loss is 10.19, weight deacy loss is 7.51, training accuracy is 0.250000, time 56.139 samples/sec
+epoch 7, total_step 118560, total loss is 16.43 , inference loss is 8.92, weight deacy loss is 7.51, training accuracy is 0.375000, time 56.532 samples/sec
+epoch 7, total_step 118580, total loss is 17.96 , inference loss is 10.45, weight deacy loss is 7.51, training accuracy is 0.281250, time 57.123 samples/sec
+epoch 7, total_step 118600, total loss is 18.51 , inference loss is 11.00, weight deacy loss is 7.51, training accuracy is 0.312500, time 56.699 samples/sec
+epoch 7, total_step 118620, total loss is 15.62 , inference loss is 8.11, weight deacy loss is 7.51, training accuracy is 0.312500, time 56.553 samples/sec
+epoch 7, total_step 118640, total loss is 15.08 , inference loss is 7.57, weight deacy loss is 7.51, training accuracy is 0.375000, time 55.760 samples/sec
+epoch 7, total_step 118660, total loss is 17.17 , inference loss is 9.66, weight deacy loss is 7.51, training accuracy is 0.312500, time 56.268 samples/sec
+epoch 7, total_step 118680, total loss is 17.59 , inference loss is 10.08, weight deacy loss is 7.51, training accuracy is 0.343750, time 56.859 samples/sec
+epoch 7, total_step 118700, total loss is 18.87 , inference loss is 11.36, weight deacy loss is 7.51, training accuracy is 0.343750, time 56.879 samples/sec
+epoch 7, total_step 118720, total loss is 13.57 , inference loss is 6.07, weight deacy loss is 7.51, training accuracy is 0.500000, time 53.878 samples/sec
+epoch 7, total_step 118740, total loss is 15.12 , inference loss is 7.61, weight deacy loss is 7.51, training accuracy is 0.375000, time 57.554 samples/sec
+epoch 7, total_step 118760, total loss is 17.55 , inference loss is 10.04, weight deacy loss is 7.51, training accuracy is 0.218750, time 57.230 samples/sec
+epoch 7, total_step 118780, total loss is 13.60 , inference loss is 6.10, weight deacy loss is 7.51, training accuracy is 0.531250, time 56.579 samples/sec
+epoch 7, total_step 118800, total loss is 20.54 , inference loss is 13.04, weight deacy loss is 7.50, training accuracy is 0.312500, time 56.192 samples/sec
+epoch 7, total_step 118820, total loss is 14.80 , inference loss is 7.30, weight deacy loss is 7.50, training accuracy is 0.375000, time 56.892 samples/sec
+epoch 7, total_step 118840, total loss is 12.17 , inference loss is 4.67, weight deacy loss is 7.50, training accuracy is 0.531250, time 55.394 samples/sec
+epoch 7, total_step 118860, total loss is 18.12 , inference loss is 10.62, weight deacy loss is 7.50, training accuracy is 0.281250, time 55.293 samples/sec
+epoch 7, total_step 118880, total loss is 23.32 , inference loss is 15.82, weight deacy loss is 7.50, training accuracy is 0.156250, time 55.120 samples/sec
+epoch 7, total_step 118900, total loss is 13.73 , inference loss is 6.23, weight deacy loss is 7.50, training accuracy is 0.500000, time 55.048 samples/sec
+epoch 7, total_step 118920, total loss is 18.08 , inference loss is 10.58, weight deacy loss is 7.50, training accuracy is 0.343750, time 55.238 samples/sec
+epoch 7, total_step 118940, total loss is 16.60 , inference loss is 9.10, weight deacy loss is 7.50, training accuracy is 0.375000, time 56.109 samples/sec
+epoch 7, total_step 118960, total loss is 18.98 , inference loss is 11.48, weight deacy loss is 7.50, training accuracy is 0.218750, time 55.320 samples/sec
+epoch 7, total_step 118980, total loss is 15.33 , inference loss is 7.83, weight deacy loss is 7.50, training accuracy is 0.406250, time 55.583 samples/sec
+epoch 7, total_step 119000, total loss is 15.91 , inference loss is 8.42, weight deacy loss is 7.50, training accuracy is 0.375000, time 56.595 samples/sec
+epoch 7, total_step 119020, total loss is 12.53 , inference loss is 5.04, weight deacy loss is 7.50, training accuracy is 0.375000, time 57.309 samples/sec
+epoch 7, total_step 119040, total loss is 15.75 , inference loss is 8.25, weight deacy loss is 7.50, training accuracy is 0.406250, time 56.276 samples/sec
+epoch 7, total_step 119060, total loss is 21.37 , inference loss is 13.87, weight deacy loss is 7.50, training accuracy is 0.187500, time 55.628 samples/sec
+epoch 7, total_step 119080, total loss is 19.14 , inference loss is 11.65, weight deacy loss is 7.50, training accuracy is 0.250000, time 56.143 samples/sec
+epoch 7, total_step 119100, total loss is 17.46 , inference loss is 9.96, weight deacy loss is 7.50, training accuracy is 0.375000, time 56.148 samples/sec
+epoch 7, total_step 119120, total loss is 16.09 , inference loss is 8.59, weight deacy loss is 7.50, training accuracy is 0.406250, time 56.463 samples/sec
+epoch 7, total_step 119140, total loss is 16.33 , inference loss is 8.83, weight deacy loss is 7.50, training accuracy is 0.312500, time 56.250 samples/sec
+epoch 7, total_step 119160, total loss is 18.74 , inference loss is 11.24, weight deacy loss is 7.49, training accuracy is 0.375000, time 57.286 samples/sec
+epoch 7, total_step 119180, total loss is 17.43 , inference loss is 9.94, weight deacy loss is 7.49, training accuracy is 0.187500, time 56.172 samples/sec
+epoch 7, total_step 119200, total loss is 17.91 , inference loss is 10.42, weight deacy loss is 7.49, training accuracy is 0.250000, time 56.164 samples/sec
+epoch 7, total_step 119220, total loss is 19.90 , inference loss is 12.41, weight deacy loss is 7.49, training accuracy is 0.312500, time 57.077 samples/sec
+epoch 7, total_step 119240, total loss is 18.58 , inference loss is 11.09, weight deacy loss is 7.49, training accuracy is 0.343750, time 56.548 samples/sec
+epoch 7, total_step 119260, total loss is 18.17 , inference loss is 10.68, weight deacy loss is 7.49, training accuracy is 0.343750, time 56.828 samples/sec
+epoch 7, total_step 119280, total loss is 12.14 , inference loss is 4.65, weight deacy loss is 7.49, training accuracy is 0.500000, time 56.225 samples/sec
+epoch 7, total_step 119300, total loss is 18.87 , inference loss is 11.38, weight deacy loss is 7.49, training accuracy is 0.156250, time 56.707 samples/sec
+epoch 7, total_step 119320, total loss is 19.45 , inference loss is 11.96, weight deacy loss is 7.49, training accuracy is 0.500000, time 56.663 samples/sec
+epoch 7, total_step 119340, total loss is 14.18 , inference loss is 6.69, weight deacy loss is 7.49, training accuracy is 0.437500, time 57.074 samples/sec
+epoch 7, total_step 119360, total loss is 17.62 , inference loss is 10.13, weight deacy loss is 7.49, training accuracy is 0.281250, time 57.045 samples/sec
+epoch 7, total_step 119380, total loss is 17.59 , inference loss is 10.10, weight deacy loss is 7.49, training accuracy is 0.437500, time 56.617 samples/sec
+epoch 7, total_step 119400, total loss is 20.75 , inference loss is 13.26, weight deacy loss is 7.49, training accuracy is 0.281250, time 57.467 samples/sec
+epoch 7, total_step 119420, total loss is 14.20 , inference loss is 6.71, weight deacy loss is 7.49, training accuracy is 0.375000, time 56.275 samples/sec
+epoch 7, total_step 119440, total loss is 21.24 , inference loss is 13.76, weight deacy loss is 7.49, training accuracy is 0.250000, time 56.327 samples/sec
+epoch 7, total_step 119460, total loss is 21.44 , inference loss is 13.95, weight deacy loss is 7.49, training accuracy is 0.218750, time 55.440 samples/sec
+epoch 7, total_step 119480, total loss is 21.79 , inference loss is 14.30, weight deacy loss is 7.49, training accuracy is 0.187500, time 55.907 samples/sec
+epoch 7, total_step 119500, total loss is 17.20 , inference loss is 9.71, weight deacy loss is 7.49, training accuracy is 0.281250, time 56.004 samples/sec
+epoch 7, total_step 119520, total loss is 18.39 , inference loss is 10.90, weight deacy loss is 7.49, training accuracy is 0.250000, time 56.601 samples/sec
+epoch 7, total_step 119540, total loss is 18.52 , inference loss is 11.03, weight deacy loss is 7.48, training accuracy is 0.281250, time 56.260 samples/sec
+epoch 7, total_step 119560, total loss is 13.10 , inference loss is 5.62, weight deacy loss is 7.48, training accuracy is 0.468750, time 56.762 samples/sec
+epoch 7, total_step 119580, total loss is 23.70 , inference loss is 16.22, weight deacy loss is 7.48, training accuracy is 0.187500, time 56.500 samples/sec
+epoch 7, total_step 119600, total loss is 14.27 , inference loss is 6.79, weight deacy loss is 7.48, training accuracy is 0.343750, time 56.648 samples/sec
+epoch 7, total_step 119620, total loss is 15.87 , inference loss is 8.39, weight deacy loss is 7.48, training accuracy is 0.406250, time 55.118 samples/sec
+epoch 7, total_step 119640, total loss is 18.98 , inference loss is 11.50, weight deacy loss is 7.48, training accuracy is 0.281250, time 56.256 samples/sec
+epoch 7, total_step 119660, total loss is 18.56 , inference loss is 11.08, weight deacy loss is 7.48, training accuracy is 0.343750, time 55.799 samples/sec
+epoch 7, total_step 119680, total loss is 19.91 , inference loss is 12.43, weight deacy loss is 7.48, training accuracy is 0.375000, time 56.250 samples/sec
+epoch 7, total_step 119700, total loss is 15.92 , inference loss is 8.44, weight deacy loss is 7.48, training accuracy is 0.375000, time 56.037 samples/sec
+epoch 7, total_step 119720, total loss is 15.79 , inference loss is 8.31, weight deacy loss is 7.48, training accuracy is 0.250000, time 56.515 samples/sec
+epoch 7, total_step 119740, total loss is 16.56 , inference loss is 9.08, weight deacy loss is 7.48, training accuracy is 0.250000, time 56.516 samples/sec
+epoch 7, total_step 119760, total loss is 17.20 , inference loss is 9.72, weight deacy loss is 7.48, training accuracy is 0.218750, time 56.259 samples/sec
+epoch 7, total_step 119780, total loss is 17.54 , inference loss is 10.06, weight deacy loss is 7.48, training accuracy is 0.500000, time 56.198 samples/sec
+epoch 7, total_step 119800, total loss is 18.57 , inference loss is 11.09, weight deacy loss is 7.48, training accuracy is 0.343750, time 55.332 samples/sec
+epoch 7, total_step 119820, total loss is 16.41 , inference loss is 8.94, weight deacy loss is 7.48, training accuracy is 0.375000, time 55.646 samples/sec
+epoch 7, total_step 119840, total loss is 17.05 , inference loss is 9.57, weight deacy loss is 7.48, training accuracy is 0.375000, time 56.130 samples/sec
+epoch 7, total_step 119860, total loss is 18.10 , inference loss is 10.63, weight deacy loss is 7.48, training accuracy is 0.281250, time 56.101 samples/sec
+epoch 7, total_step 119880, total loss is 15.21 , inference loss is 7.74, weight deacy loss is 7.47, training accuracy is 0.281250, time 56.737 samples/sec
+epoch 7, total_step 119900, total loss is 21.36 , inference loss is 13.89, weight deacy loss is 7.47, training accuracy is 0.187500, time 56.927 samples/sec
+epoch 7, total_step 119920, total loss is 18.06 , inference loss is 10.59, weight deacy loss is 7.47, training accuracy is 0.312500, time 56.939 samples/sec
+epoch 7, total_step 119940, total loss is 18.38 , inference loss is 10.91, weight deacy loss is 7.47, training accuracy is 0.406250, time 56.554 samples/sec
+epoch 7, total_step 119960, total loss is 18.31 , inference loss is 10.84, weight deacy loss is 7.47, training accuracy is 0.250000, time 56.695 samples/sec
+epoch 7, total_step 119980, total loss is 21.76 , inference loss is 14.29, weight deacy loss is 7.47, training accuracy is 0.156250, time 56.673 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.07438400000005
+best_threshold_index 111 0.9846296296296296
+best_threshold_index 111 0.9829629629629629
+best_threshold_index 111 0.9842592592592593
+best_threshold_index 107 0.9844444444444445
+best_threshold_index 111 0.9833333333333333
+best_threshold_index 111 0.9831481481481481
+best_threshold_index 111 0.9837037037037037
+best_threshold_index 111 0.9833333333333333
+best_threshold_index 111 0.9824074074074074
+best_threshold_index 111 0.9829629629629629
+[lfw][120000]XNorm: 22.156794
+[lfw][120000]Accuracy-Flip: 0.98267+-0.00731
+testing verification..
+(14000, 512)
+infer time 85.22569999999996
+best_threshold_index 114 0.9868253968253968
+best_threshold_index 116 0.9857142857142858
+best_threshold_index 115 0.9853968253968254
+best_threshold_index 116 0.9852380952380952
+best_threshold_index 116 0.9863492063492063
+best_threshold_index 115 0.9852380952380952
+best_threshold_index 114 0.9858730158730159
+best_threshold_index 114 0.9847619047619047
+best_threshold_index 114 0.9853968253968254
+best_threshold_index 116 0.9865079365079366
+[cfp_ff][120000]XNorm: 21.639786
+[cfp_ff][120000]Accuracy-Flip: 0.98471+-0.00542
+testing verification..
+(14000, 512)
+infer time 85.15504400000007
+best_threshold_index 136 0.9228571428571428
+best_threshold_index 132 0.9242857142857143
+best_threshold_index 136 0.922063492063492
+best_threshold_index 136 0.9247619047619048
+best_threshold_index 136 0.9261904761904762
+best_threshold_index 136 0.9231746031746032
+best_threshold_index 136 0.9211111111111111
+best_threshold_index 136 0.922063492063492
+best_threshold_index 136 0.9238095238095239
+best_threshold_index 136 0.9211111111111111
+[cfp_fp][120000]XNorm: 21.302120
+[cfp_fp][120000]Accuracy-Flip: 0.92214+-0.01505
+testing verification..
+(12000, 512)
+infer time 73.44457799999998
+best_threshold_index 131 0.9044444444444445
+best_threshold_index 131 0.9016666666666666
+best_threshold_index 131 0.9038888888888889
+best_threshold_index 131 0.9055555555555556
+best_threshold_index 131 0.9087037037037037
+best_threshold_index 131 0.9031481481481481
+best_threshold_index 131 0.9037037037037037
+best_threshold_index 131 0.9096296296296297
+best_threshold_index 131 0.909074074074074
+best_threshold_index 131 0.9051851851851852
+[agedb_30][120000]XNorm: 22.055358
+[agedb_30][120000]Accuracy-Flip: 0.90550+-0.02336
+test accuracy is: 0.9826666666666668
+epoch 7, total_step 120000, total loss is 18.58 , inference loss is 11.11, weight deacy loss is 7.47, training accuracy is 0.218750, time 52.979 samples/sec
+epoch 7, total_step 120020, total loss is 15.99 , inference loss is 8.52, weight deacy loss is 7.47, training accuracy is 0.250000, time 56.277 samples/sec
+epoch 7, total_step 120040, total loss is 19.09 , inference loss is 11.62, weight deacy loss is 7.47, training accuracy is 0.125000, time 56.980 samples/sec
+epoch 7, total_step 120060, total loss is 18.62 , inference loss is 11.15, weight deacy loss is 7.47, training accuracy is 0.250000, time 56.117 samples/sec
+epoch 7, total_step 120080, total loss is 15.26 , inference loss is 7.79, weight deacy loss is 7.47, training accuracy is 0.312500, time 56.202 samples/sec
+epoch 7, total_step 120100, total loss is 19.20 , inference loss is 11.74, weight deacy loss is 7.47, training accuracy is 0.156250, time 56.632 samples/sec
+epoch 7, total_step 120120, total loss is 19.19 , inference loss is 11.72, weight deacy loss is 7.47, training accuracy is 0.218750, time 56.957 samples/sec
+epoch 7, total_step 120140, total loss is 20.10 , inference loss is 12.63, weight deacy loss is 7.47, training accuracy is 0.250000, time 55.221 samples/sec
+epoch 7, total_step 120160, total loss is 20.97 , inference loss is 13.51, weight deacy loss is 7.47, training accuracy is 0.156250, time 57.515 samples/sec
+epoch 7, total_step 120180, total loss is 17.98 , inference loss is 10.51, weight deacy loss is 7.47, training accuracy is 0.343750, time 56.884 samples/sec
+epoch 7, total_step 120200, total loss is 20.89 , inference loss is 13.42, weight deacy loss is 7.47, training accuracy is 0.218750, time 56.543 samples/sec
+epoch 7, total_step 120220, total loss is 19.35 , inference loss is 11.89, weight deacy loss is 7.47, training accuracy is 0.218750, time 57.041 samples/sec
+epoch 7, total_step 120240, total loss is 17.64 , inference loss is 10.18, weight deacy loss is 7.46, training accuracy is 0.375000, time 56.865 samples/sec
+epoch 7, total_step 120260, total loss is 18.91 , inference loss is 11.45, weight deacy loss is 7.46, training accuracy is 0.250000, time 56.321 samples/sec
+epoch 7, total_step 120280, total loss is 13.21 , inference loss is 5.75, weight deacy loss is 7.46, training accuracy is 0.312500, time 56.738 samples/sec
+epoch 7, total_step 120300, total loss is 17.96 , inference loss is 10.50, weight deacy loss is 7.46, training accuracy is 0.312500, time 56.325 samples/sec
+epoch 7, total_step 120320, total loss is 16.93 , inference loss is 9.47, weight deacy loss is 7.46, training accuracy is 0.312500, time 56.725 samples/sec
+epoch 7, total_step 120340, total loss is 17.65 , inference loss is 10.19, weight deacy loss is 7.46, training accuracy is 0.218750, time 56.419 samples/sec
+epoch 7, total_step 120360, total loss is 20.61 , inference loss is 13.15, weight deacy loss is 7.46, training accuracy is 0.187500, time 56.328 samples/sec
+epoch 7, total_step 120380, total loss is 18.04 , inference loss is 10.58, weight deacy loss is 7.46, training accuracy is 0.187500, time 57.284 samples/sec
+epoch 7, total_step 120400, total loss is 19.08 , inference loss is 11.62, weight deacy loss is 7.46, training accuracy is 0.218750, time 56.258 samples/sec
+epoch 7, total_step 120420, total loss is 20.28 , inference loss is 12.82, weight deacy loss is 7.46, training accuracy is 0.312500, time 56.475 samples/sec
+epoch 7, total_step 120440, total loss is 16.26 , inference loss is 8.80, weight deacy loss is 7.46, training accuracy is 0.250000, time 55.849 samples/sec
+epoch 7, total_step 120460, total loss is 19.94 , inference loss is 12.48, weight deacy loss is 7.46, training accuracy is 0.375000, time 56.218 samples/sec
+epoch 7, total_step 120480, total loss is 15.87 , inference loss is 8.42, weight deacy loss is 7.46, training accuracy is 0.437500, time 56.439 samples/sec
+epoch 7, total_step 120500, total loss is 17.01 , inference loss is 9.55, weight deacy loss is 7.46, training accuracy is 0.250000, time 56.376 samples/sec
+epoch 7, total_step 120520, total loss is 17.22 , inference loss is 9.77, weight deacy loss is 7.46, training accuracy is 0.343750, time 56.587 samples/sec
+epoch 7, total_step 120540, total loss is 18.59 , inference loss is 11.13, weight deacy loss is 7.46, training accuracy is 0.218750, time 56.352 samples/sec
+epoch 7, total_step 120560, total loss is 17.78 , inference loss is 10.33, weight deacy loss is 7.46, training accuracy is 0.343750, time 55.510 samples/sec
+epoch 7, total_step 120580, total loss is 21.79 , inference loss is 14.34, weight deacy loss is 7.46, training accuracy is 0.187500, time 56.925 samples/sec
+epoch 7, total_step 120600, total loss is 13.28 , inference loss is 5.83, weight deacy loss is 7.45, training accuracy is 0.406250, time 56.486 samples/sec
+epoch 7, total_step 120620, total loss is 16.41 , inference loss is 8.95, weight deacy loss is 7.45, training accuracy is 0.343750, time 56.467 samples/sec
+epoch 7, total_step 120640, total loss is 17.81 , inference loss is 10.35, weight deacy loss is 7.45, training accuracy is 0.375000, time 56.134 samples/sec
+epoch 7, total_step 120660, total loss is 16.82 , inference loss is 9.37, weight deacy loss is 7.45, training accuracy is 0.312500, time 56.969 samples/sec
+epoch 7, total_step 120680, total loss is 20.29 , inference loss is 12.84, weight deacy loss is 7.45, training accuracy is 0.218750, time 56.403 samples/sec
+epoch 7, total_step 120700, total loss is 13.98 , inference loss is 6.53, weight deacy loss is 7.45, training accuracy is 0.562500, time 56.439 samples/sec
+epoch 7, total_step 120720, total loss is 18.49 , inference loss is 11.03, weight deacy loss is 7.45, training accuracy is 0.375000, time 56.091 samples/sec
+epoch 7, total_step 120740, total loss is 18.02 , inference loss is 10.57, weight deacy loss is 7.45, training accuracy is 0.250000, time 56.037 samples/sec
+epoch 7, total_step 120760, total loss is 13.98 , inference loss is 6.53, weight deacy loss is 7.45, training accuracy is 0.437500, time 56.790 samples/sec
+epoch 7, total_step 120780, total loss is 18.03 , inference loss is 10.58, weight deacy loss is 7.45, training accuracy is 0.312500, time 56.936 samples/sec
+epoch 7, total_step 120800, total loss is 14.81 , inference loss is 7.36, weight deacy loss is 7.45, training accuracy is 0.375000, time 57.111 samples/sec
+epoch 7, total_step 120820, total loss is 16.82 , inference loss is 9.37, weight deacy loss is 7.45, training accuracy is 0.218750, time 56.596 samples/sec
+epoch 7, total_step 120840, total loss is 15.72 , inference loss is 8.27, weight deacy loss is 7.45, training accuracy is 0.281250, time 57.397 samples/sec
+epoch 7, total_step 120860, total loss is 16.63 , inference loss is 9.19, weight deacy loss is 7.45, training accuracy is 0.218750, time 56.774 samples/sec
+epoch 7, total_step 120880, total loss is 15.55 , inference loss is 8.10, weight deacy loss is 7.45, training accuracy is 0.406250, time 56.926 samples/sec
+epoch 7, total_step 120900, total loss is 17.59 , inference loss is 10.15, weight deacy loss is 7.45, training accuracy is 0.156250, time 56.008 samples/sec
+epoch 7, total_step 120920, total loss is 18.81 , inference loss is 11.36, weight deacy loss is 7.45, training accuracy is 0.187500, time 56.534 samples/sec
+epoch 7, total_step 120940, total loss is 23.39 , inference loss is 15.95, weight deacy loss is 7.44, training accuracy is 0.156250, time 56.207 samples/sec
+epoch 7, total_step 120960, total loss is 19.55 , inference loss is 12.11, weight deacy loss is 7.44, training accuracy is 0.312500, time 56.960 samples/sec
+epoch 7, total_step 120980, total loss is 17.71 , inference loss is 10.27, weight deacy loss is 7.44, training accuracy is 0.312500, time 56.170 samples/sec
+epoch 7, total_step 121000, total loss is 17.17 , inference loss is 9.73, weight deacy loss is 7.44, training accuracy is 0.343750, time 56.505 samples/sec
+epoch 7, total_step 121020, total loss is 20.02 , inference loss is 12.57, weight deacy loss is 7.44, training accuracy is 0.250000, time 56.158 samples/sec
+epoch 7, total_step 121040, total loss is 17.20 , inference loss is 9.75, weight deacy loss is 7.44, training accuracy is 0.218750, time 56.602 samples/sec
+epoch 7, total_step 121060, total loss is 17.17 , inference loss is 9.73, weight deacy loss is 7.44, training accuracy is 0.312500, time 55.623 samples/sec
+epoch 7, total_step 121080, total loss is 20.06 , inference loss is 12.62, weight deacy loss is 7.44, training accuracy is 0.187500, time 57.082 samples/sec
+epoch 7, total_step 121100, total loss is 14.92 , inference loss is 7.48, weight deacy loss is 7.44, training accuracy is 0.281250, time 56.161 samples/sec
+epoch 7, total_step 121120, total loss is 16.30 , inference loss is 8.86, weight deacy loss is 7.44, training accuracy is 0.375000, time 56.374 samples/sec
+epoch 7, total_step 121140, total loss is 19.90 , inference loss is 12.46, weight deacy loss is 7.44, training accuracy is 0.250000, time 56.416 samples/sec
+epoch 7, total_step 121160, total loss is 18.42 , inference loss is 10.98, weight deacy loss is 7.44, training accuracy is 0.375000, time 56.488 samples/sec
+epoch 7, total_step 121180, total loss is 16.62 , inference loss is 9.18, weight deacy loss is 7.44, training accuracy is 0.468750, time 56.240 samples/sec
+epoch 7, total_step 121200, total loss is 22.95 , inference loss is 15.51, weight deacy loss is 7.44, training accuracy is 0.218750, time 56.095 samples/sec
+epoch 7, total_step 121220, total loss is 16.97 , inference loss is 9.54, weight deacy loss is 7.44, training accuracy is 0.312500, time 55.964 samples/sec
+epoch 7, total_step 121240, total loss is 16.87 , inference loss is 9.44, weight deacy loss is 7.44, training accuracy is 0.343750, time 55.769 samples/sec
+epoch 7, total_step 121260, total loss is 18.37 , inference loss is 10.93, weight deacy loss is 7.44, training accuracy is 0.375000, time 55.467 samples/sec
+epoch 7, total_step 121280, total loss is 18.67 , inference loss is 11.23, weight deacy loss is 7.44, training accuracy is 0.156250, time 57.209 samples/sec
+epoch 7, total_step 121300, total loss is 15.97 , inference loss is 8.53, weight deacy loss is 7.44, training accuracy is 0.218750, time 56.389 samples/sec
+epoch 7, total_step 121320, total loss is 19.45 , inference loss is 12.01, weight deacy loss is 7.43, training accuracy is 0.218750, time 56.832 samples/sec
+epoch 7, total_step 121340, total loss is 18.99 , inference loss is 11.55, weight deacy loss is 7.43, training accuracy is 0.218750, time 56.365 samples/sec
+epoch 7, total_step 121360, total loss is 16.95 , inference loss is 9.52, weight deacy loss is 7.43, training accuracy is 0.343750, time 56.634 samples/sec
+epoch 7, total_step 121380, total loss is 22.62 , inference loss is 15.19, weight deacy loss is 7.43, training accuracy is 0.093750, time 56.608 samples/sec
+epoch 7, total_step 121400, total loss is 20.53 , inference loss is 13.10, weight deacy loss is 7.43, training accuracy is 0.250000, time 56.752 samples/sec
+epoch 7, total_step 121420, total loss is 18.45 , inference loss is 11.02, weight deacy loss is 7.43, training accuracy is 0.250000, time 56.590 samples/sec
+epoch 7, total_step 121440, total loss is 18.38 , inference loss is 10.95, weight deacy loss is 7.43, training accuracy is 0.281250, time 56.837 samples/sec
+epoch 7, total_step 121460, total loss is 14.89 , inference loss is 7.46, weight deacy loss is 7.43, training accuracy is 0.375000, time 56.643 samples/sec
+epoch 7, total_step 121480, total loss is 22.99 , inference loss is 15.56, weight deacy loss is 7.43, training accuracy is 0.187500, time 56.072 samples/sec
+epoch 7, total_step 121500, total loss is 16.53 , inference loss is 9.10, weight deacy loss is 7.43, training accuracy is 0.281250, time 56.232 samples/sec
+epoch 7, total_step 121520, total loss is 18.20 , inference loss is 10.77, weight deacy loss is 7.43, training accuracy is 0.281250, time 57.309 samples/sec
+epoch 7, total_step 121540, total loss is 19.99 , inference loss is 12.56, weight deacy loss is 7.43, training accuracy is 0.281250, time 56.464 samples/sec
+epoch 7, total_step 121560, total loss is 15.43 , inference loss is 8.00, weight deacy loss is 7.43, training accuracy is 0.406250, time 57.036 samples/sec
+epoch 7, total_step 121580, total loss is 18.68 , inference loss is 11.25, weight deacy loss is 7.43, training accuracy is 0.468750, time 56.612 samples/sec
+epoch 7, total_step 121600, total loss is 14.65 , inference loss is 7.22, weight deacy loss is 7.43, training accuracy is 0.437500, time 56.587 samples/sec
+epoch 7, total_step 121620, total loss is 26.11 , inference loss is 18.69, weight deacy loss is 7.43, training accuracy is 0.125000, time 56.402 samples/sec
+epoch 7, total_step 121640, total loss is 20.36 , inference loss is 12.94, weight deacy loss is 7.43, training accuracy is 0.218750, time 56.417 samples/sec
+epoch 7, total_step 121660, total loss is 19.39 , inference loss is 11.96, weight deacy loss is 7.43, training accuracy is 0.156250, time 55.734 samples/sec
+epoch 7, total_step 121680, total loss is 17.00 , inference loss is 9.57, weight deacy loss is 7.42, training accuracy is 0.281250, time 56.563 samples/sec
+epoch 7, total_step 121700, total loss is 19.93 , inference loss is 12.50, weight deacy loss is 7.42, training accuracy is 0.187500, time 56.791 samples/sec
+epoch 7, total_step 121720, total loss is 20.90 , inference loss is 13.48, weight deacy loss is 7.42, training accuracy is 0.250000, time 56.091 samples/sec
+epoch 7, total_step 121740, total loss is 17.74 , inference loss is 10.31, weight deacy loss is 7.42, training accuracy is 0.125000, time 55.763 samples/sec
+epoch 7, total_step 121760, total loss is 16.79 , inference loss is 9.37, weight deacy loss is 7.42, training accuracy is 0.312500, time 56.069 samples/sec
+epoch 7, total_step 121780, total loss is 17.06 , inference loss is 9.64, weight deacy loss is 7.42, training accuracy is 0.156250, time 56.354 samples/sec
+epoch 7, total_step 121800, total loss is 17.51 , inference loss is 10.08, weight deacy loss is 7.42, training accuracy is 0.343750, time 57.038 samples/sec
+epoch 7, total_step 121820, total loss is 17.63 , inference loss is 10.21, weight deacy loss is 7.42, training accuracy is 0.250000, time 56.368 samples/sec
+epoch 7, total_step 121840, total loss is 18.76 , inference loss is 11.34, weight deacy loss is 7.42, training accuracy is 0.250000, time 57.014 samples/sec
+epoch 7, total_step 121860, total loss is 17.67 , inference loss is 10.25, weight deacy loss is 7.42, training accuracy is 0.250000, time 57.011 samples/sec
+epoch 7, total_step 121880, total loss is 17.04 , inference loss is 9.62, weight deacy loss is 7.42, training accuracy is 0.250000, time 57.028 samples/sec
+epoch 7, total_step 121900, total loss is 15.37 , inference loss is 7.95, weight deacy loss is 7.42, training accuracy is 0.250000, time 56.620 samples/sec
+epoch 7, total_step 121920, total loss is 17.75 , inference loss is 10.33, weight deacy loss is 7.42, training accuracy is 0.187500, time 56.460 samples/sec
+epoch 7, total_step 121940, total loss is 17.63 , inference loss is 10.22, weight deacy loss is 7.42, training accuracy is 0.250000, time 56.645 samples/sec
+epoch 7, total_step 121960, total loss is 19.89 , inference loss is 12.47, weight deacy loss is 7.42, training accuracy is 0.250000, time 56.771 samples/sec
+epoch 7, total_step 121980, total loss is 15.47 , inference loss is 8.05, weight deacy loss is 7.42, training accuracy is 0.312500, time 56.521 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.79240499999996
+best_threshold_index 108 0.9868518518518519
+best_threshold_index 108 0.985
+best_threshold_index 108 0.9866666666666667
+best_threshold_index 108 0.9864814814814815
+best_threshold_index 108 0.987037037037037
+best_threshold_index 108 0.9848148148148148
+best_threshold_index 108 0.9861111111111112
+best_threshold_index 108 0.9861111111111112
+best_threshold_index 108 0.9844444444444445
+best_threshold_index 108 0.9848148148148148
+[lfw][122000]XNorm: 22.379583
+[lfw][122000]Accuracy-Flip: 0.98583+-0.00827
+testing verification..
+(14000, 512)
+infer time 85.428685
+best_threshold_index 121 0.986984126984127
+best_threshold_index 121 0.984920634920635
+best_threshold_index 124 0.9853968253968254
+best_threshold_index 124 0.9852380952380952
+best_threshold_index 124 0.9861904761904762
+best_threshold_index 121 0.984920634920635
+best_threshold_index 121 0.9863492063492063
+best_threshold_index 124 0.9855555555555555
+best_threshold_index 121 0.9847619047619047
+best_threshold_index 124 0.9863492063492063
+[cfp_ff][122000]XNorm: 22.392082
+[cfp_ff][122000]Accuracy-Flip: 0.98471+-0.00645
+testing verification..
+(14000, 512)
+infer time 86.21842099999998
+best_threshold_index 134 0.9168253968253969
+best_threshold_index 134 0.9182539682539682
+best_threshold_index 134 0.9149206349206349
+best_threshold_index 134 0.9193650793650794
+best_threshold_index 134 0.9207936507936508
+best_threshold_index 134 0.9144444444444444
+best_threshold_index 134 0.9142857142857143
+best_threshold_index 134 0.9161904761904762
+best_threshold_index 134 0.9165079365079365
+best_threshold_index 134 0.916984126984127
+[cfp_fp][122000]XNorm: 21.989961
+[cfp_fp][122000]Accuracy-Flip: 0.91686+-0.01811
+testing verification..
+(12000, 512)
+infer time 74.127473
+best_threshold_index 139 0.9068518518518518
+best_threshold_index 139 0.9068518518518518
+best_threshold_index 139 0.9077777777777778
+best_threshold_index 139 0.9087037037037037
+best_threshold_index 139 0.9122222222222223
+best_threshold_index 139 0.9064814814814814
+best_threshold_index 139 0.9087037037037037
+best_threshold_index 139 0.9144444444444444
+best_threshold_index 139 0.9103703703703704
+best_threshold_index 139 0.9109259259259259
+[agedb_30][122000]XNorm: 22.764437
+[agedb_30][122000]Accuracy-Flip: 0.90933+-0.02236
+test accuracy is: 0.9858333333333332
+epoch 7, total_step 122000, total loss is 19.79 , inference loss is 12.38, weight deacy loss is 7.42, training accuracy is 0.156250, time 54.122 samples/sec
+epoch 7, total_step 122020, total loss is 18.39 , inference loss is 10.97, weight deacy loss is 7.42, training accuracy is 0.125000, time 55.935 samples/sec
+epoch 7, total_step 122040, total loss is 18.27 , inference loss is 10.85, weight deacy loss is 7.41, training accuracy is 0.250000, time 56.590 samples/sec
+epoch 7, total_step 122060, total loss is 19.84 , inference loss is 12.42, weight deacy loss is 7.41, training accuracy is 0.125000, time 56.229 samples/sec
+epoch 7, total_step 122080, total loss is 15.62 , inference loss is 8.20, weight deacy loss is 7.41, training accuracy is 0.375000, time 56.505 samples/sec
+epoch 7, total_step 122100, total loss is 17.39 , inference loss is 9.97, weight deacy loss is 7.41, training accuracy is 0.250000, time 56.687 samples/sec
+epoch 7, total_step 122120, total loss is 21.32 , inference loss is 13.91, weight deacy loss is 7.41, training accuracy is 0.125000, time 56.823 samples/sec
+epoch 7, total_step 122140, total loss is 15.52 , inference loss is 8.11, weight deacy loss is 7.41, training accuracy is 0.437500, time 56.407 samples/sec
+epoch 7, total_step 122160, total loss is 20.24 , inference loss is 12.83, weight deacy loss is 7.41, training accuracy is 0.156250, time 56.708 samples/sec
+epoch 7, total_step 122180, total loss is 18.18 , inference loss is 10.77, weight deacy loss is 7.41, training accuracy is 0.250000, time 56.430 samples/sec
+epoch 7, total_step 122200, total loss is 19.29 , inference loss is 11.88, weight deacy loss is 7.41, training accuracy is 0.125000, time 57.258 samples/sec
+epoch 7, total_step 122220, total loss is 18.17 , inference loss is 10.76, weight deacy loss is 7.41, training accuracy is 0.187500, time 56.676 samples/sec
+epoch 7, total_step 122240, total loss is 19.11 , inference loss is 11.70, weight deacy loss is 7.41, training accuracy is 0.250000, time 56.254 samples/sec
+epoch 7, total_step 122260, total loss is 16.41 , inference loss is 9.00, weight deacy loss is 7.41, training accuracy is 0.375000, time 56.242 samples/sec
+epoch 7, total_step 122280, total loss is 18.94 , inference loss is 11.53, weight deacy loss is 7.41, training accuracy is 0.218750, time 56.169 samples/sec
+epoch 7, total_step 122300, total loss is 18.76 , inference loss is 11.35, weight deacy loss is 7.41, training accuracy is 0.343750, time 56.079 samples/sec
+epoch 7, total_step 122320, total loss is 20.75 , inference loss is 13.34, weight deacy loss is 7.41, training accuracy is 0.218750, time 56.598 samples/sec
+epoch 7, total_step 122340, total loss is 16.67 , inference loss is 9.26, weight deacy loss is 7.41, training accuracy is 0.125000, time 56.480 samples/sec
+epoch 7, total_step 122360, total loss is 17.98 , inference loss is 10.58, weight deacy loss is 7.41, training accuracy is 0.250000, time 56.519 samples/sec
+epoch 7, total_step 122380, total loss is 18.63 , inference loss is 11.22, weight deacy loss is 7.41, training accuracy is 0.281250, time 55.250 samples/sec
+epoch 7, total_step 122400, total loss is 20.77 , inference loss is 13.36, weight deacy loss is 7.41, training accuracy is 0.281250, time 56.150 samples/sec
+epoch 7, total_step 122420, total loss is 18.87 , inference loss is 11.47, weight deacy loss is 7.40, training accuracy is 0.312500, time 56.243 samples/sec
+epoch 7, total_step 122440, total loss is 17.35 , inference loss is 9.94, weight deacy loss is 7.40, training accuracy is 0.187500, time 56.343 samples/sec
+epoch 7, total_step 122460, total loss is 17.57 , inference loss is 10.17, weight deacy loss is 7.40, training accuracy is 0.312500, time 56.616 samples/sec
+epoch 7, total_step 122480, total loss is 15.60 , inference loss is 8.20, weight deacy loss is 7.40, training accuracy is 0.343750, time 56.859 samples/sec
+epoch 7, total_step 122500, total loss is 13.88 , inference loss is 6.48, weight deacy loss is 7.40, training accuracy is 0.437500, time 56.694 samples/sec
+epoch 7, total_step 122520, total loss is 20.24 , inference loss is 12.84, weight deacy loss is 7.40, training accuracy is 0.218750, time 56.497 samples/sec
+epoch 7, total_step 122540, total loss is 19.76 , inference loss is 12.36, weight deacy loss is 7.40, training accuracy is 0.218750, time 52.758 samples/sec
+epoch 7, total_step 122560, total loss is 15.22 , inference loss is 7.82, weight deacy loss is 7.40, training accuracy is 0.406250, time 56.807 samples/sec
+epoch 7, total_step 122580, total loss is 15.62 , inference loss is 8.22, weight deacy loss is 7.40, training accuracy is 0.343750, time 56.071 samples/sec
+epoch 7, total_step 122600, total loss is 19.02 , inference loss is 11.62, weight deacy loss is 7.40, training accuracy is 0.312500, time 57.242 samples/sec
+epoch 7, total_step 122620, total loss is 19.72 , inference loss is 12.32, weight deacy loss is 7.40, training accuracy is 0.375000, time 56.194 samples/sec
+epoch 7, total_step 122640, total loss is 19.96 , inference loss is 12.56, weight deacy loss is 7.40, training accuracy is 0.250000, time 56.929 samples/sec
+
+
+
+
+End of epoch 7
+epoch 8, total_step 122660, total loss is 26.56 , inference loss is 19.17, weight deacy loss is 7.40, training accuracy is 0.000000, time 56.491 samples/sec
+epoch 8, total_step 122680, total loss is 14.44 , inference loss is 7.04, weight deacy loss is 7.40, training accuracy is 0.406250, time 56.063 samples/sec
+epoch 8, total_step 122700, total loss is 14.00 , inference loss is 6.60, weight deacy loss is 7.40, training accuracy is 0.468750, time 55.734 samples/sec
+epoch 8, total_step 122720, total loss is 12.58 , inference loss is 5.18, weight deacy loss is 7.40, training accuracy is 0.562500, time 56.505 samples/sec
+epoch 8, total_step 122740, total loss is 14.97 , inference loss is 7.57, weight deacy loss is 7.40, training accuracy is 0.593750, time 57.102 samples/sec
+epoch 8, total_step 122760, total loss is 11.52 , inference loss is 4.12, weight deacy loss is 7.40, training accuracy is 0.562500, time 56.505 samples/sec
+epoch 8, total_step 122780, total loss is 14.33 , inference loss is 6.94, weight deacy loss is 7.39, training accuracy is 0.531250, time 55.877 samples/sec
+epoch 8, total_step 122800, total loss is 17.41 , inference loss is 10.02, weight deacy loss is 7.39, training accuracy is 0.343750, time 56.254 samples/sec
+epoch 8, total_step 122820, total loss is 12.36 , inference loss is 4.97, weight deacy loss is 7.39, training accuracy is 0.562500, time 56.946 samples/sec
+epoch 8, total_step 122840, total loss is 15.19 , inference loss is 7.80, weight deacy loss is 7.39, training accuracy is 0.593750, time 57.109 samples/sec
+epoch 8, total_step 122860, total loss is 15.15 , inference loss is 7.75, weight deacy loss is 7.39, training accuracy is 0.406250, time 56.577 samples/sec
+epoch 8, total_step 122880, total loss is 15.83 , inference loss is 8.44, weight deacy loss is 7.39, training accuracy is 0.375000, time 56.905 samples/sec
+epoch 8, total_step 122900, total loss is 14.34 , inference loss is 6.95, weight deacy loss is 7.39, training accuracy is 0.531250, time 56.549 samples/sec
+epoch 8, total_step 122920, total loss is 15.15 , inference loss is 7.76, weight deacy loss is 7.39, training accuracy is 0.531250, time 55.871 samples/sec
+epoch 8, total_step 122940, total loss is 12.71 , inference loss is 5.32, weight deacy loss is 7.39, training accuracy is 0.593750, time 55.366 samples/sec
+epoch 8, total_step 122960, total loss is 12.74 , inference loss is 5.35, weight deacy loss is 7.39, training accuracy is 0.562500, time 57.010 samples/sec
+epoch 8, total_step 122980, total loss is 14.38 , inference loss is 6.99, weight deacy loss is 7.39, training accuracy is 0.593750, time 55.647 samples/sec
+epoch 8, total_step 123000, total loss is 10.86 , inference loss is 3.47, weight deacy loss is 7.39, training accuracy is 0.656250, time 56.118 samples/sec
+epoch 8, total_step 123020, total loss is 15.31 , inference loss is 7.93, weight deacy loss is 7.39, training accuracy is 0.437500, time 56.563 samples/sec
+epoch 8, total_step 123040, total loss is 15.09 , inference loss is 7.71, weight deacy loss is 7.39, training accuracy is 0.531250, time 55.581 samples/sec
+epoch 8, total_step 123060, total loss is 13.60 , inference loss is 6.21, weight deacy loss is 7.39, training accuracy is 0.531250, time 55.738 samples/sec
+epoch 8, total_step 123080, total loss is 15.91 , inference loss is 8.52, weight deacy loss is 7.39, training accuracy is 0.468750, time 56.887 samples/sec
+epoch 8, total_step 123100, total loss is 12.72 , inference loss is 5.34, weight deacy loss is 7.39, training accuracy is 0.562500, time 56.462 samples/sec
+epoch 8, total_step 123120, total loss is 14.07 , inference loss is 6.69, weight deacy loss is 7.39, training accuracy is 0.625000, time 57.112 samples/sec
+epoch 8, total_step 123140, total loss is 12.61 , inference loss is 5.22, weight deacy loss is 7.39, training accuracy is 0.625000, time 56.346 samples/sec
+epoch 8, total_step 123160, total loss is 14.84 , inference loss is 7.45, weight deacy loss is 7.39, training accuracy is 0.562500, time 56.802 samples/sec
+epoch 8, total_step 123180, total loss is 14.22 , inference loss is 6.83, weight deacy loss is 7.39, training accuracy is 0.562500, time 56.218 samples/sec
+epoch 8, total_step 123200, total loss is 16.72 , inference loss is 9.34, weight deacy loss is 7.38, training accuracy is 0.531250, time 57.236 samples/sec
+epoch 8, total_step 123220, total loss is 16.65 , inference loss is 9.27, weight deacy loss is 7.38, training accuracy is 0.656250, time 57.092 samples/sec
+epoch 8, total_step 123240, total loss is 15.16 , inference loss is 7.78, weight deacy loss is 7.38, training accuracy is 0.531250, time 56.621 samples/sec
+epoch 8, total_step 123260, total loss is 17.46 , inference loss is 10.08, weight deacy loss is 7.38, training accuracy is 0.531250, time 57.751 samples/sec
+epoch 8, total_step 123280, total loss is 19.14 , inference loss is 11.75, weight deacy loss is 7.38, training accuracy is 0.343750, time 56.970 samples/sec
+epoch 8, total_step 123300, total loss is 16.07 , inference loss is 8.69, weight deacy loss is 7.38, training accuracy is 0.500000, time 56.891 samples/sec
+epoch 8, total_step 123320, total loss is 13.97 , inference loss is 6.58, weight deacy loss is 7.38, training accuracy is 0.656250, time 56.466 samples/sec
+epoch 8, total_step 123340, total loss is 13.81 , inference loss is 6.42, weight deacy loss is 7.38, training accuracy is 0.625000, time 56.482 samples/sec
+epoch 8, total_step 123360, total loss is 14.16 , inference loss is 6.78, weight deacy loss is 7.38, training accuracy is 0.593750, time 56.538 samples/sec
+epoch 8, total_step 123380, total loss is 19.26 , inference loss is 11.88, weight deacy loss is 7.38, training accuracy is 0.468750, time 56.375 samples/sec
+epoch 8, total_step 123400, total loss is 13.68 , inference loss is 6.30, weight deacy loss is 7.38, training accuracy is 0.625000, time 55.890 samples/sec
+epoch 8, total_step 123420, total loss is 14.69 , inference loss is 7.31, weight deacy loss is 7.38, training accuracy is 0.437500, time 56.306 samples/sec
+epoch 8, total_step 123440, total loss is 14.52 , inference loss is 7.14, weight deacy loss is 7.38, training accuracy is 0.437500, time 56.663 samples/sec
+epoch 8, total_step 123460, total loss is 13.09 , inference loss is 5.71, weight deacy loss is 7.38, training accuracy is 0.562500, time 55.419 samples/sec
+epoch 8, total_step 123480, total loss is 17.18 , inference loss is 9.80, weight deacy loss is 7.38, training accuracy is 0.437500, time 56.676 samples/sec
+epoch 8, total_step 123500, total loss is 14.26 , inference loss is 6.88, weight deacy loss is 7.38, training accuracy is 0.500000, time 56.243 samples/sec
+epoch 8, total_step 123520, total loss is 15.17 , inference loss is 7.79, weight deacy loss is 7.38, training accuracy is 0.562500, time 56.956 samples/sec
+epoch 8, total_step 123540, total loss is 15.39 , inference loss is 8.01, weight deacy loss is 7.38, training accuracy is 0.375000, time 56.914 samples/sec
+epoch 8, total_step 123560, total loss is 15.53 , inference loss is 8.15, weight deacy loss is 7.38, training accuracy is 0.406250, time 57.021 samples/sec
+epoch 8, total_step 123580, total loss is 12.58 , inference loss is 5.20, weight deacy loss is 7.38, training accuracy is 0.562500, time 56.967 samples/sec
+epoch 8, total_step 123600, total loss is 18.29 , inference loss is 10.91, weight deacy loss is 7.38, training accuracy is 0.343750, time 56.152 samples/sec
+epoch 8, total_step 123620, total loss is 13.72 , inference loss is 6.34, weight deacy loss is 7.38, training accuracy is 0.562500, time 56.723 samples/sec
+epoch 8, total_step 123640, total loss is 17.87 , inference loss is 10.49, weight deacy loss is 7.38, training accuracy is 0.312500, time 56.048 samples/sec
+epoch 8, total_step 123660, total loss is 15.83 , inference loss is 8.45, weight deacy loss is 7.38, training accuracy is 0.468750, time 56.546 samples/sec
+epoch 8, total_step 123680, total loss is 14.65 , inference loss is 7.27, weight deacy loss is 7.38, training accuracy is 0.593750, time 56.870 samples/sec
+epoch 8, total_step 123700, total loss is 10.98 , inference loss is 3.61, weight deacy loss is 7.38, training accuracy is 0.500000, time 56.524 samples/sec
+epoch 8, total_step 123720, total loss is 16.27 , inference loss is 8.89, weight deacy loss is 7.38, training accuracy is 0.593750, time 57.386 samples/sec
+epoch 8, total_step 123740, total loss is 15.17 , inference loss is 7.79, weight deacy loss is 7.38, training accuracy is 0.437500, time 54.204 samples/sec
+epoch 8, total_step 123760, total loss is 16.04 , inference loss is 8.66, weight deacy loss is 7.38, training accuracy is 0.437500, time 56.143 samples/sec
+epoch 8, total_step 123780, total loss is 18.98 , inference loss is 11.60, weight deacy loss is 7.38, training accuracy is 0.343750, time 54.524 samples/sec
+epoch 8, total_step 123800, total loss is 19.32 , inference loss is 11.95, weight deacy loss is 7.38, training accuracy is 0.312500, time 56.727 samples/sec
+epoch 8, total_step 123820, total loss is 17.12 , inference loss is 9.74, weight deacy loss is 7.38, training accuracy is 0.406250, time 56.412 samples/sec
+epoch 8, total_step 123840, total loss is 18.22 , inference loss is 10.84, weight deacy loss is 7.37, training accuracy is 0.468750, time 55.662 samples/sec
+epoch 8, total_step 123860, total loss is 15.87 , inference loss is 8.50, weight deacy loss is 7.37, training accuracy is 0.531250, time 56.926 samples/sec
+epoch 8, total_step 123880, total loss is 14.40 , inference loss is 7.03, weight deacy loss is 7.37, training accuracy is 0.468750, time 56.732 samples/sec
+epoch 8, total_step 123900, total loss is 12.25 , inference loss is 4.88, weight deacy loss is 7.37, training accuracy is 0.562500, time 56.799 samples/sec
+epoch 8, total_step 123920, total loss is 15.28 , inference loss is 7.91, weight deacy loss is 7.37, training accuracy is 0.437500, time 56.533 samples/sec
+epoch 8, total_step 123940, total loss is 16.31 , inference loss is 8.93, weight deacy loss is 7.37, training accuracy is 0.468750, time 56.434 samples/sec
+epoch 8, total_step 123960, total loss is 13.22 , inference loss is 5.84, weight deacy loss is 7.37, training accuracy is 0.531250, time 55.799 samples/sec
+epoch 8, total_step 123980, total loss is 18.20 , inference loss is 10.82, weight deacy loss is 7.37, training accuracy is 0.406250, time 55.683 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.94855200000002
+best_threshold_index 116 0.9868518518518519
+best_threshold_index 116 0.9857407407407407
+best_threshold_index 114 0.9861111111111112
+best_threshold_index 116 0.9868518518518519
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 116 0.9851851851851852
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 116 0.9862962962962963
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 116 0.9851851851851852
+[lfw][124000]XNorm: 22.973721
+[lfw][124000]Accuracy-Flip: 0.98550+-0.00522
+testing verification..
+(14000, 512)
+infer time 86.42040099999998
+best_threshold_index 112 0.9831746031746031
+best_threshold_index 116 0.9825396825396825
+best_threshold_index 116 0.9825396825396825
+best_threshold_index 112 0.9822222222222222
+best_threshold_index 112 0.9820634920634921
+best_threshold_index 112 0.9812698412698413
+best_threshold_index 112 0.9826984126984127
+best_threshold_index 112 0.9822222222222222
+best_threshold_index 116 0.9817460317460317
+best_threshold_index 116 0.9831746031746031
+[cfp_ff][124000]XNorm: 22.858721
+[cfp_ff][124000]Accuracy-Flip: 0.98157+-0.00509
+testing verification..
+(14000, 512)
+infer time 86.34043600000003
+best_threshold_index 134 0.9065079365079365
+best_threshold_index 134 0.9063492063492063
+best_threshold_index 134 0.903968253968254
+best_threshold_index 134 0.9074603174603174
+best_threshold_index 137 0.9085714285714286
+best_threshold_index 134 0.9047619047619048
+best_threshold_index 134 0.9038095238095238
+best_threshold_index 134 0.9046031746031746
+best_threshold_index 134 0.9057142857142857
+best_threshold_index 134 0.9073015873015873
+[cfp_fp][124000]XNorm: 23.657879
+[cfp_fp][124000]Accuracy-Flip: 0.90500+-0.01450
+testing verification..
+(12000, 512)
+infer time 73.93691200000004
+best_threshold_index 138 0.8998148148148148
+best_threshold_index 141 0.8975925925925926
+best_threshold_index 141 0.8996296296296297
+best_threshold_index 141 0.9
+best_threshold_index 140 0.905
+best_threshold_index 141 0.9
+best_threshold_index 140 0.9009259259259259
+best_threshold_index 140 0.9053703703703704
+best_threshold_index 138 0.9025925925925926
+best_threshold_index 141 0.9014814814814814
+[agedb_30][124000]XNorm: 23.655041
+[agedb_30][124000]Accuracy-Flip: 0.89933+-0.02151
+test accuracy is: 0.9855
+epoch 8, total_step 124000, total loss is 15.83 , inference loss is 8.46, weight deacy loss is 7.37, training accuracy is 0.375000, time 52.918 samples/sec
+epoch 8, total_step 124020, total loss is 17.57 , inference loss is 10.20, weight deacy loss is 7.37, training accuracy is 0.375000, time 57.087 samples/sec
+epoch 8, total_step 124040, total loss is 14.78 , inference loss is 7.41, weight deacy loss is 7.37, training accuracy is 0.437500, time 55.808 samples/sec
+epoch 8, total_step 124060, total loss is 15.92 , inference loss is 8.54, weight deacy loss is 7.37, training accuracy is 0.500000, time 55.389 samples/sec
+epoch 8, total_step 124080, total loss is 15.91 , inference loss is 8.54, weight deacy loss is 7.37, training accuracy is 0.468750, time 56.859 samples/sec
+epoch 8, total_step 124100, total loss is 15.08 , inference loss is 7.71, weight deacy loss is 7.37, training accuracy is 0.500000, time 56.134 samples/sec
+epoch 8, total_step 124120, total loss is 12.99 , inference loss is 5.62, weight deacy loss is 7.37, training accuracy is 0.656250, time 56.405 samples/sec
+epoch 8, total_step 124140, total loss is 16.31 , inference loss is 8.94, weight deacy loss is 7.37, training accuracy is 0.468750, time 56.896 samples/sec
+epoch 8, total_step 124160, total loss is 14.45 , inference loss is 7.08, weight deacy loss is 7.37, training accuracy is 0.593750, time 56.829 samples/sec
+epoch 8, total_step 124180, total loss is 12.94 , inference loss is 5.57, weight deacy loss is 7.37, training accuracy is 0.593750, time 56.724 samples/sec
+epoch 8, total_step 124200, total loss is 14.12 , inference loss is 6.75, weight deacy loss is 7.37, training accuracy is 0.500000, time 56.455 samples/sec
+epoch 8, total_step 124220, total loss is 16.74 , inference loss is 9.37, weight deacy loss is 7.37, training accuracy is 0.312500, time 56.091 samples/sec
+epoch 8, total_step 124240, total loss is 14.28 , inference loss is 6.91, weight deacy loss is 7.37, training accuracy is 0.531250, time 56.384 samples/sec
+epoch 8, total_step 124260, total loss is 14.65 , inference loss is 7.28, weight deacy loss is 7.37, training accuracy is 0.343750, time 56.811 samples/sec
+epoch 8, total_step 124280, total loss is 12.78 , inference loss is 5.41, weight deacy loss is 7.37, training accuracy is 0.437500, time 56.376 samples/sec
+epoch 8, total_step 124300, total loss is 15.33 , inference loss is 7.96, weight deacy loss is 7.37, training accuracy is 0.343750, time 55.783 samples/sec
+epoch 8, total_step 124320, total loss is 17.30 , inference loss is 9.94, weight deacy loss is 7.37, training accuracy is 0.343750, time 55.929 samples/sec
+epoch 8, total_step 124340, total loss is 17.33 , inference loss is 9.96, weight deacy loss is 7.37, training accuracy is 0.437500, time 55.195 samples/sec
+epoch 8, total_step 124360, total loss is 12.53 , inference loss is 5.16, weight deacy loss is 7.37, training accuracy is 0.531250, time 56.630 samples/sec
+epoch 8, total_step 124380, total loss is 17.62 , inference loss is 10.26, weight deacy loss is 7.37, training accuracy is 0.500000, time 56.801 samples/sec
+epoch 8, total_step 124400, total loss is 13.83 , inference loss is 6.46, weight deacy loss is 7.37, training accuracy is 0.343750, time 56.666 samples/sec
+epoch 8, total_step 124420, total loss is 14.88 , inference loss is 7.51, weight deacy loss is 7.37, training accuracy is 0.406250, time 56.325 samples/sec
+epoch 8, total_step 124440, total loss is 15.19 , inference loss is 7.82, weight deacy loss is 7.37, training accuracy is 0.375000, time 56.547 samples/sec
+epoch 8, total_step 124460, total loss is 22.15 , inference loss is 14.79, weight deacy loss is 7.36, training accuracy is 0.343750, time 56.919 samples/sec
+epoch 8, total_step 124480, total loss is 13.86 , inference loss is 6.50, weight deacy loss is 7.36, training accuracy is 0.437500, time 56.590 samples/sec
+epoch 8, total_step 124500, total loss is 16.65 , inference loss is 9.28, weight deacy loss is 7.36, training accuracy is 0.531250, time 54.758 samples/sec
+epoch 8, total_step 124520, total loss is 13.96 , inference loss is 6.59, weight deacy loss is 7.36, training accuracy is 0.437500, time 53.456 samples/sec
+epoch 8, total_step 124540, total loss is 14.24 , inference loss is 6.88, weight deacy loss is 7.36, training accuracy is 0.406250, time 56.254 samples/sec
+epoch 8, total_step 124560, total loss is 17.88 , inference loss is 10.52, weight deacy loss is 7.36, training accuracy is 0.343750, time 56.908 samples/sec
+epoch 8, total_step 124580, total loss is 14.84 , inference loss is 7.48, weight deacy loss is 7.36, training accuracy is 0.468750, time 57.307 samples/sec
+epoch 8, total_step 124600, total loss is 18.54 , inference loss is 11.18, weight deacy loss is 7.36, training accuracy is 0.468750, time 56.689 samples/sec
+epoch 8, total_step 124620, total loss is 16.52 , inference loss is 9.16, weight deacy loss is 7.36, training accuracy is 0.468750, time 55.812 samples/sec
+epoch 8, total_step 124640, total loss is 12.07 , inference loss is 4.71, weight deacy loss is 7.36, training accuracy is 0.562500, time 56.533 samples/sec
+epoch 8, total_step 124660, total loss is 17.07 , inference loss is 9.71, weight deacy loss is 7.36, training accuracy is 0.406250, time 56.045 samples/sec
+epoch 8, total_step 124680, total loss is 14.16 , inference loss is 6.80, weight deacy loss is 7.36, training accuracy is 0.437500, time 55.939 samples/sec
+epoch 8, total_step 124700, total loss is 12.18 , inference loss is 4.82, weight deacy loss is 7.36, training accuracy is 0.656250, time 56.445 samples/sec
+epoch 8, total_step 124720, total loss is 11.91 , inference loss is 4.56, weight deacy loss is 7.36, training accuracy is 0.531250, time 55.707 samples/sec
+epoch 8, total_step 124740, total loss is 13.39 , inference loss is 6.04, weight deacy loss is 7.36, training accuracy is 0.500000, time 57.362 samples/sec
+epoch 8, total_step 124760, total loss is 15.92 , inference loss is 8.56, weight deacy loss is 7.36, training accuracy is 0.343750, time 57.007 samples/sec
+epoch 8, total_step 124780, total loss is 15.16 , inference loss is 7.80, weight deacy loss is 7.36, training accuracy is 0.375000, time 57.070 samples/sec
+epoch 8, total_step 124800, total loss is 12.88 , inference loss is 5.53, weight deacy loss is 7.36, training accuracy is 0.593750, time 55.740 samples/sec
+epoch 8, total_step 124820, total loss is 11.53 , inference loss is 4.17, weight deacy loss is 7.36, training accuracy is 0.500000, time 56.637 samples/sec
+epoch 8, total_step 124840, total loss is 15.34 , inference loss is 7.98, weight deacy loss is 7.36, training accuracy is 0.500000, time 56.715 samples/sec
+epoch 8, total_step 124860, total loss is 11.41 , inference loss is 4.06, weight deacy loss is 7.36, training accuracy is 0.437500, time 56.867 samples/sec
+epoch 8, total_step 124880, total loss is 18.59 , inference loss is 11.24, weight deacy loss is 7.35, training accuracy is 0.437500, time 56.400 samples/sec
+epoch 8, total_step 124900, total loss is 14.24 , inference loss is 6.88, weight deacy loss is 7.35, training accuracy is 0.375000, time 56.162 samples/sec
+epoch 8, total_step 124920, total loss is 16.03 , inference loss is 8.68, weight deacy loss is 7.35, training accuracy is 0.468750, time 57.528 samples/sec
+epoch 8, total_step 124940, total loss is 14.64 , inference loss is 7.29, weight deacy loss is 7.35, training accuracy is 0.500000, time 56.708 samples/sec
+epoch 8, total_step 124960, total loss is 13.90 , inference loss is 6.55, weight deacy loss is 7.35, training accuracy is 0.468750, time 56.345 samples/sec
+epoch 8, total_step 124980, total loss is 16.99 , inference loss is 9.63, weight deacy loss is 7.35, training accuracy is 0.406250, time 56.517 samples/sec
+epoch 8, total_step 125000, total loss is 15.25 , inference loss is 7.90, weight deacy loss is 7.35, training accuracy is 0.406250, time 55.633 samples/sec
+epoch 8, total_step 125020, total loss is 12.22 , inference loss is 4.87, weight deacy loss is 7.35, training accuracy is 0.656250, time 56.740 samples/sec
+epoch 8, total_step 125040, total loss is 13.39 , inference loss is 6.04, weight deacy loss is 7.35, training accuracy is 0.625000, time 55.477 samples/sec
+epoch 8, total_step 125060, total loss is 14.18 , inference loss is 6.83, weight deacy loss is 7.35, training accuracy is 0.500000, time 56.525 samples/sec
+epoch 8, total_step 125080, total loss is 13.05 , inference loss is 5.70, weight deacy loss is 7.35, training accuracy is 0.531250, time 56.984 samples/sec
+epoch 8, total_step 125100, total loss is 17.59 , inference loss is 10.25, weight deacy loss is 7.35, training accuracy is 0.406250, time 55.813 samples/sec
+epoch 8, total_step 125120, total loss is 13.41 , inference loss is 6.07, weight deacy loss is 7.35, training accuracy is 0.656250, time 56.263 samples/sec
+epoch 8, total_step 125140, total loss is 11.58 , inference loss is 4.23, weight deacy loss is 7.35, training accuracy is 0.531250, time 56.897 samples/sec
+epoch 8, total_step 125160, total loss is 14.86 , inference loss is 7.52, weight deacy loss is 7.35, training accuracy is 0.437500, time 56.093 samples/sec
+epoch 8, total_step 125180, total loss is 10.86 , inference loss is 3.51, weight deacy loss is 7.35, training accuracy is 0.593750, time 56.296 samples/sec
+epoch 8, total_step 125200, total loss is 18.74 , inference loss is 11.40, weight deacy loss is 7.35, training accuracy is 0.281250, time 56.119 samples/sec
+epoch 8, total_step 125220, total loss is 14.75 , inference loss is 7.41, weight deacy loss is 7.35, training accuracy is 0.468750, time 56.661 samples/sec
+epoch 8, total_step 125240, total loss is 17.45 , inference loss is 10.11, weight deacy loss is 7.35, training accuracy is 0.406250, time 55.724 samples/sec
+epoch 8, total_step 125260, total loss is 13.67 , inference loss is 6.33, weight deacy loss is 7.35, training accuracy is 0.562500, time 56.782 samples/sec
+epoch 8, total_step 125280, total loss is 13.41 , inference loss is 6.07, weight deacy loss is 7.34, training accuracy is 0.593750, time 56.544 samples/sec
+epoch 8, total_step 125300, total loss is 14.14 , inference loss is 6.80, weight deacy loss is 7.34, training accuracy is 0.562500, time 56.241 samples/sec
+epoch 8, total_step 125320, total loss is 14.84 , inference loss is 7.50, weight deacy loss is 7.34, training accuracy is 0.437500, time 56.662 samples/sec
+epoch 8, total_step 125340, total loss is 11.84 , inference loss is 4.50, weight deacy loss is 7.34, training accuracy is 0.593750, time 56.309 samples/sec
+epoch 8, total_step 125360, total loss is 11.77 , inference loss is 4.42, weight deacy loss is 7.34, training accuracy is 0.593750, time 56.883 samples/sec
+epoch 8, total_step 125380, total loss is 14.15 , inference loss is 6.81, weight deacy loss is 7.34, training accuracy is 0.562500, time 56.863 samples/sec
+epoch 8, total_step 125400, total loss is 16.06 , inference loss is 8.72, weight deacy loss is 7.34, training accuracy is 0.562500, time 56.275 samples/sec
+epoch 8, total_step 125420, total loss is 14.43 , inference loss is 7.09, weight deacy loss is 7.34, training accuracy is 0.531250, time 56.051 samples/sec
+epoch 8, total_step 125440, total loss is 13.74 , inference loss is 6.40, weight deacy loss is 7.34, training accuracy is 0.593750, time 56.503 samples/sec
+epoch 8, total_step 125460, total loss is 10.12 , inference loss is 2.78, weight deacy loss is 7.34, training accuracy is 0.656250, time 56.311 samples/sec
+epoch 8, total_step 125480, total loss is 17.49 , inference loss is 10.15, weight deacy loss is 7.34, training accuracy is 0.468750, time 56.234 samples/sec
+epoch 8, total_step 125500, total loss is 14.46 , inference loss is 7.12, weight deacy loss is 7.34, training accuracy is 0.437500, time 55.574 samples/sec
+epoch 8, total_step 125520, total loss is 13.65 , inference loss is 6.31, weight deacy loss is 7.34, training accuracy is 0.468750, time 55.917 samples/sec
+epoch 8, total_step 125540, total loss is 12.18 , inference loss is 4.84, weight deacy loss is 7.34, training accuracy is 0.562500, time 56.603 samples/sec
+epoch 8, total_step 125560, total loss is 14.46 , inference loss is 7.13, weight deacy loss is 7.34, training accuracy is 0.437500, time 57.072 samples/sec
+epoch 8, total_step 125580, total loss is 15.86 , inference loss is 8.52, weight deacy loss is 7.34, training accuracy is 0.468750, time 57.051 samples/sec
+epoch 8, total_step 125600, total loss is 11.64 , inference loss is 4.31, weight deacy loss is 7.34, training accuracy is 0.625000, time 56.547 samples/sec
+epoch 8, total_step 125620, total loss is 11.21 , inference loss is 3.88, weight deacy loss is 7.34, training accuracy is 0.593750, time 56.608 samples/sec
+epoch 8, total_step 125640, total loss is 15.90 , inference loss is 8.56, weight deacy loss is 7.34, training accuracy is 0.593750, time 56.779 samples/sec
+epoch 8, total_step 125660, total loss is 12.66 , inference loss is 5.32, weight deacy loss is 7.34, training accuracy is 0.406250, time 57.081 samples/sec
+epoch 8, total_step 125680, total loss is 13.24 , inference loss is 5.91, weight deacy loss is 7.33, training accuracy is 0.500000, time 56.425 samples/sec
+epoch 8, total_step 125700, total loss is 13.79 , inference loss is 6.46, weight deacy loss is 7.33, training accuracy is 0.531250, time 56.755 samples/sec
+epoch 8, total_step 125720, total loss is 11.83 , inference loss is 4.49, weight deacy loss is 7.33, training accuracy is 0.562500, time 56.303 samples/sec
+epoch 8, total_step 125740, total loss is 14.04 , inference loss is 6.71, weight deacy loss is 7.33, training accuracy is 0.468750, time 56.186 samples/sec
+epoch 8, total_step 125760, total loss is 16.69 , inference loss is 9.36, weight deacy loss is 7.33, training accuracy is 0.562500, time 56.833 samples/sec
+epoch 8, total_step 125780, total loss is 16.15 , inference loss is 8.82, weight deacy loss is 7.33, training accuracy is 0.468750, time 56.113 samples/sec
+epoch 8, total_step 125800, total loss is 14.79 , inference loss is 7.46, weight deacy loss is 7.33, training accuracy is 0.500000, time 56.445 samples/sec
+epoch 8, total_step 125820, total loss is 16.54 , inference loss is 9.21, weight deacy loss is 7.33, training accuracy is 0.406250, time 56.356 samples/sec
+epoch 8, total_step 125840, total loss is 15.29 , inference loss is 7.96, weight deacy loss is 7.33, training accuracy is 0.437500, time 56.445 samples/sec
+epoch 8, total_step 125860, total loss is 12.55 , inference loss is 5.22, weight deacy loss is 7.33, training accuracy is 0.687500, time 56.290 samples/sec
+epoch 8, total_step 125880, total loss is 15.77 , inference loss is 8.44, weight deacy loss is 7.33, training accuracy is 0.312500, time 56.875 samples/sec
+epoch 8, total_step 125900, total loss is 13.35 , inference loss is 6.02, weight deacy loss is 7.33, training accuracy is 0.343750, time 56.004 samples/sec
+epoch 8, total_step 125920, total loss is 21.76 , inference loss is 14.43, weight deacy loss is 7.33, training accuracy is 0.312500, time 56.719 samples/sec
+epoch 8, total_step 125940, total loss is 15.00 , inference loss is 7.67, weight deacy loss is 7.33, training accuracy is 0.375000, time 57.594 samples/sec
+epoch 8, total_step 125960, total loss is 14.54 , inference loss is 7.21, weight deacy loss is 7.33, training accuracy is 0.437500, time 57.067 samples/sec
+epoch 8, total_step 125980, total loss is 18.84 , inference loss is 11.51, weight deacy loss is 7.33, training accuracy is 0.437500, time 56.646 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.63108599999995
+best_threshold_index 116 0.9864814814814815
+best_threshold_index 116 0.9851851851851852
+best_threshold_index 116 0.9862962962962963
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 116 0.9862962962962963
+best_threshold_index 116 0.9851851851851852
+best_threshold_index 116 0.9861111111111112
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 116 0.9846296296296296
+best_threshold_index 116 0.9846296296296296
+[lfw][126000]XNorm: 23.254987
+[lfw][126000]Accuracy-Flip: 0.98567+-0.00597
+testing verification..
+(14000, 512)
+infer time 85.272783
+best_threshold_index 116 0.9876190476190476
+best_threshold_index 116 0.9866666666666667
+best_threshold_index 116 0.9866666666666667
+best_threshold_index 116 0.986984126984127
+best_threshold_index 116 0.9877777777777778
+best_threshold_index 116 0.9861904761904762
+best_threshold_index 116 0.9879365079365079
+best_threshold_index 116 0.9868253968253968
+best_threshold_index 116 0.9868253968253968
+best_threshold_index 116 0.9879365079365079
+[cfp_ff][126000]XNorm: 22.424108
+[cfp_ff][126000]Accuracy-Flip: 0.98714+-0.00531
+testing verification..
+(14000, 512)
+infer time 86.26068800000004
+best_threshold_index 139 0.9158730158730158
+best_threshold_index 139 0.9160317460317461
+best_threshold_index 139 0.9131746031746032
+best_threshold_index 139 0.9161904761904762
+best_threshold_index 139 0.9187301587301587
+best_threshold_index 139 0.9141269841269841
+best_threshold_index 139 0.9133333333333333
+best_threshold_index 139 0.9166666666666666
+best_threshold_index 139 0.915079365079365
+best_threshold_index 139 0.9165079365079365
+[cfp_fp][126000]XNorm: 22.850981
+[cfp_fp][126000]Accuracy-Flip: 0.91557+-0.01446
+testing verification..
+(12000, 512)
+infer time 73.73835200000005
+best_threshold_index 141 0.8987037037037037
+best_threshold_index 140 0.8983333333333333
+best_threshold_index 144 0.9014814814814814
+best_threshold_index 141 0.9016666666666666
+best_threshold_index 141 0.9055555555555556
+best_threshold_index 143 0.8998148148148148
+best_threshold_index 140 0.9014814814814814
+best_threshold_index 144 0.9061111111111111
+best_threshold_index 144 0.9061111111111111
+best_threshold_index 141 0.9037037037037037
+[agedb_30][126000]XNorm: 23.746502
+[agedb_30][126000]Accuracy-Flip: 0.89850+-0.02530
+test accuracy is: 0.9856666666666666
+epoch 8, total_step 126000, total loss is 16.45 , inference loss is 9.12, weight deacy loss is 7.33, training accuracy is 0.406250, time 54.015 samples/sec
+epoch 8, total_step 126020, total loss is 14.10 , inference loss is 6.78, weight deacy loss is 7.33, training accuracy is 0.468750, time 56.635 samples/sec
+epoch 8, total_step 126040, total loss is 18.28 , inference loss is 10.95, weight deacy loss is 7.33, training accuracy is 0.406250, time 56.125 samples/sec
+epoch 8, total_step 126060, total loss is 14.81 , inference loss is 7.49, weight deacy loss is 7.33, training accuracy is 0.593750, time 56.946 samples/sec
+epoch 8, total_step 126080, total loss is 13.01 , inference loss is 5.68, weight deacy loss is 7.33, training accuracy is 0.500000, time 57.222 samples/sec
+epoch 8, total_step 126100, total loss is 17.17 , inference loss is 9.84, weight deacy loss is 7.33, training accuracy is 0.343750, time 57.077 samples/sec
+epoch 8, total_step 126120, total loss is 15.61 , inference loss is 8.28, weight deacy loss is 7.33, training accuracy is 0.437500, time 56.332 samples/sec
+epoch 8, total_step 126140, total loss is 12.27 , inference loss is 4.95, weight deacy loss is 7.32, training accuracy is 0.562500, time 56.129 samples/sec
+epoch 8, total_step 126160, total loss is 15.68 , inference loss is 8.35, weight deacy loss is 7.32, training accuracy is 0.312500, time 56.345 samples/sec
+epoch 8, total_step 126180, total loss is 14.19 , inference loss is 6.87, weight deacy loss is 7.32, training accuracy is 0.500000, time 56.402 samples/sec
+epoch 8, total_step 126200, total loss is 17.70 , inference loss is 10.37, weight deacy loss is 7.32, training accuracy is 0.468750, time 56.600 samples/sec
+epoch 8, total_step 126220, total loss is 15.51 , inference loss is 8.19, weight deacy loss is 7.32, training accuracy is 0.406250, time 56.448 samples/sec
+epoch 8, total_step 126240, total loss is 19.50 , inference loss is 12.18, weight deacy loss is 7.32, training accuracy is 0.468750, time 56.272 samples/sec
+epoch 8, total_step 126260, total loss is 12.96 , inference loss is 5.63, weight deacy loss is 7.32, training accuracy is 0.468750, time 56.654 samples/sec
+epoch 8, total_step 126280, total loss is 17.35 , inference loss is 10.03, weight deacy loss is 7.32, training accuracy is 0.312500, time 56.317 samples/sec
+epoch 8, total_step 126300, total loss is 14.70 , inference loss is 7.38, weight deacy loss is 7.32, training accuracy is 0.468750, time 56.721 samples/sec
+epoch 8, total_step 126320, total loss is 13.73 , inference loss is 6.41, weight deacy loss is 7.32, training accuracy is 0.468750, time 56.642 samples/sec
+epoch 8, total_step 126340, total loss is 20.11 , inference loss is 12.79, weight deacy loss is 7.32, training accuracy is 0.343750, time 56.707 samples/sec
+epoch 8, total_step 126360, total loss is 12.05 , inference loss is 4.73, weight deacy loss is 7.32, training accuracy is 0.500000, time 56.782 samples/sec
+epoch 8, total_step 126380, total loss is 18.17 , inference loss is 10.85, weight deacy loss is 7.32, training accuracy is 0.343750, time 56.595 samples/sec
+epoch 8, total_step 126400, total loss is 22.27 , inference loss is 14.95, weight deacy loss is 7.32, training accuracy is 0.406250, time 56.658 samples/sec
+epoch 8, total_step 126420, total loss is 17.07 , inference loss is 9.75, weight deacy loss is 7.32, training accuracy is 0.343750, time 56.386 samples/sec
+epoch 8, total_step 126440, total loss is 15.25 , inference loss is 7.93, weight deacy loss is 7.32, training accuracy is 0.343750, time 56.323 samples/sec
+epoch 8, total_step 126460, total loss is 13.17 , inference loss is 5.85, weight deacy loss is 7.32, training accuracy is 0.500000, time 56.590 samples/sec
+epoch 8, total_step 126480, total loss is 15.93 , inference loss is 8.61, weight deacy loss is 7.32, training accuracy is 0.375000, time 56.615 samples/sec
+epoch 8, total_step 126500, total loss is 16.93 , inference loss is 9.61, weight deacy loss is 7.32, training accuracy is 0.375000, time 56.085 samples/sec
+epoch 8, total_step 126520, total loss is 17.57 , inference loss is 10.25, weight deacy loss is 7.32, training accuracy is 0.375000, time 56.047 samples/sec
+epoch 8, total_step 126540, total loss is 13.32 , inference loss is 6.00, weight deacy loss is 7.32, training accuracy is 0.531250, time 55.361 samples/sec
+epoch 8, total_step 126560, total loss is 20.27 , inference loss is 12.96, weight deacy loss is 7.32, training accuracy is 0.218750, time 56.403 samples/sec
+epoch 8, total_step 126580, total loss is 15.85 , inference loss is 8.53, weight deacy loss is 7.32, training accuracy is 0.437500, time 57.004 samples/sec
+epoch 8, total_step 126600, total loss is 19.43 , inference loss is 12.12, weight deacy loss is 7.32, training accuracy is 0.437500, time 56.282 samples/sec
+epoch 8, total_step 126620, total loss is 14.52 , inference loss is 7.21, weight deacy loss is 7.31, training accuracy is 0.437500, time 55.715 samples/sec
+epoch 8, total_step 126640, total loss is 15.59 , inference loss is 8.28, weight deacy loss is 7.31, training accuracy is 0.437500, time 56.702 samples/sec
+epoch 8, total_step 126660, total loss is 11.83 , inference loss is 4.51, weight deacy loss is 7.31, training accuracy is 0.406250, time 56.747 samples/sec
+epoch 8, total_step 126680, total loss is 16.26 , inference loss is 8.95, weight deacy loss is 7.31, training accuracy is 0.468750, time 56.621 samples/sec
+epoch 8, total_step 126700, total loss is 15.59 , inference loss is 8.28, weight deacy loss is 7.31, training accuracy is 0.406250, time 56.615 samples/sec
+epoch 8, total_step 126720, total loss is 15.28 , inference loss is 7.96, weight deacy loss is 7.31, training accuracy is 0.406250, time 57.113 samples/sec
+epoch 8, total_step 126740, total loss is 15.39 , inference loss is 8.07, weight deacy loss is 7.31, training accuracy is 0.468750, time 56.272 samples/sec
+epoch 8, total_step 126760, total loss is 18.09 , inference loss is 10.78, weight deacy loss is 7.31, training accuracy is 0.281250, time 56.374 samples/sec
+epoch 8, total_step 126780, total loss is 15.17 , inference loss is 7.86, weight deacy loss is 7.31, training accuracy is 0.375000, time 56.561 samples/sec
+epoch 8, total_step 126800, total loss is 14.98 , inference loss is 7.67, weight deacy loss is 7.31, training accuracy is 0.312500, time 56.755 samples/sec
+epoch 8, total_step 126820, total loss is 18.31 , inference loss is 11.00, weight deacy loss is 7.31, training accuracy is 0.281250, time 56.131 samples/sec
+epoch 8, total_step 126840, total loss is 12.92 , inference loss is 5.61, weight deacy loss is 7.31, training accuracy is 0.406250, time 57.585 samples/sec
+epoch 8, total_step 126860, total loss is 19.22 , inference loss is 11.91, weight deacy loss is 7.31, training accuracy is 0.250000, time 57.344 samples/sec
+epoch 8, total_step 126880, total loss is 12.65 , inference loss is 5.35, weight deacy loss is 7.31, training accuracy is 0.531250, time 56.958 samples/sec
+epoch 8, total_step 126900, total loss is 16.96 , inference loss is 9.65, weight deacy loss is 7.31, training accuracy is 0.437500, time 56.191 samples/sec
+epoch 8, total_step 126920, total loss is 17.15 , inference loss is 9.84, weight deacy loss is 7.31, training accuracy is 0.312500, time 55.868 samples/sec
+epoch 8, total_step 126940, total loss is 17.21 , inference loss is 9.90, weight deacy loss is 7.31, training accuracy is 0.375000, time 57.949 samples/sec
+epoch 8, total_step 126960, total loss is 13.63 , inference loss is 6.32, weight deacy loss is 7.31, training accuracy is 0.375000, time 57.298 samples/sec
+epoch 8, total_step 126980, total loss is 19.63 , inference loss is 12.33, weight deacy loss is 7.31, training accuracy is 0.406250, time 56.818 samples/sec
+epoch 8, total_step 127000, total loss is 13.41 , inference loss is 6.10, weight deacy loss is 7.31, training accuracy is 0.531250, time 56.327 samples/sec
+epoch 8, total_step 127020, total loss is 16.55 , inference loss is 9.24, weight deacy loss is 7.31, training accuracy is 0.437500, time 56.745 samples/sec
+epoch 8, total_step 127040, total loss is 14.52 , inference loss is 7.22, weight deacy loss is 7.31, training accuracy is 0.500000, time 56.942 samples/sec
+epoch 8, total_step 127060, total loss is 16.19 , inference loss is 8.88, weight deacy loss is 7.31, training accuracy is 0.437500, time 55.362 samples/sec
+epoch 8, total_step 127080, total loss is 18.56 , inference loss is 11.26, weight deacy loss is 7.30, training accuracy is 0.312500, time 56.313 samples/sec
+epoch 8, total_step 127100, total loss is 15.78 , inference loss is 8.47, weight deacy loss is 7.30, training accuracy is 0.468750, time 56.872 samples/sec
+epoch 8, total_step 127120, total loss is 14.69 , inference loss is 7.39, weight deacy loss is 7.30, training accuracy is 0.406250, time 56.327 samples/sec
+epoch 8, total_step 127140, total loss is 17.78 , inference loss is 10.47, weight deacy loss is 7.30, training accuracy is 0.218750, time 56.874 samples/sec
+epoch 8, total_step 127160, total loss is 16.69 , inference loss is 9.39, weight deacy loss is 7.30, training accuracy is 0.375000, time 56.625 samples/sec
+epoch 8, total_step 127180, total loss is 15.47 , inference loss is 8.16, weight deacy loss is 7.30, training accuracy is 0.406250, time 55.527 samples/sec
+epoch 8, total_step 127200, total loss is 14.86 , inference loss is 7.56, weight deacy loss is 7.30, training accuracy is 0.437500, time 56.078 samples/sec
+epoch 8, total_step 127220, total loss is 15.36 , inference loss is 8.05, weight deacy loss is 7.30, training accuracy is 0.343750, time 56.150 samples/sec
+epoch 8, total_step 127240, total loss is 17.11 , inference loss is 9.81, weight deacy loss is 7.30, training accuracy is 0.468750, time 56.525 samples/sec
+epoch 8, total_step 127260, total loss is 15.81 , inference loss is 8.51, weight deacy loss is 7.30, training accuracy is 0.437500, time 56.939 samples/sec
+epoch 8, total_step 127280, total loss is 19.04 , inference loss is 11.74, weight deacy loss is 7.30, training accuracy is 0.343750, time 56.808 samples/sec
+epoch 8, total_step 127300, total loss is 15.36 , inference loss is 8.06, weight deacy loss is 7.30, training accuracy is 0.375000, time 56.583 samples/sec
+epoch 8, total_step 127320, total loss is 14.60 , inference loss is 7.30, weight deacy loss is 7.30, training accuracy is 0.437500, time 56.806 samples/sec
+epoch 8, total_step 127340, total loss is 16.74 , inference loss is 9.44, weight deacy loss is 7.30, training accuracy is 0.343750, time 56.451 samples/sec
+epoch 8, total_step 127360, total loss is 13.78 , inference loss is 6.48, weight deacy loss is 7.30, training accuracy is 0.343750, time 56.655 samples/sec
+epoch 8, total_step 127380, total loss is 22.43 , inference loss is 15.13, weight deacy loss is 7.30, training accuracy is 0.281250, time 56.660 samples/sec
+epoch 8, total_step 127400, total loss is 15.22 , inference loss is 7.93, weight deacy loss is 7.30, training accuracy is 0.312500, time 57.148 samples/sec
+epoch 8, total_step 127420, total loss is 13.46 , inference loss is 6.17, weight deacy loss is 7.30, training accuracy is 0.562500, time 55.798 samples/sec
+epoch 8, total_step 127440, total loss is 19.83 , inference loss is 12.53, weight deacy loss is 7.30, training accuracy is 0.250000, time 57.357 samples/sec
+epoch 8, total_step 127460, total loss is 14.55 , inference loss is 7.26, weight deacy loss is 7.30, training accuracy is 0.406250, time 56.030 samples/sec
+epoch 8, total_step 127480, total loss is 15.12 , inference loss is 7.82, weight deacy loss is 7.30, training accuracy is 0.250000, time 56.871 samples/sec
+epoch 8, total_step 127500, total loss is 15.55 , inference loss is 8.26, weight deacy loss is 7.30, training accuracy is 0.468750, time 56.059 samples/sec
+epoch 8, total_step 127520, total loss is 20.25 , inference loss is 12.95, weight deacy loss is 7.29, training accuracy is 0.218750, time 56.971 samples/sec
+epoch 8, total_step 127540, total loss is 12.29 , inference loss is 5.00, weight deacy loss is 7.29, training accuracy is 0.437500, time 56.086 samples/sec
+epoch 8, total_step 127560, total loss is 18.60 , inference loss is 11.30, weight deacy loss is 7.29, training accuracy is 0.281250, time 56.821 samples/sec
+epoch 8, total_step 127580, total loss is 18.66 , inference loss is 11.37, weight deacy loss is 7.29, training accuracy is 0.343750, time 57.151 samples/sec
+epoch 8, total_step 127600, total loss is 17.23 , inference loss is 9.94, weight deacy loss is 7.29, training accuracy is 0.500000, time 57.710 samples/sec
+epoch 8, total_step 127620, total loss is 17.03 , inference loss is 9.74, weight deacy loss is 7.29, training accuracy is 0.343750, time 57.005 samples/sec
+epoch 8, total_step 127640, total loss is 16.07 , inference loss is 8.78, weight deacy loss is 7.29, training accuracy is 0.187500, time 56.807 samples/sec
+epoch 8, total_step 127660, total loss is 13.09 , inference loss is 5.80, weight deacy loss is 7.29, training accuracy is 0.531250, time 56.614 samples/sec
+epoch 8, total_step 127680, total loss is 17.52 , inference loss is 10.23, weight deacy loss is 7.29, training accuracy is 0.343750, time 55.584 samples/sec
+epoch 8, total_step 127700, total loss is 19.49 , inference loss is 12.20, weight deacy loss is 7.29, training accuracy is 0.281250, time 56.436 samples/sec
+epoch 8, total_step 127720, total loss is 17.06 , inference loss is 9.77, weight deacy loss is 7.29, training accuracy is 0.406250, time 56.764 samples/sec
+epoch 8, total_step 127740, total loss is 12.75 , inference loss is 5.45, weight deacy loss is 7.29, training accuracy is 0.437500, time 56.716 samples/sec
+epoch 8, total_step 127760, total loss is 13.34 , inference loss is 6.05, weight deacy loss is 7.29, training accuracy is 0.531250, time 56.319 samples/sec
+epoch 8, total_step 127780, total loss is 16.63 , inference loss is 9.34, weight deacy loss is 7.29, training accuracy is 0.281250, time 56.174 samples/sec
+epoch 8, total_step 127800, total loss is 14.38 , inference loss is 7.09, weight deacy loss is 7.29, training accuracy is 0.343750, time 57.013 samples/sec
+epoch 8, total_step 127820, total loss is 17.96 , inference loss is 10.67, weight deacy loss is 7.29, training accuracy is 0.343750, time 56.773 samples/sec
+epoch 8, total_step 127840, total loss is 14.08 , inference loss is 6.79, weight deacy loss is 7.29, training accuracy is 0.437500, time 56.595 samples/sec
+epoch 8, total_step 127860, total loss is 20.70 , inference loss is 13.41, weight deacy loss is 7.29, training accuracy is 0.312500, time 56.227 samples/sec
+epoch 8, total_step 127880, total loss is 16.36 , inference loss is 9.07, weight deacy loss is 7.29, training accuracy is 0.406250, time 56.288 samples/sec
+epoch 8, total_step 127900, total loss is 16.07 , inference loss is 8.78, weight deacy loss is 7.29, training accuracy is 0.343750, time 57.035 samples/sec
+epoch 8, total_step 127920, total loss is 16.17 , inference loss is 8.88, weight deacy loss is 7.29, training accuracy is 0.500000, time 55.488 samples/sec
+epoch 8, total_step 127940, total loss is 21.61 , inference loss is 14.32, weight deacy loss is 7.29, training accuracy is 0.218750, time 57.077 samples/sec
+epoch 8, total_step 127960, total loss is 11.84 , inference loss is 4.56, weight deacy loss is 7.29, training accuracy is 0.437500, time 56.624 samples/sec
+epoch 8, total_step 127980, total loss is 16.36 , inference loss is 9.08, weight deacy loss is 7.28, training accuracy is 0.437500, time 56.237 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.531728
+best_threshold_index 116 0.9866666666666667
+best_threshold_index 116 0.9853703703703703
+best_threshold_index 115 0.9859259259259259
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 113 0.9859259259259259
+best_threshold_index 116 0.985
+best_threshold_index 115 0.9855555555555555
+best_threshold_index 116 0.9857407407407407
+best_threshold_index 116 0.9844444444444445
+best_threshold_index 124 0.9846296296296296
+[lfw][128000]XNorm: 22.757576
+[lfw][128000]Accuracy-Flip: 0.98417+-0.00597
+testing verification..
+(14000, 512)
+infer time 85.86630300000006
+best_threshold_index 127 0.9866666666666667
+best_threshold_index 127 0.986031746031746
+best_threshold_index 127 0.9865079365079366
+best_threshold_index 127 0.9868253968253968
+best_threshold_index 127 0.9865079365079366
+best_threshold_index 127 0.9855555555555555
+best_threshold_index 126 0.986031746031746
+best_threshold_index 127 0.9861904761904762
+best_threshold_index 127 0.9853968253968254
+best_threshold_index 127 0.9873015873015873
+[cfp_ff][128000]XNorm: 22.282797
+[cfp_ff][128000]Accuracy-Flip: 0.98586+-0.00488
+testing verification..
+(14000, 512)
+infer time 85.28908200000008
+best_threshold_index 141 0.9266666666666666
+best_threshold_index 141 0.926031746031746
+best_threshold_index 141 0.923968253968254
+best_threshold_index 141 0.9293650793650794
+best_threshold_index 141 0.9284126984126985
+best_threshold_index 141 0.9242857142857143
+best_threshold_index 143 0.923968253968254
+best_threshold_index 141 0.9257142857142857
+best_threshold_index 141 0.9257142857142857
+best_threshold_index 141 0.9261904761904762
+[cfp_fp][128000]XNorm: 22.195448
+[cfp_fp][128000]Accuracy-Flip: 0.92529+-0.01485
+testing verification..
+(12000, 512)
+infer time 74.05538600000003
+best_threshold_index 143 0.9009259259259259
+best_threshold_index 143 0.9007407407407407
+best_threshold_index 143 0.9035185185185185
+best_threshold_index 138 0.9025925925925926
+best_threshold_index 143 0.9074074074074074
+best_threshold_index 143 0.9014814814814814
+best_threshold_index 138 0.9016666666666666
+best_threshold_index 138 0.9092592592592592
+best_threshold_index 143 0.9077777777777778
+best_threshold_index 138 0.905
+[agedb_30][128000]XNorm: 23.685583
+[agedb_30][128000]Accuracy-Flip: 0.90100+-0.02751
+test accuracy is: 0.9841666666666666
+epoch 8, total_step 128000, total loss is 18.06 , inference loss is 10.77, weight deacy loss is 7.28, training accuracy is 0.468750, time 53.856 samples/sec
+epoch 8, total_step 128020, total loss is 14.01 , inference loss is 6.72, weight deacy loss is 7.28, training accuracy is 0.406250, time 55.618 samples/sec
+epoch 8, total_step 128040, total loss is 21.87 , inference loss is 14.59, weight deacy loss is 7.28, training accuracy is 0.375000, time 56.074 samples/sec
+epoch 8, total_step 128060, total loss is 16.07 , inference loss is 8.78, weight deacy loss is 7.28, training accuracy is 0.437500, time 56.687 samples/sec
+epoch 8, total_step 128080, total loss is 18.91 , inference loss is 11.62, weight deacy loss is 7.28, training accuracy is 0.312500, time 56.686 samples/sec
+epoch 8, total_step 128100, total loss is 18.31 , inference loss is 11.03, weight deacy loss is 7.28, training accuracy is 0.343750, time 55.458 samples/sec
+epoch 8, total_step 128120, total loss is 13.07 , inference loss is 5.79, weight deacy loss is 7.28, training accuracy is 0.468750, time 55.890 samples/sec
+epoch 8, total_step 128140, total loss is 14.65 , inference loss is 7.37, weight deacy loss is 7.28, training accuracy is 0.281250, time 56.662 samples/sec
+epoch 8, total_step 128160, total loss is 13.34 , inference loss is 6.06, weight deacy loss is 7.28, training accuracy is 0.562500, time 56.104 samples/sec
+epoch 8, total_step 128180, total loss is 14.16 , inference loss is 6.88, weight deacy loss is 7.28, training accuracy is 0.468750, time 56.378 samples/sec
+epoch 8, total_step 128200, total loss is 17.42 , inference loss is 10.14, weight deacy loss is 7.28, training accuracy is 0.406250, time 57.246 samples/sec
+epoch 8, total_step 128220, total loss is 17.51 , inference loss is 10.24, weight deacy loss is 7.28, training accuracy is 0.250000, time 56.434 samples/sec
+epoch 8, total_step 128240, total loss is 12.05 , inference loss is 4.77, weight deacy loss is 7.28, training accuracy is 0.500000, time 56.959 samples/sec
+epoch 8, total_step 128260, total loss is 14.70 , inference loss is 7.42, weight deacy loss is 7.28, training accuracy is 0.437500, time 56.441 samples/sec
+epoch 8, total_step 128280, total loss is 18.10 , inference loss is 10.83, weight deacy loss is 7.28, training accuracy is 0.250000, time 56.708 samples/sec
+epoch 8, total_step 128300, total loss is 18.54 , inference loss is 11.26, weight deacy loss is 7.28, training accuracy is 0.406250, time 56.450 samples/sec
+epoch 8, total_step 128320, total loss is 11.68 , inference loss is 4.41, weight deacy loss is 7.28, training accuracy is 0.500000, time 56.745 samples/sec
+epoch 8, total_step 128340, total loss is 13.31 , inference loss is 6.03, weight deacy loss is 7.28, training accuracy is 0.625000, time 56.504 samples/sec
+epoch 8, total_step 128360, total loss is 16.72 , inference loss is 9.44, weight deacy loss is 7.28, training accuracy is 0.343750, time 56.632 samples/sec
+epoch 8, total_step 128380, total loss is 14.98 , inference loss is 7.70, weight deacy loss is 7.28, training accuracy is 0.406250, time 56.205 samples/sec
+epoch 8, total_step 128400, total loss is 11.37 , inference loss is 4.10, weight deacy loss is 7.28, training accuracy is 0.625000, time 57.213 samples/sec
+epoch 8, total_step 128420, total loss is 15.86 , inference loss is 8.59, weight deacy loss is 7.27, training accuracy is 0.500000, time 55.456 samples/sec
+epoch 8, total_step 128440, total loss is 10.52 , inference loss is 3.24, weight deacy loss is 7.27, training accuracy is 0.625000, time 55.993 samples/sec
+epoch 8, total_step 128460, total loss is 15.33 , inference loss is 8.06, weight deacy loss is 7.27, training accuracy is 0.500000, time 55.831 samples/sec
+epoch 8, total_step 128480, total loss is 15.30 , inference loss is 8.03, weight deacy loss is 7.27, training accuracy is 0.406250, time 55.986 samples/sec
+epoch 8, total_step 128500, total loss is 15.51 , inference loss is 8.24, weight deacy loss is 7.27, training accuracy is 0.468750, time 56.817 samples/sec
+epoch 8, total_step 128520, total loss is 13.02 , inference loss is 5.75, weight deacy loss is 7.27, training accuracy is 0.437500, time 55.719 samples/sec
+epoch 8, total_step 128540, total loss is 16.04 , inference loss is 8.77, weight deacy loss is 7.27, training accuracy is 0.437500, time 56.565 samples/sec
+epoch 8, total_step 128560, total loss is 22.42 , inference loss is 15.15, weight deacy loss is 7.27, training accuracy is 0.281250, time 56.490 samples/sec
+epoch 8, total_step 128580, total loss is 16.69 , inference loss is 9.42, weight deacy loss is 7.27, training accuracy is 0.343750, time 57.166 samples/sec
+epoch 8, total_step 128600, total loss is 14.46 , inference loss is 7.19, weight deacy loss is 7.27, training accuracy is 0.437500, time 56.258 samples/sec
+epoch 8, total_step 128620, total loss is 14.77 , inference loss is 7.50, weight deacy loss is 7.27, training accuracy is 0.375000, time 55.956 samples/sec
+epoch 8, total_step 128640, total loss is 15.61 , inference loss is 8.34, weight deacy loss is 7.27, training accuracy is 0.437500, time 56.140 samples/sec
+epoch 8, total_step 128660, total loss is 17.84 , inference loss is 10.57, weight deacy loss is 7.27, training accuracy is 0.375000, time 56.816 samples/sec
+epoch 8, total_step 128680, total loss is 18.57 , inference loss is 11.30, weight deacy loss is 7.27, training accuracy is 0.250000, time 55.976 samples/sec
+epoch 8, total_step 128700, total loss is 18.47 , inference loss is 11.20, weight deacy loss is 7.27, training accuracy is 0.312500, time 55.702 samples/sec
+epoch 8, total_step 128720, total loss is 18.45 , inference loss is 11.18, weight deacy loss is 7.27, training accuracy is 0.500000, time 57.911 samples/sec
+epoch 8, total_step 128740, total loss is 16.78 , inference loss is 9.52, weight deacy loss is 7.27, training accuracy is 0.250000, time 56.777 samples/sec
+epoch 8, total_step 128760, total loss is 23.88 , inference loss is 16.62, weight deacy loss is 7.27, training accuracy is 0.156250, time 56.761 samples/sec
+epoch 8, total_step 128780, total loss is 18.59 , inference loss is 11.33, weight deacy loss is 7.27, training accuracy is 0.312500, time 56.805 samples/sec
+epoch 8, total_step 128800, total loss is 10.53 , inference loss is 3.26, weight deacy loss is 7.27, training accuracy is 0.625000, time 57.115 samples/sec
+epoch 8, total_step 128820, total loss is 21.56 , inference loss is 14.30, weight deacy loss is 7.27, training accuracy is 0.312500, time 56.837 samples/sec
+epoch 8, total_step 128840, total loss is 14.31 , inference loss is 7.04, weight deacy loss is 7.27, training accuracy is 0.468750, time 56.864 samples/sec
+epoch 8, total_step 128860, total loss is 12.33 , inference loss is 5.07, weight deacy loss is 7.26, training accuracy is 0.593750, time 56.840 samples/sec
+epoch 8, total_step 128880, total loss is 20.43 , inference loss is 13.17, weight deacy loss is 7.26, training accuracy is 0.281250, time 56.443 samples/sec
+epoch 8, total_step 128900, total loss is 13.51 , inference loss is 6.24, weight deacy loss is 7.26, training accuracy is 0.468750, time 56.472 samples/sec
+epoch 8, total_step 128920, total loss is 14.94 , inference loss is 7.67, weight deacy loss is 7.26, training accuracy is 0.281250, time 56.868 samples/sec
+epoch 8, total_step 128940, total loss is 15.19 , inference loss is 7.92, weight deacy loss is 7.26, training accuracy is 0.562500, time 56.777 samples/sec
+epoch 8, total_step 128960, total loss is 17.47 , inference loss is 10.21, weight deacy loss is 7.26, training accuracy is 0.343750, time 56.390 samples/sec
+epoch 8, total_step 128980, total loss is 20.01 , inference loss is 12.75, weight deacy loss is 7.26, training accuracy is 0.312500, time 56.053 samples/sec
+epoch 8, total_step 129000, total loss is 16.38 , inference loss is 9.12, weight deacy loss is 7.26, training accuracy is 0.375000, time 55.872 samples/sec
+epoch 8, total_step 129020, total loss is 16.92 , inference loss is 9.66, weight deacy loss is 7.26, training accuracy is 0.343750, time 56.678 samples/sec
+epoch 8, total_step 129040, total loss is 15.55 , inference loss is 8.29, weight deacy loss is 7.26, training accuracy is 0.437500, time 56.503 samples/sec
+epoch 8, total_step 129060, total loss is 14.77 , inference loss is 7.51, weight deacy loss is 7.26, training accuracy is 0.375000, time 56.986 samples/sec
+epoch 8, total_step 129080, total loss is 18.55 , inference loss is 11.29, weight deacy loss is 7.26, training accuracy is 0.312500, time 56.988 samples/sec
+epoch 8, total_step 129100, total loss is 12.45 , inference loss is 5.19, weight deacy loss is 7.26, training accuracy is 0.406250, time 55.631 samples/sec
+epoch 8, total_step 129120, total loss is 19.93 , inference loss is 12.67, weight deacy loss is 7.26, training accuracy is 0.281250, time 57.017 samples/sec
+epoch 8, total_step 129140, total loss is 16.50 , inference loss is 9.24, weight deacy loss is 7.26, training accuracy is 0.437500, time 57.332 samples/sec
+epoch 8, total_step 129160, total loss is 16.34 , inference loss is 9.09, weight deacy loss is 7.26, training accuracy is 0.375000, time 56.552 samples/sec
+epoch 8, total_step 129180, total loss is 16.67 , inference loss is 9.42, weight deacy loss is 7.26, training accuracy is 0.312500, time 56.417 samples/sec
+epoch 8, total_step 129200, total loss is 12.82 , inference loss is 5.56, weight deacy loss is 7.26, training accuracy is 0.437500, time 56.449 samples/sec
+epoch 8, total_step 129220, total loss is 13.34 , inference loss is 6.08, weight deacy loss is 7.26, training accuracy is 0.468750, time 56.760 samples/sec
+epoch 8, total_step 129240, total loss is 13.85 , inference loss is 6.59, weight deacy loss is 7.26, training accuracy is 0.406250, time 56.336 samples/sec
+epoch 8, total_step 129260, total loss is 17.65 , inference loss is 10.39, weight deacy loss is 7.26, training accuracy is 0.343750, time 56.880 samples/sec
+epoch 8, total_step 129280, total loss is 15.05 , inference loss is 7.79, weight deacy loss is 7.26, training accuracy is 0.281250, time 54.958 samples/sec
+epoch 8, total_step 129300, total loss is 18.75 , inference loss is 11.49, weight deacy loss is 7.25, training accuracy is 0.312500, time 56.040 samples/sec
+epoch 8, total_step 129320, total loss is 13.11 , inference loss is 5.85, weight deacy loss is 7.25, training accuracy is 0.437500, time 56.120 samples/sec
+epoch 8, total_step 129340, total loss is 14.61 , inference loss is 7.35, weight deacy loss is 7.25, training accuracy is 0.375000, time 55.804 samples/sec
+epoch 8, total_step 129360, total loss is 15.53 , inference loss is 8.27, weight deacy loss is 7.25, training accuracy is 0.437500, time 56.997 samples/sec
+epoch 8, total_step 129380, total loss is 15.79 , inference loss is 8.54, weight deacy loss is 7.25, training accuracy is 0.500000, time 56.561 samples/sec
+epoch 8, total_step 129400, total loss is 20.47 , inference loss is 13.22, weight deacy loss is 7.25, training accuracy is 0.375000, time 56.437 samples/sec
+epoch 8, total_step 129420, total loss is 15.56 , inference loss is 8.31, weight deacy loss is 7.25, training accuracy is 0.437500, time 56.754 samples/sec
+epoch 8, total_step 129440, total loss is 16.48 , inference loss is 9.23, weight deacy loss is 7.25, training accuracy is 0.562500, time 55.956 samples/sec
+epoch 8, total_step 129460, total loss is 12.99 , inference loss is 5.74, weight deacy loss is 7.25, training accuracy is 0.500000, time 56.544 samples/sec
+epoch 8, total_step 129480, total loss is 13.50 , inference loss is 6.25, weight deacy loss is 7.25, training accuracy is 0.375000, time 56.230 samples/sec
+epoch 8, total_step 129500, total loss is 16.80 , inference loss is 9.55, weight deacy loss is 7.25, training accuracy is 0.500000, time 57.081 samples/sec
+epoch 8, total_step 129520, total loss is 16.90 , inference loss is 9.65, weight deacy loss is 7.25, training accuracy is 0.437500, time 56.172 samples/sec
+epoch 8, total_step 129540, total loss is 18.81 , inference loss is 11.56, weight deacy loss is 7.25, training accuracy is 0.312500, time 56.360 samples/sec
+epoch 8, total_step 129560, total loss is 19.96 , inference loss is 12.71, weight deacy loss is 7.25, training accuracy is 0.406250, time 57.281 samples/sec
+epoch 8, total_step 129580, total loss is 18.10 , inference loss is 10.85, weight deacy loss is 7.25, training accuracy is 0.250000, time 56.298 samples/sec
+epoch 8, total_step 129600, total loss is 17.18 , inference loss is 9.93, weight deacy loss is 7.25, training accuracy is 0.312500, time 56.871 samples/sec
+epoch 8, total_step 129620, total loss is 15.20 , inference loss is 7.95, weight deacy loss is 7.25, training accuracy is 0.500000, time 56.855 samples/sec
+epoch 8, total_step 129640, total loss is 20.48 , inference loss is 13.23, weight deacy loss is 7.25, training accuracy is 0.375000, time 56.791 samples/sec
+epoch 8, total_step 129660, total loss is 17.33 , inference loss is 10.08, weight deacy loss is 7.25, training accuracy is 0.468750, time 55.799 samples/sec
+epoch 8, total_step 129680, total loss is 15.43 , inference loss is 8.19, weight deacy loss is 7.25, training accuracy is 0.375000, time 56.807 samples/sec
+epoch 8, total_step 129700, total loss is 15.26 , inference loss is 8.01, weight deacy loss is 7.25, training accuracy is 0.468750, time 55.487 samples/sec
+epoch 8, total_step 129720, total loss is 15.39 , inference loss is 8.15, weight deacy loss is 7.25, training accuracy is 0.343750, time 56.823 samples/sec
+epoch 8, total_step 129740, total loss is 16.13 , inference loss is 8.88, weight deacy loss is 7.25, training accuracy is 0.500000, time 56.932 samples/sec
+epoch 8, total_step 129760, total loss is 18.56 , inference loss is 11.32, weight deacy loss is 7.24, training accuracy is 0.375000, time 55.591 samples/sec
+epoch 8, total_step 129780, total loss is 15.36 , inference loss is 8.12, weight deacy loss is 7.24, training accuracy is 0.406250, time 57.098 samples/sec
+epoch 8, total_step 129800, total loss is 20.51 , inference loss is 13.27, weight deacy loss is 7.24, training accuracy is 0.406250, time 56.778 samples/sec
+epoch 8, total_step 129820, total loss is 19.10 , inference loss is 11.85, weight deacy loss is 7.24, training accuracy is 0.250000, time 57.639 samples/sec
+epoch 8, total_step 129840, total loss is 16.31 , inference loss is 9.07, weight deacy loss is 7.24, training accuracy is 0.531250, time 56.453 samples/sec
+epoch 8, total_step 129860, total loss is 14.82 , inference loss is 7.57, weight deacy loss is 7.24, training accuracy is 0.406250, time 56.588 samples/sec
+epoch 8, total_step 129880, total loss is 13.96 , inference loss is 6.72, weight deacy loss is 7.24, training accuracy is 0.437500, time 56.416 samples/sec
+epoch 8, total_step 129900, total loss is 15.41 , inference loss is 8.17, weight deacy loss is 7.24, training accuracy is 0.437500, time 56.553 samples/sec
+epoch 8, total_step 129920, total loss is 16.16 , inference loss is 8.92, weight deacy loss is 7.24, training accuracy is 0.500000, time 56.508 samples/sec
+epoch 8, total_step 129940, total loss is 14.91 , inference loss is 7.67, weight deacy loss is 7.24, training accuracy is 0.500000, time 56.778 samples/sec
+epoch 8, total_step 129960, total loss is 13.61 , inference loss is 6.37, weight deacy loss is 7.24, training accuracy is 0.375000, time 56.608 samples/sec
+epoch 8, total_step 129980, total loss is 14.62 , inference loss is 7.38, weight deacy loss is 7.24, training accuracy is 0.531250, time 56.358 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.11414100000016
+best_threshold_index 114 0.9892592592592593
+best_threshold_index 114 0.9885185185185185
+best_threshold_index 111 0.9887037037037038
+best_threshold_index 111 0.9888888888888889
+best_threshold_index 111 0.9892592592592593
+best_threshold_index 114 0.9879629629629629
+best_threshold_index 111 0.9890740740740741
+best_threshold_index 111 0.9879629629629629
+best_threshold_index 114 0.9874074074074074
+best_threshold_index 114 0.9874074074074074
+[lfw][130000]XNorm: 21.850091
+[lfw][130000]Accuracy-Flip: 0.98733+-0.00611
+testing verification..
+(14000, 512)
+infer time 86.63907899999991
+best_threshold_index 115 0.9865079365079366
+best_threshold_index 120 0.9865079365079366
+best_threshold_index 119 0.9861904761904762
+best_threshold_index 120 0.986984126984127
+best_threshold_index 120 0.9880952380952381
+best_threshold_index 120 0.9865079365079366
+best_threshold_index 120 0.9873015873015873
+best_threshold_index 119 0.986984126984127
+best_threshold_index 120 0.986984126984127
+best_threshold_index 120 0.9879365079365079
+[cfp_ff][130000]XNorm: 21.706737
+[cfp_ff][130000]Accuracy-Flip: 0.98586+-0.00505
+testing verification..
+(14000, 512)
+infer time 86.139927
+best_threshold_index 135 0.9249206349206349
+best_threshold_index 135 0.9265079365079365
+best_threshold_index 135 0.9255555555555556
+best_threshold_index 135 0.9253968253968254
+best_threshold_index 137 0.9282539682539682
+best_threshold_index 137 0.9250793650793651
+best_threshold_index 137 0.9234920634920635
+best_threshold_index 137 0.9255555555555556
+best_threshold_index 137 0.9246031746031746
+best_threshold_index 135 0.9252380952380952
+[cfp_fp][130000]XNorm: 21.711862
+[cfp_fp][130000]Accuracy-Flip: 0.92371+-0.01083
+testing verification..
+(12000, 512)
+infer time 73.95175199999989
+best_threshold_index 131 0.9042592592592592
+best_threshold_index 131 0.9020370370370371
+best_threshold_index 131 0.9044444444444445
+best_threshold_index 131 0.9066666666666666
+best_threshold_index 131 0.9088888888888889
+best_threshold_index 133 0.902962962962963
+best_threshold_index 132 0.9044444444444445
+best_threshold_index 134 0.9098148148148149
+best_threshold_index 131 0.9068518518518518
+best_threshold_index 131 0.907962962962963
+[agedb_30][130000]XNorm: 22.900995
+[agedb_30][130000]Accuracy-Flip: 0.90300+-0.02197
+test accuracy is: 0.9873333333333333
+epoch 8, total_step 130000, total loss is 15.63 , inference loss is 8.39, weight deacy loss is 7.24, training accuracy is 0.375000, time 53.907 samples/sec
+epoch 8, total_step 130020, total loss is 16.56 , inference loss is 9.32, weight deacy loss is 7.24, training accuracy is 0.312500, time 56.921 samples/sec
+epoch 8, total_step 130040, total loss is 13.98 , inference loss is 6.74, weight deacy loss is 7.24, training accuracy is 0.468750, time 58.020 samples/sec
+epoch 8, total_step 130060, total loss is 14.95 , inference loss is 7.71, weight deacy loss is 7.24, training accuracy is 0.437500, time 57.008 samples/sec
+epoch 8, total_step 130080, total loss is 12.56 , inference loss is 5.32, weight deacy loss is 7.24, training accuracy is 0.562500, time 57.329 samples/sec
+epoch 8, total_step 130100, total loss is 15.94 , inference loss is 8.70, weight deacy loss is 7.24, training accuracy is 0.281250, time 56.489 samples/sec
+epoch 8, total_step 130120, total loss is 12.14 , inference loss is 4.90, weight deacy loss is 7.24, training accuracy is 0.656250, time 56.685 samples/sec
+epoch 8, total_step 130140, total loss is 16.26 , inference loss is 9.02, weight deacy loss is 7.24, training accuracy is 0.437500, time 56.749 samples/sec
+epoch 8, total_step 130160, total loss is 13.75 , inference loss is 6.51, weight deacy loss is 7.24, training accuracy is 0.281250, time 55.896 samples/sec
+epoch 8, total_step 130180, total loss is 17.69 , inference loss is 10.45, weight deacy loss is 7.24, training accuracy is 0.375000, time 56.283 samples/sec
+epoch 8, total_step 130200, total loss is 17.54 , inference loss is 10.30, weight deacy loss is 7.23, training accuracy is 0.406250, time 57.044 samples/sec
+epoch 8, total_step 130220, total loss is 13.89 , inference loss is 6.66, weight deacy loss is 7.23, training accuracy is 0.500000, time 56.045 samples/sec
+epoch 8, total_step 130240, total loss is 14.43 , inference loss is 7.20, weight deacy loss is 7.23, training accuracy is 0.343750, time 56.497 samples/sec
+epoch 8, total_step 130260, total loss is 13.19 , inference loss is 5.96, weight deacy loss is 7.23, training accuracy is 0.343750, time 56.927 samples/sec
+epoch 8, total_step 130280, total loss is 13.10 , inference loss is 5.86, weight deacy loss is 7.23, training accuracy is 0.437500, time 57.401 samples/sec
+epoch 8, total_step 130300, total loss is 13.10 , inference loss is 5.87, weight deacy loss is 7.23, training accuracy is 0.562500, time 56.750 samples/sec
+epoch 8, total_step 130320, total loss is 12.63 , inference loss is 5.40, weight deacy loss is 7.23, training accuracy is 0.500000, time 57.035 samples/sec
+epoch 8, total_step 130340, total loss is 15.01 , inference loss is 7.78, weight deacy loss is 7.23, training accuracy is 0.343750, time 56.212 samples/sec
+epoch 8, total_step 130360, total loss is 18.24 , inference loss is 11.01, weight deacy loss is 7.23, training accuracy is 0.406250, time 56.458 samples/sec
+epoch 8, total_step 130380, total loss is 13.27 , inference loss is 6.04, weight deacy loss is 7.23, training accuracy is 0.468750, time 56.798 samples/sec
+epoch 8, total_step 130400, total loss is 15.42 , inference loss is 8.19, weight deacy loss is 7.23, training accuracy is 0.531250, time 55.545 samples/sec
+epoch 8, total_step 130420, total loss is 17.84 , inference loss is 10.61, weight deacy loss is 7.23, training accuracy is 0.406250, time 57.325 samples/sec
+epoch 8, total_step 130440, total loss is 18.89 , inference loss is 11.66, weight deacy loss is 7.23, training accuracy is 0.437500, time 57.266 samples/sec
+epoch 8, total_step 130460, total loss is 14.34 , inference loss is 7.11, weight deacy loss is 7.23, training accuracy is 0.343750, time 55.732 samples/sec
+epoch 8, total_step 130480, total loss is 13.79 , inference loss is 6.56, weight deacy loss is 7.23, training accuracy is 0.437500, time 56.280 samples/sec
+epoch 8, total_step 130500, total loss is 12.95 , inference loss is 5.72, weight deacy loss is 7.23, training accuracy is 0.437500, time 55.958 samples/sec
+epoch 8, total_step 130520, total loss is 14.00 , inference loss is 6.77, weight deacy loss is 7.23, training accuracy is 0.406250, time 55.222 samples/sec
+epoch 8, total_step 130540, total loss is 15.24 , inference loss is 8.02, weight deacy loss is 7.23, training accuracy is 0.250000, time 56.293 samples/sec
+epoch 8, total_step 130560, total loss is 17.28 , inference loss is 10.06, weight deacy loss is 7.23, training accuracy is 0.343750, time 57.040 samples/sec
+epoch 8, total_step 130580, total loss is 17.83 , inference loss is 10.60, weight deacy loss is 7.23, training accuracy is 0.375000, time 56.528 samples/sec
+epoch 8, total_step 130600, total loss is 22.02 , inference loss is 14.80, weight deacy loss is 7.23, training accuracy is 0.312500, time 57.558 samples/sec
+epoch 8, total_step 130620, total loss is 16.24 , inference loss is 9.01, weight deacy loss is 7.22, training accuracy is 0.500000, time 57.137 samples/sec
+epoch 8, total_step 130640, total loss is 19.73 , inference loss is 12.51, weight deacy loss is 7.22, training accuracy is 0.281250, time 56.260 samples/sec
+epoch 8, total_step 130660, total loss is 15.60 , inference loss is 8.37, weight deacy loss is 7.22, training accuracy is 0.406250, time 55.933 samples/sec
+epoch 8, total_step 130680, total loss is 15.33 , inference loss is 8.10, weight deacy loss is 7.22, training accuracy is 0.312500, time 56.819 samples/sec
+epoch 8, total_step 130700, total loss is 15.44 , inference loss is 8.22, weight deacy loss is 7.22, training accuracy is 0.468750, time 57.058 samples/sec
+epoch 8, total_step 130720, total loss is 16.94 , inference loss is 9.72, weight deacy loss is 7.22, training accuracy is 0.312500, time 56.157 samples/sec
+epoch 8, total_step 130740, total loss is 13.46 , inference loss is 6.24, weight deacy loss is 7.22, training accuracy is 0.406250, time 56.067 samples/sec
+epoch 8, total_step 130760, total loss is 16.00 , inference loss is 8.78, weight deacy loss is 7.22, training accuracy is 0.437500, time 56.029 samples/sec
+epoch 8, total_step 130780, total loss is 18.71 , inference loss is 11.49, weight deacy loss is 7.22, training accuracy is 0.375000, time 56.688 samples/sec
+epoch 8, total_step 130800, total loss is 15.86 , inference loss is 8.64, weight deacy loss is 7.22, training accuracy is 0.312500, time 56.734 samples/sec
+epoch 8, total_step 130820, total loss is 15.34 , inference loss is 8.12, weight deacy loss is 7.22, training accuracy is 0.343750, time 56.398 samples/sec
+epoch 8, total_step 130840, total loss is 16.75 , inference loss is 9.53, weight deacy loss is 7.22, training accuracy is 0.250000, time 57.044 samples/sec
+epoch 8, total_step 130860, total loss is 18.59 , inference loss is 11.38, weight deacy loss is 7.22, training accuracy is 0.250000, time 56.348 samples/sec
+epoch 8, total_step 130880, total loss is 19.95 , inference loss is 12.73, weight deacy loss is 7.22, training accuracy is 0.281250, time 56.092 samples/sec
+epoch 8, total_step 130900, total loss is 14.40 , inference loss is 7.18, weight deacy loss is 7.22, training accuracy is 0.250000, time 57.366 samples/sec
+epoch 8, total_step 130920, total loss is 14.04 , inference loss is 6.83, weight deacy loss is 7.22, training accuracy is 0.406250, time 56.247 samples/sec
+epoch 8, total_step 130940, total loss is 16.52 , inference loss is 9.30, weight deacy loss is 7.22, training accuracy is 0.437500, time 56.439 samples/sec
+epoch 8, total_step 130960, total loss is 15.14 , inference loss is 7.92, weight deacy loss is 7.22, training accuracy is 0.343750, time 56.481 samples/sec
+epoch 8, total_step 130980, total loss is 16.70 , inference loss is 9.49, weight deacy loss is 7.22, training accuracy is 0.343750, time 57.180 samples/sec
+epoch 8, total_step 131000, total loss is 15.78 , inference loss is 8.56, weight deacy loss is 7.22, training accuracy is 0.468750, time 56.054 samples/sec
+epoch 8, total_step 131020, total loss is 16.08 , inference loss is 8.86, weight deacy loss is 7.22, training accuracy is 0.281250, time 55.656 samples/sec
+epoch 8, total_step 131040, total loss is 17.62 , inference loss is 10.41, weight deacy loss is 7.22, training accuracy is 0.312500, time 56.156 samples/sec
+epoch 8, total_step 131060, total loss is 16.83 , inference loss is 9.62, weight deacy loss is 7.21, training accuracy is 0.281250, time 57.323 samples/sec
+epoch 8, total_step 131080, total loss is 16.68 , inference loss is 9.46, weight deacy loss is 7.21, training accuracy is 0.406250, time 56.664 samples/sec
+epoch 8, total_step 131100, total loss is 15.34 , inference loss is 8.13, weight deacy loss is 7.21, training accuracy is 0.437500, time 55.325 samples/sec
+epoch 8, total_step 131120, total loss is 15.38 , inference loss is 8.16, weight deacy loss is 7.21, training accuracy is 0.375000, time 57.463 samples/sec
+epoch 8, total_step 131140, total loss is 17.87 , inference loss is 10.66, weight deacy loss is 7.21, training accuracy is 0.375000, time 56.367 samples/sec
+epoch 8, total_step 131160, total loss is 20.03 , inference loss is 12.82, weight deacy loss is 7.21, training accuracy is 0.312500, time 56.603 samples/sec
+epoch 8, total_step 131180, total loss is 16.69 , inference loss is 9.48, weight deacy loss is 7.21, training accuracy is 0.312500, time 56.770 samples/sec
+epoch 8, total_step 131200, total loss is 11.87 , inference loss is 4.66, weight deacy loss is 7.21, training accuracy is 0.625000, time 56.752 samples/sec
+epoch 8, total_step 131220, total loss is 17.75 , inference loss is 10.54, weight deacy loss is 7.21, training accuracy is 0.375000, time 56.152 samples/sec
+epoch 8, total_step 131240, total loss is 15.54 , inference loss is 8.33, weight deacy loss is 7.21, training accuracy is 0.343750, time 56.878 samples/sec
+epoch 8, total_step 131260, total loss is 18.04 , inference loss is 10.83, weight deacy loss is 7.21, training accuracy is 0.218750, time 57.072 samples/sec
+epoch 8, total_step 131280, total loss is 13.34 , inference loss is 6.13, weight deacy loss is 7.21, training accuracy is 0.437500, time 56.413 samples/sec
+epoch 8, total_step 131300, total loss is 15.95 , inference loss is 8.74, weight deacy loss is 7.21, training accuracy is 0.312500, time 56.439 samples/sec
+epoch 8, total_step 131320, total loss is 15.88 , inference loss is 8.67, weight deacy loss is 7.21, training accuracy is 0.343750, time 56.716 samples/sec
+epoch 8, total_step 131340, total loss is 19.35 , inference loss is 12.14, weight deacy loss is 7.21, training accuracy is 0.281250, time 54.819 samples/sec
+epoch 8, total_step 131360, total loss is 17.40 , inference loss is 10.19, weight deacy loss is 7.21, training accuracy is 0.437500, time 56.950 samples/sec
+epoch 8, total_step 131380, total loss is 16.37 , inference loss is 9.16, weight deacy loss is 7.21, training accuracy is 0.437500, time 56.843 samples/sec
+epoch 8, total_step 131400, total loss is 16.29 , inference loss is 9.08, weight deacy loss is 7.21, training accuracy is 0.406250, time 55.626 samples/sec
+epoch 8, total_step 131420, total loss is 16.27 , inference loss is 9.07, weight deacy loss is 7.21, training accuracy is 0.281250, time 56.508 samples/sec
+epoch 8, total_step 131440, total loss is 15.30 , inference loss is 8.09, weight deacy loss is 7.21, training accuracy is 0.375000, time 55.720 samples/sec
+epoch 8, total_step 131460, total loss is 14.48 , inference loss is 7.28, weight deacy loss is 7.21, training accuracy is 0.562500, time 55.888 samples/sec
+epoch 8, total_step 131480, total loss is 17.45 , inference loss is 10.25, weight deacy loss is 7.21, training accuracy is 0.250000, time 56.782 samples/sec
+epoch 8, total_step 131500, total loss is 17.07 , inference loss is 9.87, weight deacy loss is 7.21, training accuracy is 0.343750, time 56.087 samples/sec
+epoch 8, total_step 131520, total loss is 16.19 , inference loss is 8.98, weight deacy loss is 7.21, training accuracy is 0.375000, time 56.528 samples/sec
+epoch 8, total_step 131540, total loss is 19.21 , inference loss is 12.00, weight deacy loss is 7.21, training accuracy is 0.312500, time 57.037 samples/sec
+epoch 8, total_step 131560, total loss is 15.88 , inference loss is 8.68, weight deacy loss is 7.20, training accuracy is 0.375000, time 56.165 samples/sec
+epoch 8, total_step 131580, total loss is 13.56 , inference loss is 6.35, weight deacy loss is 7.20, training accuracy is 0.406250, time 56.508 samples/sec
+epoch 8, total_step 131600, total loss is 13.26 , inference loss is 6.06, weight deacy loss is 7.20, training accuracy is 0.437500, time 56.651 samples/sec
+epoch 8, total_step 131620, total loss is 17.42 , inference loss is 10.22, weight deacy loss is 7.20, training accuracy is 0.406250, time 56.559 samples/sec
+epoch 8, total_step 131640, total loss is 14.80 , inference loss is 7.59, weight deacy loss is 7.20, training accuracy is 0.375000, time 56.769 samples/sec
+epoch 8, total_step 131660, total loss is 15.75 , inference loss is 8.55, weight deacy loss is 7.20, training accuracy is 0.281250, time 56.649 samples/sec
+epoch 8, total_step 131680, total loss is 19.73 , inference loss is 12.53, weight deacy loss is 7.20, training accuracy is 0.156250, time 55.922 samples/sec
+epoch 8, total_step 131700, total loss is 14.53 , inference loss is 7.33, weight deacy loss is 7.20, training accuracy is 0.437500, time 56.997 samples/sec
+epoch 8, total_step 131720, total loss is 16.72 , inference loss is 9.52, weight deacy loss is 7.20, training accuracy is 0.343750, time 57.009 samples/sec
+epoch 8, total_step 131740, total loss is 15.60 , inference loss is 8.40, weight deacy loss is 7.20, training accuracy is 0.437500, time 55.731 samples/sec
+epoch 8, total_step 131760, total loss is 15.04 , inference loss is 7.84, weight deacy loss is 7.20, training accuracy is 0.531250, time 56.166 samples/sec
+epoch 8, total_step 131780, total loss is 15.48 , inference loss is 8.28, weight deacy loss is 7.20, training accuracy is 0.531250, time 56.654 samples/sec
+epoch 8, total_step 131800, total loss is 17.79 , inference loss is 10.59, weight deacy loss is 7.20, training accuracy is 0.500000, time 55.335 samples/sec
+epoch 8, total_step 131820, total loss is 14.56 , inference loss is 7.37, weight deacy loss is 7.20, training accuracy is 0.312500, time 57.230 samples/sec
+epoch 8, total_step 131840, total loss is 15.07 , inference loss is 7.87, weight deacy loss is 7.20, training accuracy is 0.281250, time 56.314 samples/sec
+epoch 8, total_step 131860, total loss is 15.41 , inference loss is 8.21, weight deacy loss is 7.20, training accuracy is 0.312500, time 55.808 samples/sec
+epoch 8, total_step 131880, total loss is 18.00 , inference loss is 10.81, weight deacy loss is 7.20, training accuracy is 0.312500, time 56.430 samples/sec
+epoch 8, total_step 131900, total loss is 15.87 , inference loss is 8.67, weight deacy loss is 7.20, training accuracy is 0.312500, time 56.651 samples/sec
+epoch 8, total_step 131920, total loss is 13.87 , inference loss is 6.67, weight deacy loss is 7.20, training accuracy is 0.406250, time 56.207 samples/sec
+epoch 8, total_step 131940, total loss is 14.56 , inference loss is 7.36, weight deacy loss is 7.20, training accuracy is 0.500000, time 56.161 samples/sec
+epoch 8, total_step 131960, total loss is 17.59 , inference loss is 10.40, weight deacy loss is 7.20, training accuracy is 0.406250, time 56.571 samples/sec
+epoch 8, total_step 131980, total loss is 14.78 , inference loss is 7.58, weight deacy loss is 7.20, training accuracy is 0.437500, time 56.390 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.89879600000009
+best_threshold_index 115 0.9875925925925926
+best_threshold_index 114 0.9862962962962963
+best_threshold_index 115 0.987037037037037
+best_threshold_index 116 0.9874074074074074
+best_threshold_index 117 0.9866666666666667
+best_threshold_index 115 0.9864814814814815
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9862962962962963
+best_threshold_index 117 0.9859259259259259
+best_threshold_index 114 0.9861111111111112
+[lfw][132000]XNorm: 23.446644
+[lfw][132000]Accuracy-Flip: 0.98583+-0.00484
+testing verification..
+(14000, 512)
+infer time 86.77709699999994
+best_threshold_index 113 0.9850793650793651
+best_threshold_index 120 0.9846031746031746
+best_threshold_index 120 0.9836507936507937
+best_threshold_index 120 0.9838095238095238
+best_threshold_index 120 0.9846031746031746
+best_threshold_index 120 0.9839682539682539
+best_threshold_index 120 0.9844444444444445
+best_threshold_index 120 0.9834920634920635
+best_threshold_index 120 0.983015873015873
+best_threshold_index 120 0.984920634920635
+[cfp_ff][132000]XNorm: 22.515348
+[cfp_ff][132000]Accuracy-Flip: 0.98343+-0.00674
+testing verification..
+(14000, 512)
+infer time 86.54306500000014
+best_threshold_index 141 0.9171428571428571
+best_threshold_index 133 0.9176190476190477
+best_threshold_index 133 0.9166666666666666
+best_threshold_index 133 0.9180952380952381
+best_threshold_index 133 0.9201587301587302
+best_threshold_index 139 0.9157142857142857
+best_threshold_index 133 0.915079365079365
+best_threshold_index 141 0.916984126984127
+best_threshold_index 133 0.9177777777777778
+best_threshold_index 133 0.9171428571428571
+[cfp_fp][132000]XNorm: 22.417045
+[cfp_fp][132000]Accuracy-Flip: 0.91429+-0.01143
+testing verification..
+(12000, 512)
+infer time 72.59680300000008
+best_threshold_index 136 0.9037037037037037
+best_threshold_index 134 0.9011111111111111
+best_threshold_index 136 0.9033333333333333
+best_threshold_index 134 0.9059259259259259
+best_threshold_index 134 0.9075925925925926
+best_threshold_index 136 0.9012962962962963
+best_threshold_index 133 0.904074074074074
+best_threshold_index 136 0.9087037037037037
+best_threshold_index 134 0.9061111111111111
+best_threshold_index 134 0.9053703703703704
+[agedb_30][132000]XNorm: 23.671532
+[agedb_30][132000]Accuracy-Flip: 0.90133+-0.02109
+test accuracy is: 0.9858333333333335
+epoch 8, total_step 132000, total loss is 12.55 , inference loss is 5.35, weight deacy loss is 7.20, training accuracy is 0.406250, time 53.758 samples/sec
+epoch 8, total_step 132020, total loss is 12.43 , inference loss is 5.23, weight deacy loss is 7.19, training accuracy is 0.281250, time 56.133 samples/sec
+epoch 8, total_step 132040, total loss is 17.43 , inference loss is 10.24, weight deacy loss is 7.19, training accuracy is 0.281250, time 57.455 samples/sec
+epoch 8, total_step 132060, total loss is 16.91 , inference loss is 9.72, weight deacy loss is 7.19, training accuracy is 0.406250, time 56.218 samples/sec
+epoch 8, total_step 132080, total loss is 15.95 , inference loss is 8.76, weight deacy loss is 7.19, training accuracy is 0.531250, time 56.591 samples/sec
+epoch 8, total_step 132100, total loss is 18.75 , inference loss is 11.55, weight deacy loss is 7.19, training accuracy is 0.312500, time 56.251 samples/sec
+epoch 8, total_step 132120, total loss is 16.72 , inference loss is 9.52, weight deacy loss is 7.19, training accuracy is 0.312500, time 56.947 samples/sec
+epoch 8, total_step 132140, total loss is 15.77 , inference loss is 8.57, weight deacy loss is 7.19, training accuracy is 0.343750, time 56.807 samples/sec
+epoch 8, total_step 132160, total loss is 14.63 , inference loss is 7.44, weight deacy loss is 7.19, training accuracy is 0.500000, time 57.249 samples/sec
+epoch 8, total_step 132180, total loss is 12.90 , inference loss is 5.71, weight deacy loss is 7.19, training accuracy is 0.406250, time 57.049 samples/sec
+epoch 8, total_step 132200, total loss is 14.32 , inference loss is 7.13, weight deacy loss is 7.19, training accuracy is 0.562500, time 56.078 samples/sec
+epoch 8, total_step 132220, total loss is 13.77 , inference loss is 6.58, weight deacy loss is 7.19, training accuracy is 0.437500, time 56.871 samples/sec
+epoch 8, total_step 132240, total loss is 15.57 , inference loss is 8.38, weight deacy loss is 7.19, training accuracy is 0.312500, time 56.063 samples/sec
+epoch 8, total_step 132260, total loss is 19.15 , inference loss is 11.96, weight deacy loss is 7.19, training accuracy is 0.406250, time 56.386 samples/sec
+epoch 8, total_step 132280, total loss is 13.52 , inference loss is 6.33, weight deacy loss is 7.19, training accuracy is 0.562500, time 56.236 samples/sec
+epoch 8, total_step 132300, total loss is 15.24 , inference loss is 8.05, weight deacy loss is 7.19, training accuracy is 0.437500, time 56.282 samples/sec
+epoch 8, total_step 132320, total loss is 16.75 , inference loss is 9.56, weight deacy loss is 7.19, training accuracy is 0.406250, time 56.510 samples/sec
+epoch 8, total_step 132340, total loss is 13.37 , inference loss is 6.19, weight deacy loss is 7.19, training accuracy is 0.500000, time 56.894 samples/sec
+epoch 8, total_step 132360, total loss is 16.32 , inference loss is 9.14, weight deacy loss is 7.19, training accuracy is 0.250000, time 57.040 samples/sec
+epoch 8, total_step 132380, total loss is 17.97 , inference loss is 10.78, weight deacy loss is 7.19, training accuracy is 0.187500, time 55.929 samples/sec
+epoch 8, total_step 132400, total loss is 15.05 , inference loss is 7.87, weight deacy loss is 7.19, training accuracy is 0.468750, time 57.082 samples/sec
+epoch 8, total_step 132420, total loss is 14.73 , inference loss is 7.54, weight deacy loss is 7.19, training accuracy is 0.312500, time 56.899 samples/sec
+epoch 8, total_step 132440, total loss is 19.71 , inference loss is 12.52, weight deacy loss is 7.19, training accuracy is 0.218750, time 57.097 samples/sec
+epoch 8, total_step 132460, total loss is 17.22 , inference loss is 10.04, weight deacy loss is 7.19, training accuracy is 0.312500, time 56.865 samples/sec
+epoch 8, total_step 132480, total loss is 18.73 , inference loss is 11.55, weight deacy loss is 7.19, training accuracy is 0.187500, time 56.508 samples/sec
+epoch 8, total_step 132500, total loss is 15.80 , inference loss is 8.62, weight deacy loss is 7.18, training accuracy is 0.343750, time 57.465 samples/sec
+epoch 8, total_step 132520, total loss is 17.37 , inference loss is 10.19, weight deacy loss is 7.18, training accuracy is 0.343750, time 56.266 samples/sec
+epoch 8, total_step 132540, total loss is 16.30 , inference loss is 9.12, weight deacy loss is 7.18, training accuracy is 0.375000, time 56.623 samples/sec
+epoch 8, total_step 132560, total loss is 14.63 , inference loss is 7.44, weight deacy loss is 7.18, training accuracy is 0.406250, time 56.344 samples/sec
+epoch 8, total_step 132580, total loss is 14.87 , inference loss is 7.68, weight deacy loss is 7.18, training accuracy is 0.375000, time 56.846 samples/sec
+epoch 8, total_step 132600, total loss is 13.46 , inference loss is 6.28, weight deacy loss is 7.18, training accuracy is 0.406250, time 56.439 samples/sec
+epoch 8, total_step 132620, total loss is 15.17 , inference loss is 7.99, weight deacy loss is 7.18, training accuracy is 0.406250, time 54.801 samples/sec
+epoch 8, total_step 132640, total loss is 13.71 , inference loss is 6.53, weight deacy loss is 7.18, training accuracy is 0.500000, time 56.061 samples/sec
+epoch 8, total_step 132660, total loss is 17.17 , inference loss is 9.99, weight deacy loss is 7.18, training accuracy is 0.406250, time 56.442 samples/sec
+epoch 8, total_step 132680, total loss is 16.10 , inference loss is 8.92, weight deacy loss is 7.18, training accuracy is 0.437500, time 56.574 samples/sec
+epoch 8, total_step 132700, total loss is 13.11 , inference loss is 5.93, weight deacy loss is 7.18, training accuracy is 0.531250, time 56.403 samples/sec
+epoch 8, total_step 132720, total loss is 14.43 , inference loss is 7.25, weight deacy loss is 7.18, training accuracy is 0.437500, time 55.748 samples/sec
+epoch 8, total_step 132740, total loss is 15.64 , inference loss is 8.46, weight deacy loss is 7.18, training accuracy is 0.281250, time 55.059 samples/sec
+epoch 8, total_step 132760, total loss is 18.30 , inference loss is 11.12, weight deacy loss is 7.18, training accuracy is 0.343750, time 56.687 samples/sec
+epoch 8, total_step 132780, total loss is 17.38 , inference loss is 10.20, weight deacy loss is 7.18, training accuracy is 0.375000, time 56.609 samples/sec
+epoch 8, total_step 132800, total loss is 14.40 , inference loss is 7.22, weight deacy loss is 7.18, training accuracy is 0.468750, time 56.862 samples/sec
+epoch 8, total_step 132820, total loss is 17.92 , inference loss is 10.74, weight deacy loss is 7.18, training accuracy is 0.406250, time 56.087 samples/sec
+epoch 8, total_step 132840, total loss is 15.71 , inference loss is 8.53, weight deacy loss is 7.18, training accuracy is 0.375000, time 56.478 samples/sec
+epoch 8, total_step 132860, total loss is 21.89 , inference loss is 14.71, weight deacy loss is 7.18, training accuracy is 0.125000, time 56.156 samples/sec
+epoch 8, total_step 132880, total loss is 13.87 , inference loss is 6.69, weight deacy loss is 7.18, training accuracy is 0.343750, time 56.736 samples/sec
+epoch 8, total_step 132900, total loss is 11.56 , inference loss is 4.39, weight deacy loss is 7.18, training accuracy is 0.406250, time 56.006 samples/sec
+epoch 8, total_step 132920, total loss is 17.17 , inference loss is 10.00, weight deacy loss is 7.17, training accuracy is 0.156250, time 55.366 samples/sec
+epoch 8, total_step 132940, total loss is 12.90 , inference loss is 5.73, weight deacy loss is 7.17, training accuracy is 0.500000, time 56.337 samples/sec
+epoch 8, total_step 132960, total loss is 16.57 , inference loss is 9.40, weight deacy loss is 7.17, training accuracy is 0.437500, time 56.505 samples/sec
+epoch 8, total_step 132980, total loss is 15.86 , inference loss is 8.69, weight deacy loss is 7.17, training accuracy is 0.312500, time 56.572 samples/sec
+epoch 8, total_step 133000, total loss is 13.10 , inference loss is 5.93, weight deacy loss is 7.17, training accuracy is 0.406250, time 56.797 samples/sec
+epoch 8, total_step 133020, total loss is 17.34 , inference loss is 10.17, weight deacy loss is 7.17, training accuracy is 0.281250, time 54.942 samples/sec
+epoch 8, total_step 133040, total loss is 11.50 , inference loss is 4.32, weight deacy loss is 7.17, training accuracy is 0.437500, time 56.573 samples/sec
+epoch 8, total_step 133060, total loss is 13.77 , inference loss is 6.60, weight deacy loss is 7.17, training accuracy is 0.500000, time 56.266 samples/sec
+epoch 8, total_step 133080, total loss is 15.35 , inference loss is 8.18, weight deacy loss is 7.17, training accuracy is 0.406250, time 56.249 samples/sec
+epoch 8, total_step 133100, total loss is 17.15 , inference loss is 9.98, weight deacy loss is 7.17, training accuracy is 0.250000, time 56.397 samples/sec
+epoch 8, total_step 133120, total loss is 11.69 , inference loss is 4.52, weight deacy loss is 7.17, training accuracy is 0.468750, time 56.312 samples/sec
+epoch 8, total_step 133140, total loss is 16.11 , inference loss is 8.94, weight deacy loss is 7.17, training accuracy is 0.468750, time 57.160 samples/sec
+epoch 8, total_step 133160, total loss is 14.16 , inference loss is 6.99, weight deacy loss is 7.17, training accuracy is 0.406250, time 56.124 samples/sec
+epoch 8, total_step 133180, total loss is 13.16 , inference loss is 5.99, weight deacy loss is 7.17, training accuracy is 0.500000, time 57.068 samples/sec
+epoch 8, total_step 133200, total loss is 20.89 , inference loss is 13.72, weight deacy loss is 7.17, training accuracy is 0.281250, time 55.242 samples/sec
+epoch 8, total_step 133220, total loss is 12.25 , inference loss is 5.08, weight deacy loss is 7.17, training accuracy is 0.468750, time 56.476 samples/sec
+epoch 8, total_step 133240, total loss is 15.12 , inference loss is 7.96, weight deacy loss is 7.17, training accuracy is 0.218750, time 55.088 samples/sec
+epoch 8, total_step 133260, total loss is 16.58 , inference loss is 9.41, weight deacy loss is 7.17, training accuracy is 0.343750, time 56.807 samples/sec
+epoch 8, total_step 133280, total loss is 15.53 , inference loss is 8.36, weight deacy loss is 7.17, training accuracy is 0.250000, time 56.869 samples/sec
+epoch 8, total_step 133300, total loss is 17.10 , inference loss is 9.93, weight deacy loss is 7.17, training accuracy is 0.343750, time 56.545 samples/sec
+epoch 8, total_step 133320, total loss is 14.16 , inference loss is 6.99, weight deacy loss is 7.17, training accuracy is 0.312500, time 56.006 samples/sec
+epoch 8, total_step 133340, total loss is 17.09 , inference loss is 9.92, weight deacy loss is 7.17, training accuracy is 0.406250, time 56.580 samples/sec
+epoch 8, total_step 133360, total loss is 14.55 , inference loss is 7.38, weight deacy loss is 7.17, training accuracy is 0.500000, time 56.256 samples/sec
+epoch 8, total_step 133380, total loss is 15.38 , inference loss is 8.22, weight deacy loss is 7.16, training accuracy is 0.531250, time 56.799 samples/sec
+epoch 8, total_step 133400, total loss is 20.78 , inference loss is 13.62, weight deacy loss is 7.16, training accuracy is 0.406250, time 56.907 samples/sec
+epoch 8, total_step 133420, total loss is 16.75 , inference loss is 9.59, weight deacy loss is 7.16, training accuracy is 0.312500, time 56.087 samples/sec
+epoch 8, total_step 133440, total loss is 16.00 , inference loss is 8.83, weight deacy loss is 7.16, training accuracy is 0.281250, time 56.743 samples/sec
+epoch 8, total_step 133460, total loss is 16.93 , inference loss is 9.77, weight deacy loss is 7.16, training accuracy is 0.468750, time 56.429 samples/sec
+epoch 8, total_step 133480, total loss is 12.46 , inference loss is 5.30, weight deacy loss is 7.16, training accuracy is 0.406250, time 55.427 samples/sec
+epoch 8, total_step 133500, total loss is 17.87 , inference loss is 10.71, weight deacy loss is 7.16, training accuracy is 0.312500, time 56.157 samples/sec
+epoch 8, total_step 133520, total loss is 19.22 , inference loss is 12.05, weight deacy loss is 7.16, training accuracy is 0.187500, time 54.644 samples/sec
+epoch 8, total_step 133540, total loss is 17.14 , inference loss is 9.98, weight deacy loss is 7.16, training accuracy is 0.343750, time 56.209 samples/sec
+epoch 8, total_step 133560, total loss is 12.52 , inference loss is 5.36, weight deacy loss is 7.16, training accuracy is 0.531250, time 56.431 samples/sec
+epoch 8, total_step 133580, total loss is 12.74 , inference loss is 5.58, weight deacy loss is 7.16, training accuracy is 0.343750, time 56.444 samples/sec
+epoch 8, total_step 133600, total loss is 14.62 , inference loss is 7.46, weight deacy loss is 7.16, training accuracy is 0.406250, time 57.136 samples/sec
+epoch 8, total_step 133620, total loss is 16.34 , inference loss is 9.18, weight deacy loss is 7.16, training accuracy is 0.406250, time 55.855 samples/sec
+epoch 8, total_step 133640, total loss is 18.03 , inference loss is 10.87, weight deacy loss is 7.16, training accuracy is 0.312500, time 56.229 samples/sec
+epoch 8, total_step 133660, total loss is 17.77 , inference loss is 10.61, weight deacy loss is 7.16, training accuracy is 0.343750, time 57.054 samples/sec
+epoch 8, total_step 133680, total loss is 15.02 , inference loss is 7.86, weight deacy loss is 7.16, training accuracy is 0.437500, time 57.169 samples/sec
+epoch 8, total_step 133700, total loss is 14.60 , inference loss is 7.44, weight deacy loss is 7.16, training accuracy is 0.375000, time 56.907 samples/sec
+epoch 8, total_step 133720, total loss is 13.63 , inference loss is 6.47, weight deacy loss is 7.16, training accuracy is 0.437500, time 56.867 samples/sec
+epoch 8, total_step 133740, total loss is 15.17 , inference loss is 8.02, weight deacy loss is 7.16, training accuracy is 0.343750, time 55.637 samples/sec
+epoch 8, total_step 133760, total loss is 16.88 , inference loss is 9.72, weight deacy loss is 7.16, training accuracy is 0.312500, time 56.038 samples/sec
+epoch 8, total_step 133780, total loss is 17.55 , inference loss is 10.39, weight deacy loss is 7.16, training accuracy is 0.406250, time 56.759 samples/sec
+epoch 8, total_step 133800, total loss is 17.19 , inference loss is 10.03, weight deacy loss is 7.16, training accuracy is 0.187500, time 56.956 samples/sec
+epoch 8, total_step 133820, total loss is 14.40 , inference loss is 7.25, weight deacy loss is 7.16, training accuracy is 0.437500, time 56.768 samples/sec
+epoch 8, total_step 133840, total loss is 11.86 , inference loss is 4.70, weight deacy loss is 7.16, training accuracy is 0.468750, time 57.320 samples/sec
+epoch 8, total_step 133860, total loss is 15.57 , inference loss is 8.42, weight deacy loss is 7.15, training accuracy is 0.281250, time 56.889 samples/sec
+epoch 8, total_step 133880, total loss is 18.37 , inference loss is 11.22, weight deacy loss is 7.15, training accuracy is 0.343750, time 56.484 samples/sec
+epoch 8, total_step 133900, total loss is 20.40 , inference loss is 13.25, weight deacy loss is 7.15, training accuracy is 0.218750, time 57.010 samples/sec
+epoch 8, total_step 133920, total loss is 14.95 , inference loss is 7.79, weight deacy loss is 7.15, training accuracy is 0.437500, time 56.414 samples/sec
+epoch 8, total_step 133940, total loss is 17.43 , inference loss is 10.28, weight deacy loss is 7.15, training accuracy is 0.343750, time 56.418 samples/sec
+epoch 8, total_step 133960, total loss is 17.99 , inference loss is 10.83, weight deacy loss is 7.15, training accuracy is 0.343750, time 56.564 samples/sec
+epoch 8, total_step 133980, total loss is 17.56 , inference loss is 10.41, weight deacy loss is 7.15, training accuracy is 0.343750, time 56.692 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.16299100000008
+best_threshold_index 111 0.9861111111111112
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 110 0.9855555555555555
+best_threshold_index 110 0.9862962962962963
+best_threshold_index 110 0.9859259259259259
+best_threshold_index 111 0.9846296296296296
+best_threshold_index 112 0.9855555555555555
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 110 0.9846296296296296
+best_threshold_index 111 0.9846296296296296
+[lfw][134000]XNorm: 21.987046
+[lfw][134000]Accuracy-Flip: 0.98467+-0.00526
+testing verification..
+(14000, 512)
+infer time 86.03707599999991
+best_threshold_index 115 0.986984126984127
+best_threshold_index 115 0.986984126984127
+best_threshold_index 115 0.9861904761904762
+best_threshold_index 115 0.9865079365079366
+best_threshold_index 114 0.9873015873015873
+best_threshold_index 115 0.9861904761904762
+best_threshold_index 115 0.9871428571428571
+best_threshold_index 115 0.9858730158730159
+best_threshold_index 115 0.986031746031746
+best_threshold_index 119 0.9866666666666667
+[cfp_ff][134000]XNorm: 21.715415
+[cfp_ff][134000]Accuracy-Flip: 0.98557+-0.00525
+testing verification..
+(14000, 512)
+infer time 85.07648500000003
+best_threshold_index 133 0.9261904761904762
+best_threshold_index 134 0.9265079365079365
+best_threshold_index 139 0.9249206349206349
+best_threshold_index 133 0.9258730158730158
+best_threshold_index 134 0.9285714285714286
+best_threshold_index 136 0.9242857142857143
+best_threshold_index 134 0.9249206349206349
+best_threshold_index 136 0.9242857142857143
+best_threshold_index 139 0.9257142857142857
+best_threshold_index 134 0.926031746031746
+[cfp_fp][134000]XNorm: 21.494457
+[cfp_fp][134000]Accuracy-Flip: 0.92271+-0.01072
+testing verification..
+(12000, 512)
+infer time 73.96677300000003
+best_threshold_index 142 0.8996296296296297
+best_threshold_index 141 0.8975925925925926
+best_threshold_index 137 0.8975925925925926
+best_threshold_index 141 0.8994444444444445
+best_threshold_index 141 0.9038888888888889
+best_threshold_index 138 0.8968518518518519
+best_threshold_index 138 0.8985185185185185
+best_threshold_index 141 0.9046296296296297
+best_threshold_index 141 0.9018518518518519
+best_threshold_index 142 0.9011111111111111
+[agedb_30][134000]XNorm: 22.368805
+[agedb_30][134000]Accuracy-Flip: 0.89517+-0.02019
+test accuracy is: 0.9846666666666668
+epoch 8, total_step 134000, total loss is 17.55 , inference loss is 10.39, weight deacy loss is 7.15, training accuracy is 0.375000, time 53.859 samples/sec
+epoch 8, total_step 134020, total loss is 14.27 , inference loss is 7.12, weight deacy loss is 7.15, training accuracy is 0.562500, time 56.867 samples/sec
+epoch 8, total_step 134040, total loss is 15.85 , inference loss is 8.70, weight deacy loss is 7.15, training accuracy is 0.375000, time 56.982 samples/sec
+epoch 8, total_step 134060, total loss is 16.63 , inference loss is 9.48, weight deacy loss is 7.15, training accuracy is 0.437500, time 56.918 samples/sec
+epoch 8, total_step 134080, total loss is 12.71 , inference loss is 5.56, weight deacy loss is 7.15, training accuracy is 0.406250, time 56.696 samples/sec
+epoch 8, total_step 134100, total loss is 15.90 , inference loss is 8.75, weight deacy loss is 7.15, training accuracy is 0.500000, time 55.959 samples/sec
+epoch 8, total_step 134120, total loss is 16.41 , inference loss is 9.26, weight deacy loss is 7.15, training accuracy is 0.343750, time 56.688 samples/sec
+epoch 8, total_step 134140, total loss is 17.43 , inference loss is 10.28, weight deacy loss is 7.15, training accuracy is 0.343750, time 55.618 samples/sec
+epoch 8, total_step 134160, total loss is 14.45 , inference loss is 7.30, weight deacy loss is 7.15, training accuracy is 0.343750, time 56.096 samples/sec
+epoch 8, total_step 134180, total loss is 16.06 , inference loss is 8.91, weight deacy loss is 7.15, training accuracy is 0.375000, time 56.235 samples/sec
+epoch 8, total_step 134200, total loss is 17.45 , inference loss is 10.30, weight deacy loss is 7.15, training accuracy is 0.250000, time 56.210 samples/sec
+epoch 8, total_step 134220, total loss is 19.95 , inference loss is 12.81, weight deacy loss is 7.15, training accuracy is 0.343750, time 55.806 samples/sec
+epoch 8, total_step 134240, total loss is 16.70 , inference loss is 9.55, weight deacy loss is 7.15, training accuracy is 0.281250, time 56.523 samples/sec
+epoch 8, total_step 134260, total loss is 18.29 , inference loss is 11.14, weight deacy loss is 7.15, training accuracy is 0.343750, time 57.113 samples/sec
+epoch 8, total_step 134280, total loss is 20.27 , inference loss is 13.13, weight deacy loss is 7.15, training accuracy is 0.218750, time 56.303 samples/sec
+epoch 8, total_step 134300, total loss is 18.99 , inference loss is 11.84, weight deacy loss is 7.15, training accuracy is 0.312500, time 55.449 samples/sec
+epoch 8, total_step 134320, total loss is 14.50 , inference loss is 7.36, weight deacy loss is 7.14, training accuracy is 0.281250, time 56.728 samples/sec
+epoch 8, total_step 134340, total loss is 17.32 , inference loss is 10.18, weight deacy loss is 7.14, training accuracy is 0.281250, time 57.118 samples/sec
+epoch 8, total_step 134360, total loss is 20.11 , inference loss is 12.97, weight deacy loss is 7.14, training accuracy is 0.187500, time 56.625 samples/sec
+epoch 8, total_step 134380, total loss is 17.80 , inference loss is 10.66, weight deacy loss is 7.14, training accuracy is 0.187500, time 55.413 samples/sec
+epoch 8, total_step 134400, total loss is 13.78 , inference loss is 6.63, weight deacy loss is 7.14, training accuracy is 0.343750, time 56.649 samples/sec
+epoch 8, total_step 134420, total loss is 16.82 , inference loss is 9.68, weight deacy loss is 7.14, training accuracy is 0.187500, time 56.272 samples/sec
+epoch 8, total_step 134440, total loss is 15.64 , inference loss is 8.49, weight deacy loss is 7.14, training accuracy is 0.375000, time 55.984 samples/sec
+epoch 8, total_step 134460, total loss is 17.82 , inference loss is 10.67, weight deacy loss is 7.14, training accuracy is 0.250000, time 56.598 samples/sec
+epoch 8, total_step 134480, total loss is 19.93 , inference loss is 12.79, weight deacy loss is 7.14, training accuracy is 0.218750, time 56.714 samples/sec
+epoch 8, total_step 134500, total loss is 14.63 , inference loss is 7.49, weight deacy loss is 7.14, training accuracy is 0.343750, time 56.858 samples/sec
+epoch 8, total_step 134520, total loss is 15.36 , inference loss is 8.21, weight deacy loss is 7.14, training accuracy is 0.343750, time 56.430 samples/sec
+epoch 8, total_step 134540, total loss is 16.77 , inference loss is 9.63, weight deacy loss is 7.14, training accuracy is 0.250000, time 56.743 samples/sec
+epoch 8, total_step 134560, total loss is 15.12 , inference loss is 7.98, weight deacy loss is 7.14, training accuracy is 0.312500, time 55.988 samples/sec
+epoch 8, total_step 134580, total loss is 15.47 , inference loss is 8.33, weight deacy loss is 7.14, training accuracy is 0.312500, time 56.827 samples/sec
+epoch 8, total_step 134600, total loss is 20.26 , inference loss is 13.13, weight deacy loss is 7.14, training accuracy is 0.281250, time 56.647 samples/sec
+epoch 8, total_step 134620, total loss is 19.28 , inference loss is 12.15, weight deacy loss is 7.14, training accuracy is 0.312500, time 57.630 samples/sec
+epoch 8, total_step 134640, total loss is 14.42 , inference loss is 7.28, weight deacy loss is 7.14, training accuracy is 0.437500, time 56.783 samples/sec
+epoch 8, total_step 134660, total loss is 17.09 , inference loss is 9.96, weight deacy loss is 7.14, training accuracy is 0.250000, time 56.780 samples/sec
+epoch 8, total_step 134680, total loss is 14.25 , inference loss is 7.11, weight deacy loss is 7.14, training accuracy is 0.437500, time 56.181 samples/sec
+epoch 8, total_step 134700, total loss is 16.24 , inference loss is 9.11, weight deacy loss is 7.14, training accuracy is 0.312500, time 56.594 samples/sec
+epoch 8, total_step 134720, total loss is 18.97 , inference loss is 11.83, weight deacy loss is 7.14, training accuracy is 0.218750, time 56.091 samples/sec
+epoch 8, total_step 134740, total loss is 16.20 , inference loss is 9.06, weight deacy loss is 7.14, training accuracy is 0.500000, time 55.850 samples/sec
+epoch 8, total_step 134760, total loss is 15.28 , inference loss is 8.15, weight deacy loss is 7.14, training accuracy is 0.375000, time 56.735 samples/sec
+epoch 8, total_step 134780, total loss is 19.50 , inference loss is 12.36, weight deacy loss is 7.13, training accuracy is 0.187500, time 57.992 samples/sec
+epoch 8, total_step 134800, total loss is 17.83 , inference loss is 10.69, weight deacy loss is 7.13, training accuracy is 0.312500, time 56.727 samples/sec
+epoch 8, total_step 134820, total loss is 16.36 , inference loss is 9.23, weight deacy loss is 7.13, training accuracy is 0.406250, time 57.570 samples/sec
+epoch 8, total_step 134840, total loss is 14.07 , inference loss is 6.94, weight deacy loss is 7.13, training accuracy is 0.531250, time 56.757 samples/sec
+epoch 8, total_step 134860, total loss is 17.49 , inference loss is 10.36, weight deacy loss is 7.13, training accuracy is 0.312500, time 56.357 samples/sec
+epoch 8, total_step 134880, total loss is 18.60 , inference loss is 11.47, weight deacy loss is 7.13, training accuracy is 0.375000, time 57.063 samples/sec
+epoch 8, total_step 134900, total loss is 17.58 , inference loss is 10.45, weight deacy loss is 7.13, training accuracy is 0.406250, time 56.883 samples/sec
+epoch 8, total_step 134920, total loss is 15.62 , inference loss is 8.49, weight deacy loss is 7.13, training accuracy is 0.375000, time 56.552 samples/sec
+epoch 8, total_step 134940, total loss is 15.14 , inference loss is 8.01, weight deacy loss is 7.13, training accuracy is 0.375000, time 56.322 samples/sec
+epoch 8, total_step 134960, total loss is 18.54 , inference loss is 11.41, weight deacy loss is 7.13, training accuracy is 0.468750, time 56.483 samples/sec
+epoch 8, total_step 134980, total loss is 19.39 , inference loss is 12.26, weight deacy loss is 7.13, training accuracy is 0.218750, time 57.031 samples/sec
+epoch 8, total_step 135000, total loss is 17.53 , inference loss is 10.40, weight deacy loss is 7.13, training accuracy is 0.281250, time 56.283 samples/sec
+epoch 8, total_step 135020, total loss is 16.14 , inference loss is 9.01, weight deacy loss is 7.13, training accuracy is 0.218750, time 56.792 samples/sec
+epoch 8, total_step 135040, total loss is 14.85 , inference loss is 7.72, weight deacy loss is 7.13, training accuracy is 0.406250, time 56.630 samples/sec
+epoch 8, total_step 135060, total loss is 20.43 , inference loss is 13.30, weight deacy loss is 7.13, training accuracy is 0.156250, time 56.807 samples/sec
+epoch 8, total_step 135080, total loss is 19.53 , inference loss is 12.41, weight deacy loss is 7.13, training accuracy is 0.218750, time 56.462 samples/sec
+epoch 8, total_step 135100, total loss is 17.86 , inference loss is 10.73, weight deacy loss is 7.13, training accuracy is 0.187500, time 56.178 samples/sec
+epoch 8, total_step 135120, total loss is 15.21 , inference loss is 8.09, weight deacy loss is 7.13, training accuracy is 0.406250, time 56.695 samples/sec
+epoch 8, total_step 135140, total loss is 16.67 , inference loss is 9.54, weight deacy loss is 7.13, training accuracy is 0.375000, time 56.930 samples/sec
+epoch 8, total_step 135160, total loss is 15.52 , inference loss is 8.40, weight deacy loss is 7.13, training accuracy is 0.343750, time 57.242 samples/sec
+epoch 8, total_step 135180, total loss is 16.08 , inference loss is 8.95, weight deacy loss is 7.13, training accuracy is 0.187500, time 56.565 samples/sec
+epoch 8, total_step 135200, total loss is 16.63 , inference loss is 9.50, weight deacy loss is 7.13, training accuracy is 0.343750, time 57.408 samples/sec
+epoch 8, total_step 135220, total loss is 17.48 , inference loss is 10.36, weight deacy loss is 7.12, training accuracy is 0.281250, time 56.783 samples/sec
+epoch 8, total_step 135240, total loss is 13.49 , inference loss is 6.36, weight deacy loss is 7.12, training accuracy is 0.468750, time 55.932 samples/sec
+epoch 8, total_step 135260, total loss is 18.19 , inference loss is 11.07, weight deacy loss is 7.12, training accuracy is 0.250000, time 55.821 samples/sec
+epoch 8, total_step 135280, total loss is 15.27 , inference loss is 8.15, weight deacy loss is 7.12, training accuracy is 0.437500, time 55.140 samples/sec
+epoch 8, total_step 135300, total loss is 13.55 , inference loss is 6.43, weight deacy loss is 7.12, training accuracy is 0.406250, time 56.815 samples/sec
+epoch 8, total_step 135320, total loss is 16.26 , inference loss is 9.14, weight deacy loss is 7.12, training accuracy is 0.250000, time 56.653 samples/sec
+epoch 8, total_step 135340, total loss is 17.67 , inference loss is 10.55, weight deacy loss is 7.12, training accuracy is 0.312500, time 55.836 samples/sec
+epoch 8, total_step 135360, total loss is 16.76 , inference loss is 9.64, weight deacy loss is 7.12, training accuracy is 0.437500, time 55.230 samples/sec
+epoch 8, total_step 135380, total loss is 16.57 , inference loss is 9.45, weight deacy loss is 7.12, training accuracy is 0.281250, time 56.755 samples/sec
+epoch 8, total_step 135400, total loss is 16.11 , inference loss is 8.99, weight deacy loss is 7.12, training accuracy is 0.281250, time 56.932 samples/sec
+epoch 8, total_step 135420, total loss is 19.53 , inference loss is 12.41, weight deacy loss is 7.12, training accuracy is 0.218750, time 56.518 samples/sec
+epoch 8, total_step 135440, total loss is 15.92 , inference loss is 8.80, weight deacy loss is 7.12, training accuracy is 0.468750, time 55.380 samples/sec
+epoch 8, total_step 135460, total loss is 17.02 , inference loss is 9.91, weight deacy loss is 7.12, training accuracy is 0.312500, time 56.568 samples/sec
+epoch 8, total_step 135480, total loss is 13.73 , inference loss is 6.62, weight deacy loss is 7.12, training accuracy is 0.375000, time 56.712 samples/sec
+epoch 8, total_step 135500, total loss is 15.33 , inference loss is 8.22, weight deacy loss is 7.12, training accuracy is 0.343750, time 56.640 samples/sec
+epoch 8, total_step 135520, total loss is 19.07 , inference loss is 11.95, weight deacy loss is 7.12, training accuracy is 0.250000, time 56.374 samples/sec
+epoch 8, total_step 135540, total loss is 19.20 , inference loss is 12.08, weight deacy loss is 7.12, training accuracy is 0.218750, time 55.713 samples/sec
+epoch 8, total_step 135560, total loss is 18.00 , inference loss is 10.88, weight deacy loss is 7.12, training accuracy is 0.500000, time 57.756 samples/sec
+epoch 8, total_step 135580, total loss is 14.61 , inference loss is 7.50, weight deacy loss is 7.12, training accuracy is 0.281250, time 55.899 samples/sec
+epoch 8, total_step 135600, total loss is 16.42 , inference loss is 9.31, weight deacy loss is 7.12, training accuracy is 0.250000, time 56.559 samples/sec
+epoch 8, total_step 135620, total loss is 16.10 , inference loss is 8.98, weight deacy loss is 7.12, training accuracy is 0.281250, time 56.124 samples/sec
+epoch 8, total_step 135640, total loss is 18.35 , inference loss is 11.23, weight deacy loss is 7.12, training accuracy is 0.375000, time 56.659 samples/sec
+epoch 8, total_step 135660, total loss is 15.22 , inference loss is 8.11, weight deacy loss is 7.11, training accuracy is 0.343750, time 56.888 samples/sec
+epoch 8, total_step 135680, total loss is 16.53 , inference loss is 9.42, weight deacy loss is 7.11, training accuracy is 0.375000, time 56.971 samples/sec
+epoch 8, total_step 135700, total loss is 18.95 , inference loss is 11.83, weight deacy loss is 7.11, training accuracy is 0.250000, time 56.331 samples/sec
+epoch 8, total_step 135720, total loss is 15.59 , inference loss is 8.48, weight deacy loss is 7.11, training accuracy is 0.250000, time 54.885 samples/sec
+epoch 8, total_step 135740, total loss is 18.08 , inference loss is 10.97, weight deacy loss is 7.11, training accuracy is 0.187500, time 56.728 samples/sec
+epoch 8, total_step 135760, total loss is 15.17 , inference loss is 8.06, weight deacy loss is 7.11, training accuracy is 0.343750, time 56.190 samples/sec
+epoch 8, total_step 135780, total loss is 15.85 , inference loss is 8.74, weight deacy loss is 7.11, training accuracy is 0.468750, time 56.378 samples/sec
+epoch 8, total_step 135800, total loss is 16.96 , inference loss is 9.85, weight deacy loss is 7.11, training accuracy is 0.250000, time 56.640 samples/sec
+epoch 8, total_step 135820, total loss is 16.19 , inference loss is 9.08, weight deacy loss is 7.11, training accuracy is 0.406250, time 56.253 samples/sec
+epoch 8, total_step 135840, total loss is 16.52 , inference loss is 9.41, weight deacy loss is 7.11, training accuracy is 0.375000, time 55.918 samples/sec
+epoch 8, total_step 135860, total loss is 18.77 , inference loss is 11.66, weight deacy loss is 7.11, training accuracy is 0.375000, time 56.633 samples/sec
+epoch 8, total_step 135880, total loss is 16.88 , inference loss is 9.77, weight deacy loss is 7.11, training accuracy is 0.281250, time 57.499 samples/sec
+epoch 8, total_step 135900, total loss is 12.65 , inference loss is 5.54, weight deacy loss is 7.11, training accuracy is 0.406250, time 56.053 samples/sec
+epoch 8, total_step 135920, total loss is 13.02 , inference loss is 5.91, weight deacy loss is 7.11, training accuracy is 0.406250, time 56.732 samples/sec
+epoch 8, total_step 135940, total loss is 18.69 , inference loss is 11.58, weight deacy loss is 7.11, training accuracy is 0.156250, time 56.144 samples/sec
+epoch 8, total_step 135960, total loss is 17.00 , inference loss is 9.89, weight deacy loss is 7.11, training accuracy is 0.312500, time 56.281 samples/sec
+epoch 8, total_step 135980, total loss is 13.02 , inference loss is 5.91, weight deacy loss is 7.11, training accuracy is 0.500000, time 57.494 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.96682899999996
+best_threshold_index 111 0.9861111111111112
+best_threshold_index 114 0.9840740740740741
+best_threshold_index 111 0.9859259259259259
+best_threshold_index 111 0.9851851851851852
+best_threshold_index 114 0.9857407407407407
+best_threshold_index 111 0.9837037037037037
+best_threshold_index 111 0.9846296296296296
+best_threshold_index 111 0.9840740740740741
+best_threshold_index 114 0.9833333333333333
+best_threshold_index 113 0.9835185185185186
+[lfw][136000]XNorm: 22.763571
+[lfw][136000]Accuracy-Flip: 0.98333+-0.00891
+testing verification..
+(14000, 512)
+infer time 86.17239
+best_threshold_index 121 0.9866666666666667
+best_threshold_index 121 0.986031746031746
+best_threshold_index 119 0.9866666666666667
+best_threshold_index 119 0.9852380952380952
+best_threshold_index 121 0.9873015873015873
+best_threshold_index 119 0.9857142857142858
+best_threshold_index 119 0.9865079365079366
+best_threshold_index 121 0.9858730158730159
+best_threshold_index 119 0.986031746031746
+best_threshold_index 119 0.9861904761904762
+[cfp_ff][136000]XNorm: 22.415945
+[cfp_ff][136000]Accuracy-Flip: 0.98543+-0.00498
+testing verification..
+(14000, 512)
+infer time 85.67269999999994
+best_threshold_index 137 0.9230158730158731
+best_threshold_index 133 0.9244444444444444
+best_threshold_index 137 0.923968253968254
+best_threshold_index 137 0.9265079365079365
+best_threshold_index 133 0.9274603174603174
+best_threshold_index 138 0.9230158730158731
+best_threshold_index 137 0.9219047619047619
+best_threshold_index 137 0.9233333333333333
+best_threshold_index 133 0.9252380952380952
+best_threshold_index 138 0.9263492063492064
+[cfp_fp][136000]XNorm: 21.743434
+[cfp_fp][136000]Accuracy-Flip: 0.92171+-0.01650
+testing verification..
+(12000, 512)
+infer time 74.22791999999998
+best_threshold_index 143 0.9
+best_threshold_index 138 0.8977777777777778
+best_threshold_index 140 0.9
+best_threshold_index 140 0.9014814814814814
+best_threshold_index 140 0.9053703703703704
+best_threshold_index 143 0.8992592592592593
+best_threshold_index 143 0.9014814814814814
+best_threshold_index 143 0.9068518518518518
+best_threshold_index 140 0.9053703703703704
+best_threshold_index 140 0.9018518518518519
+[agedb_30][136000]XNorm: 23.050744
+[agedb_30][136000]Accuracy-Flip: 0.89800+-0.02395
+test accuracy is: 0.9833333333333334
+epoch 8, total_step 136000, total loss is 17.58 , inference loss is 10.47, weight deacy loss is 7.11, training accuracy is 0.437500, time 54.456 samples/sec
+epoch 8, total_step 136020, total loss is 14.21 , inference loss is 7.10, weight deacy loss is 7.11, training accuracy is 0.375000, time 56.667 samples/sec
+epoch 8, total_step 136040, total loss is 14.74 , inference loss is 7.63, weight deacy loss is 7.11, training accuracy is 0.343750, time 55.232 samples/sec
+epoch 8, total_step 136060, total loss is 21.56 , inference loss is 14.45, weight deacy loss is 7.11, training accuracy is 0.250000, time 56.140 samples/sec
+epoch 8, total_step 136080, total loss is 18.35 , inference loss is 11.24, weight deacy loss is 7.11, training accuracy is 0.406250, time 56.305 samples/sec
+epoch 8, total_step 136100, total loss is 17.60 , inference loss is 10.49, weight deacy loss is 7.10, training accuracy is 0.312500, time 57.463 samples/sec
+epoch 8, total_step 136120, total loss is 14.03 , inference loss is 6.92, weight deacy loss is 7.10, training accuracy is 0.531250, time 56.581 samples/sec
+epoch 8, total_step 136140, total loss is 20.01 , inference loss is 12.90, weight deacy loss is 7.10, training accuracy is 0.437500, time 57.205 samples/sec
+epoch 8, total_step 136160, total loss is 18.92 , inference loss is 11.81, weight deacy loss is 7.10, training accuracy is 0.375000, time 56.368 samples/sec
+epoch 8, total_step 136180, total loss is 18.31 , inference loss is 11.21, weight deacy loss is 7.10, training accuracy is 0.343750, time 56.300 samples/sec
+epoch 8, total_step 136200, total loss is 11.55 , inference loss is 4.45, weight deacy loss is 7.10, training accuracy is 0.406250, time 56.027 samples/sec
+epoch 8, total_step 136220, total loss is 21.51 , inference loss is 14.41, weight deacy loss is 7.10, training accuracy is 0.218750, time 57.118 samples/sec
+epoch 8, total_step 136240, total loss is 16.04 , inference loss is 8.94, weight deacy loss is 7.10, training accuracy is 0.312500, time 56.828 samples/sec
+epoch 8, total_step 136260, total loss is 21.28 , inference loss is 14.18, weight deacy loss is 7.10, training accuracy is 0.156250, time 57.208 samples/sec
+epoch 8, total_step 136280, total loss is 16.40 , inference loss is 9.30, weight deacy loss is 7.10, training accuracy is 0.218750, time 56.039 samples/sec
+epoch 8, total_step 136300, total loss is 15.12 , inference loss is 8.02, weight deacy loss is 7.10, training accuracy is 0.218750, time 56.628 samples/sec
+epoch 8, total_step 136320, total loss is 17.44 , inference loss is 10.34, weight deacy loss is 7.10, training accuracy is 0.250000, time 56.734 samples/sec
+epoch 8, total_step 136340, total loss is 13.85 , inference loss is 6.75, weight deacy loss is 7.10, training accuracy is 0.375000, time 55.886 samples/sec
+epoch 8, total_step 136360, total loss is 19.44 , inference loss is 12.34, weight deacy loss is 7.10, training accuracy is 0.125000, time 56.833 samples/sec
+epoch 8, total_step 136380, total loss is 12.94 , inference loss is 5.84, weight deacy loss is 7.10, training accuracy is 0.343750, time 56.197 samples/sec
+epoch 8, total_step 136400, total loss is 18.01 , inference loss is 10.91, weight deacy loss is 7.10, training accuracy is 0.281250, time 55.784 samples/sec
+epoch 8, total_step 136420, total loss is 20.09 , inference loss is 12.99, weight deacy loss is 7.10, training accuracy is 0.218750, time 55.649 samples/sec
+epoch 8, total_step 136440, total loss is 14.85 , inference loss is 7.75, weight deacy loss is 7.10, training accuracy is 0.437500, time 56.774 samples/sec
+epoch 8, total_step 136460, total loss is 14.63 , inference loss is 7.54, weight deacy loss is 7.10, training accuracy is 0.312500, time 56.695 samples/sec
+epoch 8, total_step 136480, total loss is 17.19 , inference loss is 10.10, weight deacy loss is 7.10, training accuracy is 0.312500, time 56.739 samples/sec
+epoch 8, total_step 136500, total loss is 18.42 , inference loss is 11.32, weight deacy loss is 7.10, training accuracy is 0.250000, time 54.441 samples/sec
+epoch 8, total_step 136520, total loss is 16.91 , inference loss is 9.82, weight deacy loss is 7.09, training accuracy is 0.312500, time 55.509 samples/sec
+epoch 8, total_step 136540, total loss is 21.75 , inference loss is 14.66, weight deacy loss is 7.09, training accuracy is 0.187500, time 56.776 samples/sec
+epoch 8, total_step 136560, total loss is 15.85 , inference loss is 8.76, weight deacy loss is 7.09, training accuracy is 0.281250, time 56.871 samples/sec
+epoch 8, total_step 136580, total loss is 16.26 , inference loss is 9.17, weight deacy loss is 7.09, training accuracy is 0.375000, time 54.837 samples/sec
+epoch 8, total_step 136600, total loss is 19.58 , inference loss is 12.49, weight deacy loss is 7.09, training accuracy is 0.281250, time 55.340 samples/sec
+epoch 8, total_step 136620, total loss is 18.26 , inference loss is 11.17, weight deacy loss is 7.09, training accuracy is 0.250000, time 55.016 samples/sec
+epoch 8, total_step 136640, total loss is 17.76 , inference loss is 10.67, weight deacy loss is 7.09, training accuracy is 0.187500, time 55.561 samples/sec
+epoch 8, total_step 136660, total loss is 18.99 , inference loss is 11.90, weight deacy loss is 7.09, training accuracy is 0.250000, time 54.510 samples/sec
+epoch 8, total_step 136680, total loss is 20.24 , inference loss is 13.15, weight deacy loss is 7.09, training accuracy is 0.218750, time 54.941 samples/sec
+epoch 8, total_step 136700, total loss is 15.12 , inference loss is 8.03, weight deacy loss is 7.09, training accuracy is 0.218750, time 56.590 samples/sec
+epoch 8, total_step 136720, total loss is 16.38 , inference loss is 9.29, weight deacy loss is 7.09, training accuracy is 0.375000, time 55.882 samples/sec
+epoch 8, total_step 136740, total loss is 15.69 , inference loss is 8.60, weight deacy loss is 7.09, training accuracy is 0.312500, time 55.867 samples/sec
+epoch 8, total_step 136760, total loss is 18.35 , inference loss is 11.26, weight deacy loss is 7.09, training accuracy is 0.312500, time 57.165 samples/sec
+epoch 8, total_step 136780, total loss is 18.68 , inference loss is 11.59, weight deacy loss is 7.09, training accuracy is 0.281250, time 56.583 samples/sec
+epoch 8, total_step 136800, total loss is 18.29 , inference loss is 11.20, weight deacy loss is 7.09, training accuracy is 0.343750, time 57.186 samples/sec
+epoch 8, total_step 136820, total loss is 18.60 , inference loss is 11.51, weight deacy loss is 7.09, training accuracy is 0.312500, time 54.955 samples/sec
+epoch 8, total_step 136840, total loss is 18.65 , inference loss is 11.56, weight deacy loss is 7.09, training accuracy is 0.406250, time 56.468 samples/sec
+epoch 8, total_step 136860, total loss is 15.85 , inference loss is 8.76, weight deacy loss is 7.09, training accuracy is 0.281250, time 56.456 samples/sec
+epoch 8, total_step 136880, total loss is 20.82 , inference loss is 13.74, weight deacy loss is 7.09, training accuracy is 0.250000, time 56.564 samples/sec
+epoch 8, total_step 136900, total loss is 19.66 , inference loss is 12.57, weight deacy loss is 7.09, training accuracy is 0.187500, time 56.085 samples/sec
+epoch 8, total_step 136920, total loss is 19.84 , inference loss is 12.75, weight deacy loss is 7.09, training accuracy is 0.218750, time 56.029 samples/sec
+epoch 8, total_step 136940, total loss is 18.07 , inference loss is 10.98, weight deacy loss is 7.09, training accuracy is 0.343750, time 56.632 samples/sec
+epoch 8, total_step 136960, total loss is 15.49 , inference loss is 8.41, weight deacy loss is 7.09, training accuracy is 0.375000, time 56.908 samples/sec
+epoch 8, total_step 136980, total loss is 20.77 , inference loss is 13.69, weight deacy loss is 7.08, training accuracy is 0.281250, time 56.822 samples/sec
+epoch 8, total_step 137000, total loss is 18.60 , inference loss is 11.52, weight deacy loss is 7.08, training accuracy is 0.312500, time 56.208 samples/sec
+epoch 8, total_step 137020, total loss is 17.93 , inference loss is 10.85, weight deacy loss is 7.08, training accuracy is 0.156250, time 55.919 samples/sec
+epoch 8, total_step 137040, total loss is 18.79 , inference loss is 11.70, weight deacy loss is 7.08, training accuracy is 0.343750, time 56.577 samples/sec
+epoch 8, total_step 137060, total loss is 16.34 , inference loss is 9.26, weight deacy loss is 7.08, training accuracy is 0.250000, time 56.475 samples/sec
+epoch 8, total_step 137080, total loss is 23.66 , inference loss is 16.58, weight deacy loss is 7.08, training accuracy is 0.218750, time 56.427 samples/sec
+epoch 8, total_step 137100, total loss is 14.74 , inference loss is 7.65, weight deacy loss is 7.08, training accuracy is 0.375000, time 56.625 samples/sec
+epoch 8, total_step 137120, total loss is 19.18 , inference loss is 12.10, weight deacy loss is 7.08, training accuracy is 0.281250, time 56.655 samples/sec
+epoch 8, total_step 137140, total loss is 17.48 , inference loss is 10.40, weight deacy loss is 7.08, training accuracy is 0.281250, time 56.226 samples/sec
+epoch 8, total_step 137160, total loss is 13.36 , inference loss is 6.28, weight deacy loss is 7.08, training accuracy is 0.406250, time 56.507 samples/sec
+epoch 8, total_step 137180, total loss is 14.11 , inference loss is 7.03, weight deacy loss is 7.08, training accuracy is 0.343750, time 56.421 samples/sec
+epoch 8, total_step 137200, total loss is 16.78 , inference loss is 9.70, weight deacy loss is 7.08, training accuracy is 0.250000, time 56.879 samples/sec
+epoch 8, total_step 137220, total loss is 15.82 , inference loss is 8.74, weight deacy loss is 7.08, training accuracy is 0.281250, time 56.446 samples/sec
+epoch 8, total_step 137240, total loss is 16.35 , inference loss is 9.27, weight deacy loss is 7.08, training accuracy is 0.281250, time 54.855 samples/sec
+epoch 8, total_step 137260, total loss is 15.12 , inference loss is 8.04, weight deacy loss is 7.08, training accuracy is 0.375000, time 55.986 samples/sec
+epoch 8, total_step 137280, total loss is 15.88 , inference loss is 8.80, weight deacy loss is 7.08, training accuracy is 0.218750, time 56.155 samples/sec
+epoch 8, total_step 137300, total loss is 17.88 , inference loss is 10.80, weight deacy loss is 7.08, training accuracy is 0.343750, time 55.393 samples/sec
+epoch 8, total_step 137320, total loss is 14.45 , inference loss is 7.38, weight deacy loss is 7.08, training accuracy is 0.500000, time 56.273 samples/sec
+epoch 8, total_step 137340, total loss is 18.45 , inference loss is 11.38, weight deacy loss is 7.08, training accuracy is 0.156250, time 55.733 samples/sec
+epoch 8, total_step 137360, total loss is 18.32 , inference loss is 11.24, weight deacy loss is 7.08, training accuracy is 0.187500, time 56.206 samples/sec
+epoch 8, total_step 137380, total loss is 21.44 , inference loss is 14.37, weight deacy loss is 7.08, training accuracy is 0.250000, time 56.457 samples/sec
+epoch 8, total_step 137400, total loss is 15.82 , inference loss is 8.75, weight deacy loss is 7.08, training accuracy is 0.312500, time 55.535 samples/sec
+epoch 8, total_step 137420, total loss is 13.61 , inference loss is 6.54, weight deacy loss is 7.08, training accuracy is 0.375000, time 56.638 samples/sec
+epoch 8, total_step 137440, total loss is 15.54 , inference loss is 8.47, weight deacy loss is 7.07, training accuracy is 0.281250, time 56.037 samples/sec
+epoch 8, total_step 137460, total loss is 14.90 , inference loss is 7.82, weight deacy loss is 7.07, training accuracy is 0.218750, time 56.516 samples/sec
+epoch 8, total_step 137480, total loss is 17.55 , inference loss is 10.48, weight deacy loss is 7.07, training accuracy is 0.187500, time 56.647 samples/sec
+epoch 8, total_step 137500, total loss is 19.20 , inference loss is 12.12, weight deacy loss is 7.07, training accuracy is 0.281250, time 55.677 samples/sec
+epoch 8, total_step 137520, total loss is 21.17 , inference loss is 14.10, weight deacy loss is 7.07, training accuracy is 0.312500, time 56.622 samples/sec
+epoch 8, total_step 137540, total loss is 18.51 , inference loss is 11.44, weight deacy loss is 7.07, training accuracy is 0.312500, time 56.499 samples/sec
+epoch 8, total_step 137560, total loss is 15.44 , inference loss is 8.36, weight deacy loss is 7.07, training accuracy is 0.187500, time 55.839 samples/sec
+epoch 8, total_step 137580, total loss is 15.09 , inference loss is 8.02, weight deacy loss is 7.07, training accuracy is 0.156250, time 56.210 samples/sec
+epoch 8, total_step 137600, total loss is 17.72 , inference loss is 10.65, weight deacy loss is 7.07, training accuracy is 0.187500, time 56.515 samples/sec
+epoch 8, total_step 137620, total loss is 16.09 , inference loss is 9.01, weight deacy loss is 7.07, training accuracy is 0.250000, time 54.903 samples/sec
+epoch 8, total_step 137640, total loss is 17.86 , inference loss is 10.79, weight deacy loss is 7.07, training accuracy is 0.250000, time 57.097 samples/sec
+epoch 8, total_step 137660, total loss is 20.34 , inference loss is 13.27, weight deacy loss is 7.07, training accuracy is 0.250000, time 56.331 samples/sec
+epoch 8, total_step 137680, total loss is 17.69 , inference loss is 10.62, weight deacy loss is 7.07, training accuracy is 0.218750, time 56.466 samples/sec
+epoch 8, total_step 137700, total loss is 18.97 , inference loss is 11.90, weight deacy loss is 7.07, training accuracy is 0.250000, time 56.634 samples/sec
+epoch 8, total_step 137720, total loss is 17.64 , inference loss is 10.57, weight deacy loss is 7.07, training accuracy is 0.250000, time 56.635 samples/sec
+epoch 8, total_step 137740, total loss is 20.56 , inference loss is 13.49, weight deacy loss is 7.07, training accuracy is 0.187500, time 56.813 samples/sec
+epoch 8, total_step 137760, total loss is 17.63 , inference loss is 10.56, weight deacy loss is 7.07, training accuracy is 0.218750, time 56.331 samples/sec
+epoch 8, total_step 137780, total loss is 16.18 , inference loss is 9.12, weight deacy loss is 7.07, training accuracy is 0.281250, time 57.108 samples/sec
+epoch 8, total_step 137800, total loss is 16.60 , inference loss is 9.54, weight deacy loss is 7.07, training accuracy is 0.250000, time 56.147 samples/sec
+epoch 8, total_step 137820, total loss is 15.80 , inference loss is 8.73, weight deacy loss is 7.07, training accuracy is 0.343750, time 56.100 samples/sec
+epoch 8, total_step 137840, total loss is 14.27 , inference loss is 7.20, weight deacy loss is 7.07, training accuracy is 0.312500, time 56.182 samples/sec
+epoch 8, total_step 137860, total loss is 17.66 , inference loss is 10.59, weight deacy loss is 7.07, training accuracy is 0.406250, time 55.817 samples/sec
+epoch 8, total_step 137880, total loss is 19.08 , inference loss is 12.01, weight deacy loss is 7.06, training accuracy is 0.156250, time 55.178 samples/sec
+epoch 8, total_step 137900, total loss is 17.38 , inference loss is 10.31, weight deacy loss is 7.06, training accuracy is 0.281250, time 57.168 samples/sec
+epoch 8, total_step 137920, total loss is 14.89 , inference loss is 7.82, weight deacy loss is 7.06, training accuracy is 0.187500, time 57.396 samples/sec
+epoch 8, total_step 137940, total loss is 16.90 , inference loss is 9.84, weight deacy loss is 7.06, training accuracy is 0.312500, time 56.463 samples/sec2022-12-02 03:47:26.304768: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:111] Filling up shuffle buffer (this may take a while): 7256 of 10000
+2022-12-02 03:47:29.888176: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:162] Shuffle buffer filled.
+
+epoch 8, total_step 137960, total loss is 16.73 , inference loss is 9.67, weight deacy loss is 7.06, training accuracy is 0.343750, time 56.232 samples/sec
+epoch 8, total_step 137980, total loss is 17.19 , inference loss is 10.13, weight deacy loss is 7.06, training accuracy is 0.375000, time 56.165 samples/sec
+
+
+
+
+End of epoch 8
+testing verification..
+(12000, 512)
+infer time 73.97039299999999
+best_threshold_index 120 0.9874074074074074
+best_threshold_index 120 0.9864814814814815
+best_threshold_index 120 0.987037037037037
+best_threshold_index 120 0.9877777777777778
+best_threshold_index 120 0.9868518518518519
+best_threshold_index 124 0.9864814814814815
+best_threshold_index 120 0.9874074074074074
+best_threshold_index 120 0.9872222222222222
+best_threshold_index 120 0.9857407407407407
+best_threshold_index 120 0.9862962962962963
+[lfw][138000]XNorm: 21.574000
+[lfw][138000]Accuracy-Flip: 0.98633+-0.00521
+testing verification..
+(14000, 512)
+infer time 86.19998700000006
+best_threshold_index 115 0.9882539682539683
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 115 0.9871428571428571
+best_threshold_index 115 0.9874603174603175
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 115 0.9871428571428571
+best_threshold_index 115 0.9882539682539683
+best_threshold_index 115 0.9876190476190476
+best_threshold_index 115 0.9874603174603175
+best_threshold_index 115 0.9882539682539683
+[cfp_ff][138000]XNorm: 21.594550
+[cfp_ff][138000]Accuracy-Flip: 0.98771+-0.00368
+testing verification..
+(14000, 512)
+infer time 86.08775099999993
+best_threshold_index 135 0.9241269841269841
+best_threshold_index 142 0.9252380952380952
+best_threshold_index 135 0.9226984126984127
+best_threshold_index 135 0.9255555555555556
+best_threshold_index 142 0.926984126984127
+best_threshold_index 135 0.9222222222222223
+best_threshold_index 142 0.9222222222222223
+best_threshold_index 142 0.9238095238095239
+best_threshold_index 135 0.9238095238095239
+best_threshold_index 142 0.926031746031746
+[cfp_fp][138000]XNorm: 21.313305
+[cfp_fp][138000]Accuracy-Flip: 0.92014+-0.01405
+testing verification..
+(12000, 512)
+infer time 74.00460100000006
+best_threshold_index 141 0.9033333333333333
+best_threshold_index 138 0.9016666666666666
+best_threshold_index 138 0.9055555555555556
+best_threshold_index 138 0.905
+best_threshold_index 138 0.9088888888888889
+best_threshold_index 138 0.9027777777777778
+best_threshold_index 139 0.9048148148148148
+best_threshold_index 139 0.9085185185185185
+best_threshold_index 138 0.9066666666666666
+best_threshold_index 138 0.9061111111111111
+[agedb_30][138000]XNorm: 21.952532
+[agedb_30][138000]Accuracy-Flip: 0.90283+-0.01985
+test accuracy is: 0.9863333333333333
+epoch 9, total_step 138000, total loss is 18.16 , inference loss is 11.10, weight deacy loss is 7.06, training accuracy is 0.250000, time 54.179 samples/sec
+epoch 9, total_step 138020, total loss is 13.00 , inference loss is 5.93, weight deacy loss is 7.06, training accuracy is 0.406250, time 56.335 samples/sec
+epoch 9, total_step 138040, total loss is 12.65 , inference loss is 5.59, weight deacy loss is 7.06, training accuracy is 0.500000, time 56.741 samples/sec
+epoch 9, total_step 138060, total loss is 12.97 , inference loss is 5.91, weight deacy loss is 7.06, training accuracy is 0.468750, time 57.460 samples/sec
+epoch 9, total_step 138080, total loss is 11.88 , inference loss is 4.82, weight deacy loss is 7.06, training accuracy is 0.656250, time 56.586 samples/sec
+epoch 9, total_step 138100, total loss is 11.20 , inference loss is 4.14, weight deacy loss is 7.06, training accuracy is 0.625000, time 56.560 samples/sec
+epoch 9, total_step 138120, total loss is 10.47 , inference loss is 3.41, weight deacy loss is 7.06, training accuracy is 0.625000, time 56.150 samples/sec
+epoch 9, total_step 138140, total loss is 12.08 , inference loss is 5.02, weight deacy loss is 7.06, training accuracy is 0.500000, time 56.400 samples/sec
+epoch 9, total_step 138160, total loss is 12.42 , inference loss is 5.36, weight deacy loss is 7.06, training accuracy is 0.687500, time 56.575 samples/sec
+epoch 9, total_step 138180, total loss is 11.78 , inference loss is 4.72, weight deacy loss is 7.06, training accuracy is 0.656250, time 56.842 samples/sec
+epoch 9, total_step 138200, total loss is 10.59 , inference loss is 3.53, weight deacy loss is 7.06, training accuracy is 0.656250, time 56.195 samples/sec
+epoch 9, total_step 138220, total loss is 13.77 , inference loss is 6.71, weight deacy loss is 7.06, training accuracy is 0.656250, time 55.508 samples/sec
+epoch 9, total_step 138240, total loss is 12.73 , inference loss is 5.67, weight deacy loss is 7.06, training accuracy is 0.500000, time 57.407 samples/sec
+epoch 9, total_step 138260, total loss is 10.83 , inference loss is 3.78, weight deacy loss is 7.06, training accuracy is 0.625000, time 56.109 samples/sec
+epoch 9, total_step 138280, total loss is 10.74 , inference loss is 3.68, weight deacy loss is 7.06, training accuracy is 0.656250, time 56.699 samples/sec
+epoch 9, total_step 138300, total loss is 14.09 , inference loss is 7.04, weight deacy loss is 7.06, training accuracy is 0.562500, time 56.071 samples/sec
+epoch 9, total_step 138320, total loss is 13.40 , inference loss is 6.35, weight deacy loss is 7.05, training accuracy is 0.593750, time 56.816 samples/sec
+epoch 9, total_step 138340, total loss is 10.14 , inference loss is 3.08, weight deacy loss is 7.05, training accuracy is 0.593750, time 56.479 samples/sec
+epoch 9, total_step 138360, total loss is 13.72 , inference loss is 6.67, weight deacy loss is 7.05, training accuracy is 0.593750, time 55.650 samples/sec
+epoch 9, total_step 138380, total loss is 12.03 , inference loss is 4.98, weight deacy loss is 7.05, training accuracy is 0.656250, time 56.285 samples/sec
+epoch 9, total_step 138400, total loss is 10.26 , inference loss is 3.21, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.405 samples/sec
+epoch 9, total_step 138420, total loss is 15.77 , inference loss is 8.71, weight deacy loss is 7.05, training accuracy is 0.531250, time 55.767 samples/sec
+epoch 9, total_step 138440, total loss is 10.65 , inference loss is 3.59, weight deacy loss is 7.05, training accuracy is 0.718750, time 56.910 samples/sec
+epoch 9, total_step 138460, total loss is 13.49 , inference loss is 6.44, weight deacy loss is 7.05, training accuracy is 0.531250, time 56.580 samples/sec
+epoch 9, total_step 138480, total loss is 11.61 , inference loss is 4.55, weight deacy loss is 7.05, training accuracy is 0.625000, time 56.598 samples/sec
+epoch 9, total_step 138500, total loss is 13.97 , inference loss is 6.92, weight deacy loss is 7.05, training accuracy is 0.468750, time 55.996 samples/sec
+epoch 9, total_step 138520, total loss is 17.37 , inference loss is 10.32, weight deacy loss is 7.05, training accuracy is 0.562500, time 57.071 samples/sec
+epoch 9, total_step 138540, total loss is 16.31 , inference loss is 9.26, weight deacy loss is 7.05, training accuracy is 0.562500, time 56.465 samples/sec
+epoch 9, total_step 138560, total loss is 11.16 , inference loss is 4.11, weight deacy loss is 7.05, training accuracy is 0.687500, time 56.623 samples/sec
+epoch 9, total_step 138580, total loss is 14.81 , inference loss is 7.76, weight deacy loss is 7.05, training accuracy is 0.343750, time 56.006 samples/sec
+epoch 9, total_step 138600, total loss is 14.88 , inference loss is 7.83, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.557 samples/sec
+epoch 9, total_step 138620, total loss is 12.39 , inference loss is 5.34, weight deacy loss is 7.05, training accuracy is 0.468750, time 56.480 samples/sec
+epoch 9, total_step 138640, total loss is 17.04 , inference loss is 9.99, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.185 samples/sec
+epoch 9, total_step 138660, total loss is 12.45 , inference loss is 5.40, weight deacy loss is 7.05, training accuracy is 0.593750, time 56.537 samples/sec
+epoch 9, total_step 138680, total loss is 9.74 , inference loss is 2.69, weight deacy loss is 7.05, training accuracy is 0.593750, time 56.666 samples/sec
+epoch 9, total_step 138700, total loss is 15.05 , inference loss is 8.00, weight deacy loss is 7.05, training accuracy is 0.593750, time 56.633 samples/sec
+epoch 9, total_step 138720, total loss is 12.59 , inference loss is 5.54, weight deacy loss is 7.05, training accuracy is 0.625000, time 56.316 samples/sec
+epoch 9, total_step 138740, total loss is 16.41 , inference loss is 9.36, weight deacy loss is 7.05, training accuracy is 0.468750, time 57.001 samples/sec
+epoch 9, total_step 138760, total loss is 16.53 , inference loss is 9.48, weight deacy loss is 7.05, training accuracy is 0.468750, time 57.183 samples/sec
+epoch 9, total_step 138780, total loss is 15.98 , inference loss is 8.93, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.226 samples/sec
+epoch 9, total_step 138800, total loss is 14.80 , inference loss is 7.75, weight deacy loss is 7.05, training accuracy is 0.531250, time 55.972 samples/sec
+epoch 9, total_step 138820, total loss is 12.90 , inference loss is 5.85, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.841 samples/sec
+epoch 9, total_step 138840, total loss is 13.51 , inference loss is 6.46, weight deacy loss is 7.05, training accuracy is 0.656250, time 56.905 samples/sec
+epoch 9, total_step 138860, total loss is 13.61 , inference loss is 6.56, weight deacy loss is 7.05, training accuracy is 0.531250, time 56.876 samples/sec
+epoch 9, total_step 138880, total loss is 14.59 , inference loss is 7.55, weight deacy loss is 7.05, training accuracy is 0.531250, time 56.867 samples/sec
+epoch 9, total_step 138900, total loss is 13.22 , inference loss is 6.17, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.459 samples/sec
+epoch 9, total_step 138920, total loss is 16.68 , inference loss is 9.63, weight deacy loss is 7.05, training accuracy is 0.468750, time 56.565 samples/sec
+epoch 9, total_step 138940, total loss is 14.36 , inference loss is 7.31, weight deacy loss is 7.05, training accuracy is 0.312500, time 56.635 samples/sec
+epoch 9, total_step 138960, total loss is 13.60 , inference loss is 6.55, weight deacy loss is 7.05, training accuracy is 0.593750, time 57.021 samples/sec
+epoch 9, total_step 138980, total loss is 14.60 , inference loss is 7.55, weight deacy loss is 7.05, training accuracy is 0.562500, time 56.215 samples/sec
+epoch 9, total_step 139000, total loss is 14.68 , inference loss is 7.63, weight deacy loss is 7.05, training accuracy is 0.500000, time 57.072 samples/sec
+epoch 9, total_step 139020, total loss is 13.33 , inference loss is 6.29, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.540 samples/sec
+epoch 9, total_step 139040, total loss is 16.36 , inference loss is 9.31, weight deacy loss is 7.05, training accuracy is 0.437500, time 57.223 samples/sec
+epoch 9, total_step 139060, total loss is 17.97 , inference loss is 10.92, weight deacy loss is 7.05, training accuracy is 0.312500, time 56.246 samples/sec
+epoch 9, total_step 139080, total loss is 17.72 , inference loss is 10.68, weight deacy loss is 7.05, training accuracy is 0.437500, time 56.602 samples/sec
+epoch 9, total_step 139100, total loss is 11.98 , inference loss is 4.94, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.052 samples/sec
+epoch 9, total_step 139120, total loss is 17.07 , inference loss is 10.02, weight deacy loss is 7.05, training accuracy is 0.406250, time 55.813 samples/sec
+epoch 9, total_step 139140, total loss is 19.45 , inference loss is 12.41, weight deacy loss is 7.05, training accuracy is 0.406250, time 57.113 samples/sec
+epoch 9, total_step 139160, total loss is 19.25 , inference loss is 12.20, weight deacy loss is 7.05, training accuracy is 0.375000, time 56.362 samples/sec
+epoch 9, total_step 139180, total loss is 19.07 , inference loss is 12.02, weight deacy loss is 7.05, training accuracy is 0.406250, time 55.527 samples/sec
+epoch 9, total_step 139200, total loss is 13.56 , inference loss is 6.51, weight deacy loss is 7.05, training accuracy is 0.468750, time 55.075 samples/sec
+epoch 9, total_step 139220, total loss is 18.89 , inference loss is 11.85, weight deacy loss is 7.05, training accuracy is 0.343750, time 56.997 samples/sec
+epoch 9, total_step 139240, total loss is 16.00 , inference loss is 8.96, weight deacy loss is 7.05, training accuracy is 0.593750, time 56.501 samples/sec
+epoch 9, total_step 139260, total loss is 12.00 , inference loss is 4.95, weight deacy loss is 7.05, training accuracy is 0.625000, time 57.171 samples/sec
+epoch 9, total_step 139280, total loss is 16.03 , inference loss is 8.98, weight deacy loss is 7.05, training accuracy is 0.437500, time 56.844 samples/sec
+epoch 9, total_step 139300, total loss is 11.36 , inference loss is 4.32, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.699 samples/sec
+epoch 9, total_step 139320, total loss is 17.84 , inference loss is 10.80, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.090 samples/sec
+epoch 9, total_step 139340, total loss is 21.08 , inference loss is 14.03, weight deacy loss is 7.05, training accuracy is 0.281250, time 56.687 samples/sec
+epoch 9, total_step 139360, total loss is 16.15 , inference loss is 9.11, weight deacy loss is 7.05, training accuracy is 0.500000, time 56.109 samples/sec
+epoch 9, total_step 139380, total loss is 11.94 , inference loss is 4.90, weight deacy loss is 7.05, training accuracy is 0.468750, time 56.300 samples/sec
+epoch 9, total_step 139400, total loss is 13.37 , inference loss is 6.32, weight deacy loss is 7.05, training accuracy is 0.406250, time 56.760 samples/sec
+epoch 9, total_step 139420, total loss is 14.40 , inference loss is 7.36, weight deacy loss is 7.05, training accuracy is 0.468750, time 57.003 samples/sec
+epoch 9, total_step 139440, total loss is 12.71 , inference loss is 5.66, weight deacy loss is 7.04, training accuracy is 0.593750, time 56.849 samples/sec
+epoch 9, total_step 139460, total loss is 15.52 , inference loss is 8.47, weight deacy loss is 7.04, training accuracy is 0.375000, time 55.778 samples/sec
+epoch 9, total_step 139480, total loss is 13.56 , inference loss is 6.51, weight deacy loss is 7.04, training accuracy is 0.437500, time 56.961 samples/sec
+epoch 9, total_step 139500, total loss is 13.97 , inference loss is 6.92, weight deacy loss is 7.04, training accuracy is 0.500000, time 56.300 samples/sec
+epoch 9, total_step 139520, total loss is 14.30 , inference loss is 7.26, weight deacy loss is 7.04, training accuracy is 0.500000, time 55.967 samples/sec
+epoch 9, total_step 139540, total loss is 16.48 , inference loss is 9.44, weight deacy loss is 7.04, training accuracy is 0.406250, time 56.927 samples/sec
+epoch 9, total_step 139560, total loss is 16.11 , inference loss is 9.07, weight deacy loss is 7.04, training accuracy is 0.500000, time 56.162 samples/sec
+epoch 9, total_step 139580, total loss is 15.54 , inference loss is 8.50, weight deacy loss is 7.04, training accuracy is 0.406250, time 57.135 samples/sec
+epoch 9, total_step 139600, total loss is 14.09 , inference loss is 7.05, weight deacy loss is 7.04, training accuracy is 0.375000, time 56.346 samples/sec
+epoch 9, total_step 139620, total loss is 22.43 , inference loss is 15.38, weight deacy loss is 7.04, training accuracy is 0.312500, time 56.503 samples/sec
+epoch 9, total_step 139640, total loss is 14.56 , inference loss is 7.51, weight deacy loss is 7.04, training accuracy is 0.531250, time 55.786 samples/sec
+epoch 9, total_step 139660, total loss is 13.90 , inference loss is 6.85, weight deacy loss is 7.04, training accuracy is 0.468750, time 56.501 samples/sec
+epoch 9, total_step 139680, total loss is 17.86 , inference loss is 10.82, weight deacy loss is 7.04, training accuracy is 0.375000, time 56.936 samples/sec
+epoch 9, total_step 139700, total loss is 17.50 , inference loss is 10.45, weight deacy loss is 7.04, training accuracy is 0.375000, time 56.928 samples/sec
+epoch 9, total_step 139720, total loss is 14.19 , inference loss is 7.15, weight deacy loss is 7.04, training accuracy is 0.531250, time 55.805 samples/sec
+epoch 9, total_step 139740, total loss is 14.25 , inference loss is 7.20, weight deacy loss is 7.04, training accuracy is 0.500000, time 56.034 samples/sec
+epoch 9, total_step 139760, total loss is 14.49 , inference loss is 7.45, weight deacy loss is 7.04, training accuracy is 0.406250, time 56.336 samples/sec
+epoch 9, total_step 139780, total loss is 10.50 , inference loss is 3.45, weight deacy loss is 7.04, training accuracy is 0.593750, time 56.560 samples/sec
+epoch 9, total_step 139800, total loss is 13.67 , inference loss is 6.63, weight deacy loss is 7.04, training accuracy is 0.375000, time 56.378 samples/sec
+epoch 9, total_step 139820, total loss is 15.58 , inference loss is 8.54, weight deacy loss is 7.04, training accuracy is 0.312500, time 56.572 samples/sec
+epoch 9, total_step 139840, total loss is 15.46 , inference loss is 8.42, weight deacy loss is 7.04, training accuracy is 0.312500, time 56.474 samples/sec
+epoch 9, total_step 139860, total loss is 14.70 , inference loss is 7.66, weight deacy loss is 7.04, training accuracy is 0.500000, time 55.958 samples/sec
+epoch 9, total_step 139880, total loss is 13.90 , inference loss is 6.86, weight deacy loss is 7.04, training accuracy is 0.531250, time 56.843 samples/sec
+epoch 9, total_step 139900, total loss is 16.37 , inference loss is 9.33, weight deacy loss is 7.04, training accuracy is 0.562500, time 56.046 samples/sec
+epoch 9, total_step 139920, total loss is 14.72 , inference loss is 7.68, weight deacy loss is 7.04, training accuracy is 0.562500, time 56.518 samples/sec
+epoch 9, total_step 139940, total loss is 14.18 , inference loss is 7.15, weight deacy loss is 7.04, training accuracy is 0.500000, time 56.245 samples/sec
+epoch 9, total_step 139960, total loss is 10.41 , inference loss is 3.37, weight deacy loss is 7.04, training accuracy is 0.593750, time 56.686 samples/sec
+epoch 9, total_step 139980, total loss is 12.26 , inference loss is 5.22, weight deacy loss is 7.04, training accuracy is 0.625000, time 56.421 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.972546
+best_threshold_index 117 0.9859259259259259
+best_threshold_index 117 0.9853703703703703
+best_threshold_index 117 0.9844444444444445
+best_threshold_index 116 0.9853703703703703
+best_threshold_index 117 0.9857407407407407
+best_threshold_index 117 0.9842592592592593
+best_threshold_index 117 0.9855555555555555
+best_threshold_index 117 0.985
+best_threshold_index 117 0.9844444444444445
+best_threshold_index 117 0.9838888888888889
+[lfw][140000]XNorm: 21.207522
+[lfw][140000]Accuracy-Flip: 0.98450+-0.00641
+testing verification..
+(14000, 512)
+infer time 86.36017600000005
+best_threshold_index 122 0.9861904761904762
+best_threshold_index 122 0.9855555555555555
+best_threshold_index 122 0.9858730158730159
+best_threshold_index 122 0.9842857142857143
+best_threshold_index 122 0.9855555555555555
+best_threshold_index 122 0.9841269841269841
+best_threshold_index 122 0.986031746031746
+best_threshold_index 122 0.984920634920635
+best_threshold_index 122 0.984920634920635
+best_threshold_index 120 0.9855555555555555
+[cfp_ff][140000]XNorm: 20.465177
+[cfp_ff][140000]Accuracy-Flip: 0.98457+-0.00651
+testing verification..
+(14000, 512)
+infer time 86.25392600000013
+best_threshold_index 138 0.9217460317460318
+best_threshold_index 138 0.9206349206349206
+best_threshold_index 138 0.9196825396825397
+best_threshold_index 138 0.9215873015873016
+best_threshold_index 138 0.9223809523809524
+best_threshold_index 138 0.9173015873015873
+best_threshold_index 138 0.9184126984126985
+best_threshold_index 138 0.9185714285714286
+best_threshold_index 139 0.9195238095238095
+best_threshold_index 139 0.9207936507936508
+[cfp_fp][140000]XNorm: 21.593301
+[cfp_fp][140000]Accuracy-Flip: 0.91943+-0.01406
+testing verification..
+(12000, 512)
+infer time 73.845825
+best_threshold_index 142 0.9092592592592592
+best_threshold_index 140 0.9081481481481481
+best_threshold_index 142 0.9109259259259259
+best_threshold_index 142 0.9112962962962963
+best_threshold_index 140 0.9166666666666666
+best_threshold_index 140 0.9096296296296297
+best_threshold_index 142 0.9094444444444445
+best_threshold_index 142 0.9159259259259259
+best_threshold_index 142 0.9148148148148149
+best_threshold_index 140 0.9118518518518518
+[agedb_30][140000]XNorm: 22.579918
+[agedb_30][140000]Accuracy-Flip: 0.90917+-0.02554
+test accuracy is: 0.9844999999999999
+epoch 9, total_step 140000, total loss is 14.96 , inference loss is 7.92, weight deacy loss is 7.04, training accuracy is 0.437500, time 55.136 samples/sec
+epoch 9, total_step 140020, total loss is 16.58 , inference loss is 9.55, weight deacy loss is 7.04, training accuracy is 0.468750, time 56.617 samples/sec
+epoch 9, total_step 140040, total loss is 12.77 , inference loss is 5.73, weight deacy loss is 7.04, training accuracy is 0.531250, time 56.360 samples/sec
+epoch 9, total_step 140060, total loss is 13.52 , inference loss is 6.48, weight deacy loss is 7.04, training accuracy is 0.625000, time 56.156 samples/sec
+epoch 9, total_step 140080, total loss is 13.30 , inference loss is 6.26, weight deacy loss is 7.04, training accuracy is 0.625000, time 57.117 samples/sec
+epoch 9, total_step 140100, total loss is 14.52 , inference loss is 7.49, weight deacy loss is 7.04, training accuracy is 0.531250, time 56.468 samples/sec
+epoch 9, total_step 140120, total loss is 13.15 , inference loss is 6.11, weight deacy loss is 7.04, training accuracy is 0.562500, time 56.777 samples/sec
+epoch 9, total_step 140140, total loss is 13.33 , inference loss is 6.30, weight deacy loss is 7.04, training accuracy is 0.562500, time 56.446 samples/sec
+epoch 9, total_step 140160, total loss is 14.11 , inference loss is 7.07, weight deacy loss is 7.04, training accuracy is 0.468750, time 56.821 samples/sec
+epoch 9, total_step 140180, total loss is 10.26 , inference loss is 3.23, weight deacy loss is 7.03, training accuracy is 0.625000, time 57.320 samples/sec
+epoch 9, total_step 140200, total loss is 12.26 , inference loss is 5.23, weight deacy loss is 7.03, training accuracy is 0.656250, time 56.229 samples/sec
+epoch 9, total_step 140220, total loss is 16.40 , inference loss is 9.37, weight deacy loss is 7.03, training accuracy is 0.562500, time 56.245 samples/sec
+epoch 9, total_step 140240, total loss is 15.69 , inference loss is 8.66, weight deacy loss is 7.03, training accuracy is 0.500000, time 56.391 samples/sec
+epoch 9, total_step 140260, total loss is 14.35 , inference loss is 7.32, weight deacy loss is 7.03, training accuracy is 0.500000, time 56.420 samples/sec
+epoch 9, total_step 140280, total loss is 17.86 , inference loss is 10.83, weight deacy loss is 7.03, training accuracy is 0.437500, time 56.621 samples/sec
+epoch 9, total_step 140300, total loss is 11.94 , inference loss is 4.91, weight deacy loss is 7.03, training accuracy is 0.562500, time 56.419 samples/sec
+epoch 9, total_step 140320, total loss is 13.98 , inference loss is 6.94, weight deacy loss is 7.03, training accuracy is 0.500000, time 56.599 samples/sec
+epoch 9, total_step 140340, total loss is 13.93 , inference loss is 6.89, weight deacy loss is 7.03, training accuracy is 0.625000, time 56.933 samples/sec
+epoch 9, total_step 140360, total loss is 10.38 , inference loss is 3.35, weight deacy loss is 7.03, training accuracy is 0.687500, time 56.341 samples/sec
+epoch 9, total_step 140380, total loss is 13.85 , inference loss is 6.82, weight deacy loss is 7.03, training accuracy is 0.531250, time 55.010 samples/sec
+epoch 9, total_step 140400, total loss is 13.15 , inference loss is 6.12, weight deacy loss is 7.03, training accuracy is 0.531250, time 57.528 samples/sec
+epoch 9, total_step 140420, total loss is 11.31 , inference loss is 4.28, weight deacy loss is 7.03, training accuracy is 0.531250, time 56.638 samples/sec
+epoch 9, total_step 140440, total loss is 16.17 , inference loss is 9.14, weight deacy loss is 7.03, training accuracy is 0.437500, time 56.483 samples/sec
+epoch 9, total_step 140460, total loss is 12.16 , inference loss is 5.14, weight deacy loss is 7.03, training accuracy is 0.500000, time 56.469 samples/sec
+epoch 9, total_step 140480, total loss is 14.64 , inference loss is 7.61, weight deacy loss is 7.03, training accuracy is 0.406250, time 56.953 samples/sec
+epoch 9, total_step 140500, total loss is 14.43 , inference loss is 7.40, weight deacy loss is 7.03, training accuracy is 0.562500, time 57.486 samples/sec
+epoch 9, total_step 140520, total loss is 12.29 , inference loss is 5.26, weight deacy loss is 7.03, training accuracy is 0.562500, time 56.722 samples/sec
+epoch 9, total_step 140540, total loss is 10.00 , inference loss is 2.98, weight deacy loss is 7.03, training accuracy is 0.656250, time 56.927 samples/sec
+epoch 9, total_step 140560, total loss is 11.72 , inference loss is 4.70, weight deacy loss is 7.03, training accuracy is 0.656250, time 55.720 samples/sec
+epoch 9, total_step 140580, total loss is 11.02 , inference loss is 3.99, weight deacy loss is 7.03, training accuracy is 0.718750, time 56.717 samples/sec
+epoch 9, total_step 140600, total loss is 13.55 , inference loss is 6.52, weight deacy loss is 7.03, training accuracy is 0.531250, time 54.946 samples/sec
+epoch 9, total_step 140620, total loss is 14.57 , inference loss is 7.54, weight deacy loss is 7.03, training accuracy is 0.593750, time 56.461 samples/sec
+epoch 9, total_step 140640, total loss is 13.27 , inference loss is 6.24, weight deacy loss is 7.03, training accuracy is 0.593750, time 56.993 samples/sec
+epoch 9, total_step 140660, total loss is 14.88 , inference loss is 7.85, weight deacy loss is 7.03, training accuracy is 0.562500, time 55.855 samples/sec
+epoch 9, total_step 140680, total loss is 12.16 , inference loss is 5.13, weight deacy loss is 7.03, training accuracy is 0.625000, time 57.062 samples/sec
+epoch 9, total_step 140700, total loss is 16.91 , inference loss is 9.88, weight deacy loss is 7.03, training accuracy is 0.562500, time 56.780 samples/sec
+epoch 9, total_step 140720, total loss is 13.23 , inference loss is 6.20, weight deacy loss is 7.02, training accuracy is 0.468750, time 56.669 samples/sec
+epoch 9, total_step 140740, total loss is 8.91 , inference loss is 1.89, weight deacy loss is 7.02, training accuracy is 0.687500, time 56.561 samples/sec
+epoch 9, total_step 140760, total loss is 12.10 , inference loss is 5.08, weight deacy loss is 7.02, training accuracy is 0.625000, time 56.991 samples/sec
+epoch 9, total_step 140780, total loss is 15.99 , inference loss is 8.97, weight deacy loss is 7.02, training accuracy is 0.593750, time 56.881 samples/sec
+epoch 9, total_step 140800, total loss is 11.60 , inference loss is 4.58, weight deacy loss is 7.02, training accuracy is 0.687500, time 57.134 samples/sec
+epoch 9, total_step 140820, total loss is 12.22 , inference loss is 5.20, weight deacy loss is 7.02, training accuracy is 0.500000, time 56.938 samples/sec
+epoch 9, total_step 140840, total loss is 13.28 , inference loss is 6.26, weight deacy loss is 7.02, training accuracy is 0.500000, time 55.836 samples/sec
+epoch 9, total_step 140860, total loss is 11.56 , inference loss is 4.54, weight deacy loss is 7.02, training accuracy is 0.625000, time 56.920 samples/sec
+epoch 9, total_step 140880, total loss is 13.53 , inference loss is 6.51, weight deacy loss is 7.02, training accuracy is 0.656250, time 57.197 samples/sec
+epoch 9, total_step 140900, total loss is 10.55 , inference loss is 3.53, weight deacy loss is 7.02, training accuracy is 0.625000, time 57.605 samples/sec
+epoch 9, total_step 140920, total loss is 10.66 , inference loss is 3.64, weight deacy loss is 7.02, training accuracy is 0.531250, time 56.957 samples/sec
+epoch 9, total_step 140940, total loss is 10.56 , inference loss is 3.54, weight deacy loss is 7.02, training accuracy is 0.625000, time 56.285 samples/sec
+epoch 9, total_step 140960, total loss is 14.54 , inference loss is 7.52, weight deacy loss is 7.02, training accuracy is 0.406250, time 56.651 samples/sec
+epoch 9, total_step 140980, total loss is 11.16 , inference loss is 4.15, weight deacy loss is 7.02, training accuracy is 0.625000, time 56.177 samples/sec
+epoch 9, total_step 141000, total loss is 14.05 , inference loss is 7.03, weight deacy loss is 7.02, training accuracy is 0.562500, time 55.735 samples/sec
+epoch 9, total_step 141020, total loss is 13.42 , inference loss is 6.40, weight deacy loss is 7.02, training accuracy is 0.531250, time 56.927 samples/sec
+epoch 9, total_step 141040, total loss is 16.12 , inference loss is 9.10, weight deacy loss is 7.02, training accuracy is 0.468750, time 56.710 samples/sec
+epoch 9, total_step 141060, total loss is 14.68 , inference loss is 7.66, weight deacy loss is 7.02, training accuracy is 0.500000, time 57.348 samples/sec
+epoch 9, total_step 141080, total loss is 14.20 , inference loss is 7.18, weight deacy loss is 7.02, training accuracy is 0.437500, time 57.012 samples/sec
+epoch 9, total_step 141100, total loss is 12.67 , inference loss is 5.66, weight deacy loss is 7.02, training accuracy is 0.531250, time 57.385 samples/sec
+epoch 9, total_step 141120, total loss is 11.59 , inference loss is 4.57, weight deacy loss is 7.02, training accuracy is 0.625000, time 56.559 samples/sec
+epoch 9, total_step 141140, total loss is 13.32 , inference loss is 6.31, weight deacy loss is 7.02, training accuracy is 0.531250, time 56.937 samples/sec
+epoch 9, total_step 141160, total loss is 9.45 , inference loss is 2.43, weight deacy loss is 7.02, training accuracy is 0.687500, time 56.782 samples/sec
+epoch 9, total_step 141180, total loss is 10.85 , inference loss is 3.83, weight deacy loss is 7.02, training accuracy is 0.625000, time 56.727 samples/sec
+epoch 9, total_step 141200, total loss is 16.24 , inference loss is 9.23, weight deacy loss is 7.02, training accuracy is 0.468750, time 56.568 samples/sec
+epoch 9, total_step 141220, total loss is 13.83 , inference loss is 6.81, weight deacy loss is 7.02, training accuracy is 0.468750, time 54.726 samples/sec
+epoch 9, total_step 141240, total loss is 13.03 , inference loss is 6.02, weight deacy loss is 7.02, training accuracy is 0.562500, time 56.856 samples/sec
+epoch 9, total_step 141260, total loss is 15.39 , inference loss is 8.38, weight deacy loss is 7.02, training accuracy is 0.343750, time 56.553 samples/sec
+epoch 9, total_step 141280, total loss is 15.89 , inference loss is 8.87, weight deacy loss is 7.02, training accuracy is 0.468750, time 56.607 samples/sec
+epoch 9, total_step 141300, total loss is 15.14 , inference loss is 8.12, weight deacy loss is 7.01, training accuracy is 0.406250, time 55.803 samples/sec
+epoch 9, total_step 141320, total loss is 13.35 , inference loss is 6.34, weight deacy loss is 7.01, training accuracy is 0.593750, time 56.662 samples/sec
+epoch 9, total_step 141340, total loss is 13.82 , inference loss is 6.80, weight deacy loss is 7.01, training accuracy is 0.375000, time 56.728 samples/sec
+epoch 9, total_step 141360, total loss is 11.22 , inference loss is 4.20, weight deacy loss is 7.01, training accuracy is 0.437500, time 56.035 samples/sec
+epoch 9, total_step 141380, total loss is 13.23 , inference loss is 6.21, weight deacy loss is 7.01, training accuracy is 0.500000, time 56.039 samples/sec
+epoch 9, total_step 141400, total loss is 15.68 , inference loss is 8.66, weight deacy loss is 7.01, training accuracy is 0.406250, time 56.354 samples/sec
+epoch 9, total_step 141420, total loss is 12.91 , inference loss is 5.90, weight deacy loss is 7.01, training accuracy is 0.437500, time 56.009 samples/sec
+epoch 9, total_step 141440, total loss is 12.57 , inference loss is 5.56, weight deacy loss is 7.01, training accuracy is 0.656250, time 56.076 samples/sec
+epoch 9, total_step 141460, total loss is 14.00 , inference loss is 6.98, weight deacy loss is 7.01, training accuracy is 0.406250, time 57.016 samples/sec
+epoch 9, total_step 141480, total loss is 19.16 , inference loss is 12.15, weight deacy loss is 7.01, training accuracy is 0.375000, time 56.598 samples/sec
+epoch 9, total_step 141500, total loss is 16.31 , inference loss is 9.30, weight deacy loss is 7.01, training accuracy is 0.406250, time 56.826 samples/sec
+epoch 9, total_step 141520, total loss is 11.52 , inference loss is 4.51, weight deacy loss is 7.01, training accuracy is 0.562500, time 56.546 samples/sec
+epoch 9, total_step 141540, total loss is 15.21 , inference loss is 8.19, weight deacy loss is 7.01, training accuracy is 0.531250, time 56.379 samples/sec
+epoch 9, total_step 141560, total loss is 15.10 , inference loss is 8.09, weight deacy loss is 7.01, training accuracy is 0.406250, time 56.654 samples/sec
+epoch 9, total_step 141580, total loss is 14.54 , inference loss is 7.53, weight deacy loss is 7.01, training accuracy is 0.625000, time 56.373 samples/sec
+epoch 9, total_step 141600, total loss is 16.73 , inference loss is 9.72, weight deacy loss is 7.01, training accuracy is 0.375000, time 57.699 samples/sec
+epoch 9, total_step 141620, total loss is 15.33 , inference loss is 8.32, weight deacy loss is 7.01, training accuracy is 0.531250, time 56.904 samples/sec
+epoch 9, total_step 141640, total loss is 14.91 , inference loss is 7.90, weight deacy loss is 7.01, training accuracy is 0.406250, time 55.837 samples/sec
+epoch 9, total_step 141660, total loss is 18.78 , inference loss is 11.77, weight deacy loss is 7.01, training accuracy is 0.343750, time 56.959 samples/sec
+epoch 9, total_step 141680, total loss is 16.24 , inference loss is 9.23, weight deacy loss is 7.01, training accuracy is 0.406250, time 55.947 samples/sec
+epoch 9, total_step 141700, total loss is 14.07 , inference loss is 7.06, weight deacy loss is 7.01, training accuracy is 0.468750, time 56.400 samples/sec
+epoch 9, total_step 141720, total loss is 17.18 , inference loss is 10.17, weight deacy loss is 7.01, training accuracy is 0.406250, time 56.328 samples/sec
+epoch 9, total_step 141740, total loss is 21.53 , inference loss is 14.52, weight deacy loss is 7.01, training accuracy is 0.281250, time 56.739 samples/sec
+epoch 9, total_step 141760, total loss is 19.86 , inference loss is 12.85, weight deacy loss is 7.01, training accuracy is 0.312500, time 56.590 samples/sec
+epoch 9, total_step 141780, total loss is 13.20 , inference loss is 6.19, weight deacy loss is 7.01, training accuracy is 0.562500, time 56.497 samples/sec
+epoch 9, total_step 141800, total loss is 10.80 , inference loss is 3.79, weight deacy loss is 7.01, training accuracy is 0.562500, time 56.886 samples/sec
+epoch 9, total_step 141820, total loss is 11.68 , inference loss is 4.67, weight deacy loss is 7.01, training accuracy is 0.500000, time 56.709 samples/sec
+epoch 9, total_step 141840, total loss is 12.31 , inference loss is 5.31, weight deacy loss is 7.01, training accuracy is 0.593750, time 57.018 samples/sec
+epoch 9, total_step 141860, total loss is 15.95 , inference loss is 8.95, weight deacy loss is 7.01, training accuracy is 0.500000, time 56.448 samples/sec
+epoch 9, total_step 141880, total loss is 19.25 , inference loss is 12.24, weight deacy loss is 7.01, training accuracy is 0.406250, time 56.511 samples/sec
+epoch 9, total_step 141900, total loss is 16.52 , inference loss is 9.51, weight deacy loss is 7.01, training accuracy is 0.312500, time 56.205 samples/sec
+epoch 9, total_step 141920, total loss is 15.83 , inference loss is 8.83, weight deacy loss is 7.01, training accuracy is 0.468750, time 55.277 samples/sec
+epoch 9, total_step 141940, total loss is 13.41 , inference loss is 6.40, weight deacy loss is 7.01, training accuracy is 0.437500, time 56.701 samples/sec
+epoch 9, total_step 141960, total loss is 19.60 , inference loss is 12.60, weight deacy loss is 7.01, training accuracy is 0.343750, time 57.175 samples/sec
+epoch 9, total_step 141980, total loss is 14.53 , inference loss is 7.52, weight deacy loss is 7.01, training accuracy is 0.468750, time 56.563 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.4602470000001
+best_threshold_index 122 0.9872222222222222
+best_threshold_index 122 0.9859259259259259
+best_threshold_index 122 0.9864814814814815
+best_threshold_index 121 0.9866666666666667
+best_threshold_index 118 0.9864814814814815
+best_threshold_index 122 0.9861111111111112
+best_threshold_index 120 0.9872222222222222
+best_threshold_index 122 0.9859259259259259
+best_threshold_index 122 0.9851851851851852
+best_threshold_index 122 0.9861111111111112
+[lfw][142000]XNorm: 21.416350
+[lfw][142000]Accuracy-Flip: 0.98500+-0.00658
+testing verification..
+(14000, 512)
+infer time 86.2714489999999
+best_threshold_index 125 0.9882539682539683
+best_threshold_index 125 0.9876190476190476
+best_threshold_index 125 0.9882539682539683
+best_threshold_index 119 0.9876190476190476
+best_threshold_index 125 0.9882539682539683
+best_threshold_index 125 0.9884126984126984
+best_threshold_index 123 0.9874603174603175
+best_threshold_index 125 0.9873015873015873
+best_threshold_index 125 0.9874603174603175
+best_threshold_index 125 0.9882539682539683
+[cfp_ff][142000]XNorm: 21.064214
+[cfp_ff][142000]Accuracy-Flip: 0.98657+-0.00357
+testing verification..
+(14000, 512)
+infer time 86.28457000000003
+best_threshold_index 143 0.9244444444444444
+best_threshold_index 143 0.923968253968254
+best_threshold_index 143 0.9225396825396825
+best_threshold_index 143 0.9247619047619048
+best_threshold_index 143 0.9287301587301587
+best_threshold_index 143 0.9230158730158731
+best_threshold_index 143 0.9215873015873016
+best_threshold_index 143 0.9242857142857143
+best_threshold_index 143 0.9234920634920635
+best_threshold_index 143 0.9246031746031746
+[cfp_fp][142000]XNorm: 21.226528
+[cfp_fp][142000]Accuracy-Flip: 0.92414+-0.01624
+testing verification..
+(12000, 512)
+infer time 74.27712500000015
+best_threshold_index 146 0.9
+best_threshold_index 146 0.8981481481481481
+best_threshold_index 146 0.9009259259259259
+best_threshold_index 137 0.9009259259259259
+best_threshold_index 137 0.9055555555555556
+best_threshold_index 146 0.9
+best_threshold_index 146 0.9009259259259259
+best_threshold_index 143 0.9066666666666666
+best_threshold_index 146 0.9055555555555556
+best_threshold_index 137 0.9051851851851852
+[agedb_30][142000]XNorm: 22.423330
+[agedb_30][142000]Accuracy-Flip: 0.89800+-0.02763
+test accuracy is: 0.985
+epoch 9, total_step 142000, total loss is 15.19 , inference loss is 8.18, weight deacy loss is 7.01, training accuracy is 0.468750, time 53.319 samples/sec
+epoch 9, total_step 142020, total loss is 13.21 , inference loss is 6.20, weight deacy loss is 7.00, training accuracy is 0.375000, time 55.657 samples/sec
+epoch 9, total_step 142040, total loss is 17.38 , inference loss is 10.38, weight deacy loss is 7.00, training accuracy is 0.312500, time 56.245 samples/sec
+epoch 9, total_step 142060, total loss is 16.02 , inference loss is 9.01, weight deacy loss is 7.00, training accuracy is 0.375000, time 56.331 samples/sec
+epoch 9, total_step 142080, total loss is 15.47 , inference loss is 8.47, weight deacy loss is 7.00, training accuracy is 0.468750, time 56.513 samples/sec
+epoch 9, total_step 142100, total loss is 16.21 , inference loss is 9.21, weight deacy loss is 7.00, training accuracy is 0.437500, time 56.864 samples/sec
+epoch 9, total_step 142120, total loss is 14.88 , inference loss is 7.88, weight deacy loss is 7.00, training accuracy is 0.468750, time 56.655 samples/sec
+epoch 9, total_step 142140, total loss is 16.10 , inference loss is 9.10, weight deacy loss is 7.00, training accuracy is 0.500000, time 55.836 samples/sec
+epoch 9, total_step 142160, total loss is 19.59 , inference loss is 12.59, weight deacy loss is 7.00, training accuracy is 0.468750, time 56.743 samples/sec
+epoch 9, total_step 142180, total loss is 15.34 , inference loss is 8.33, weight deacy loss is 7.00, training accuracy is 0.531250, time 56.689 samples/sec
+epoch 9, total_step 142200, total loss is 13.57 , inference loss is 6.57, weight deacy loss is 7.00, training accuracy is 0.625000, time 56.273 samples/sec
+epoch 9, total_step 142220, total loss is 13.48 , inference loss is 6.48, weight deacy loss is 7.00, training accuracy is 0.468750, time 56.341 samples/sec
+epoch 9, total_step 142240, total loss is 13.46 , inference loss is 6.46, weight deacy loss is 7.00, training accuracy is 0.406250, time 56.708 samples/sec
+epoch 9, total_step 142260, total loss is 13.99 , inference loss is 6.98, weight deacy loss is 7.00, training accuracy is 0.468750, time 57.077 samples/sec
+epoch 9, total_step 142280, total loss is 20.02 , inference loss is 13.02, weight deacy loss is 7.00, training accuracy is 0.156250, time 56.716 samples/sec
+epoch 9, total_step 142300, total loss is 15.51 , inference loss is 8.50, weight deacy loss is 7.00, training accuracy is 0.250000, time 57.414 samples/sec
+epoch 9, total_step 142320, total loss is 15.54 , inference loss is 8.54, weight deacy loss is 7.00, training accuracy is 0.375000, time 56.219 samples/sec
+epoch 9, total_step 142340, total loss is 18.92 , inference loss is 11.92, weight deacy loss is 7.00, training accuracy is 0.562500, time 56.194 samples/sec
+epoch 9, total_step 142360, total loss is 13.73 , inference loss is 6.73, weight deacy loss is 7.00, training accuracy is 0.437500, time 55.971 samples/sec
+epoch 9, total_step 142380, total loss is 13.11 , inference loss is 6.11, weight deacy loss is 7.00, training accuracy is 0.406250, time 56.152 samples/sec
+epoch 9, total_step 142400, total loss is 16.54 , inference loss is 9.54, weight deacy loss is 7.00, training accuracy is 0.406250, time 57.385 samples/sec
+epoch 9, total_step 142420, total loss is 16.11 , inference loss is 9.11, weight deacy loss is 7.00, training accuracy is 0.406250, time 56.489 samples/sec
+epoch 9, total_step 142440, total loss is 18.06 , inference loss is 11.06, weight deacy loss is 7.00, training accuracy is 0.375000, time 56.980 samples/sec
+epoch 9, total_step 142460, total loss is 17.42 , inference loss is 10.42, weight deacy loss is 7.00, training accuracy is 0.593750, time 56.258 samples/sec
+epoch 9, total_step 142480, total loss is 17.50 , inference loss is 10.50, weight deacy loss is 7.00, training accuracy is 0.375000, time 56.173 samples/sec
+epoch 9, total_step 142500, total loss is 15.54 , inference loss is 8.54, weight deacy loss is 7.00, training accuracy is 0.312500, time 56.256 samples/sec
+epoch 9, total_step 142520, total loss is 15.82 , inference loss is 8.82, weight deacy loss is 7.00, training accuracy is 0.375000, time 56.289 samples/sec
+epoch 9, total_step 142540, total loss is 13.24 , inference loss is 6.24, weight deacy loss is 7.00, training accuracy is 0.500000, time 56.495 samples/sec
+epoch 9, total_step 142560, total loss is 12.83 , inference loss is 5.84, weight deacy loss is 7.00, training accuracy is 0.406250, time 56.484 samples/sec
+epoch 9, total_step 142580, total loss is 13.91 , inference loss is 6.91, weight deacy loss is 7.00, training accuracy is 0.562500, time 55.913 samples/sec
+epoch 9, total_step 142600, total loss is 19.50 , inference loss is 12.50, weight deacy loss is 7.00, training accuracy is 0.312500, time 56.993 samples/sec
+epoch 9, total_step 142620, total loss is 12.71 , inference loss is 5.71, weight deacy loss is 7.00, training accuracy is 0.500000, time 56.425 samples/sec
+epoch 9, total_step 142640, total loss is 16.71 , inference loss is 9.72, weight deacy loss is 7.00, training accuracy is 0.375000, time 55.833 samples/sec
+epoch 9, total_step 142660, total loss is 18.39 , inference loss is 11.40, weight deacy loss is 6.99, training accuracy is 0.375000, time 54.848 samples/sec
+epoch 9, total_step 142680, total loss is 14.00 , inference loss is 7.01, weight deacy loss is 6.99, training accuracy is 0.562500, time 56.159 samples/sec
+epoch 9, total_step 142700, total loss is 17.58 , inference loss is 10.58, weight deacy loss is 6.99, training accuracy is 0.343750, time 56.538 samples/sec
+epoch 9, total_step 142720, total loss is 17.52 , inference loss is 10.53, weight deacy loss is 6.99, training accuracy is 0.406250, time 56.543 samples/sec
+epoch 9, total_step 142740, total loss is 11.98 , inference loss is 4.98, weight deacy loss is 6.99, training accuracy is 0.531250, time 56.907 samples/sec
+epoch 9, total_step 142760, total loss is 14.86 , inference loss is 7.87, weight deacy loss is 6.99, training accuracy is 0.437500, time 56.383 samples/sec
+epoch 9, total_step 142780, total loss is 14.49 , inference loss is 7.50, weight deacy loss is 6.99, training accuracy is 0.437500, time 56.792 samples/sec
+epoch 9, total_step 142800, total loss is 17.46 , inference loss is 10.47, weight deacy loss is 6.99, training accuracy is 0.468750, time 55.850 samples/sec
+epoch 9, total_step 142820, total loss is 17.33 , inference loss is 10.33, weight deacy loss is 6.99, training accuracy is 0.343750, time 56.154 samples/sec
+epoch 9, total_step 142840, total loss is 14.45 , inference loss is 7.45, weight deacy loss is 6.99, training accuracy is 0.500000, time 56.725 samples/sec
+epoch 9, total_step 142860, total loss is 14.83 , inference loss is 7.84, weight deacy loss is 6.99, training accuracy is 0.468750, time 56.827 samples/sec
+epoch 9, total_step 142880, total loss is 17.16 , inference loss is 10.17, weight deacy loss is 6.99, training accuracy is 0.312500, time 56.402 samples/sec
+epoch 9, total_step 142900, total loss is 13.13 , inference loss is 6.13, weight deacy loss is 6.99, training accuracy is 0.500000, time 56.295 samples/sec
+epoch 9, total_step 142920, total loss is 19.60 , inference loss is 12.61, weight deacy loss is 6.99, training accuracy is 0.250000, time 56.543 samples/sec
+epoch 9, total_step 142940, total loss is 16.69 , inference loss is 9.70, weight deacy loss is 6.99, training accuracy is 0.468750, time 56.391 samples/sec
+epoch 9, total_step 142960, total loss is 20.05 , inference loss is 13.06, weight deacy loss is 6.99, training accuracy is 0.312500, time 56.467 samples/sec
+epoch 9, total_step 142980, total loss is 16.26 , inference loss is 9.27, weight deacy loss is 6.99, training accuracy is 0.281250, time 56.329 samples/sec
+epoch 9, total_step 143000, total loss is 16.83 , inference loss is 9.84, weight deacy loss is 6.99, training accuracy is 0.468750, time 55.307 samples/sec
+epoch 9, total_step 143020, total loss is 15.77 , inference loss is 8.78, weight deacy loss is 6.99, training accuracy is 0.437500, time 56.301 samples/sec
+epoch 9, total_step 143040, total loss is 16.89 , inference loss is 9.90, weight deacy loss is 6.99, training accuracy is 0.312500, time 56.800 samples/sec
+epoch 9, total_step 143060, total loss is 14.92 , inference loss is 7.94, weight deacy loss is 6.99, training accuracy is 0.500000, time 57.366 samples/sec
+epoch 9, total_step 143080, total loss is 18.84 , inference loss is 11.85, weight deacy loss is 6.99, training accuracy is 0.375000, time 56.233 samples/sec
+epoch 9, total_step 143100, total loss is 15.18 , inference loss is 8.19, weight deacy loss is 6.99, training accuracy is 0.468750, time 57.001 samples/sec
+epoch 9, total_step 143120, total loss is 13.41 , inference loss is 6.42, weight deacy loss is 6.99, training accuracy is 0.406250, time 56.420 samples/sec
+epoch 9, total_step 143140, total loss is 18.80 , inference loss is 11.82, weight deacy loss is 6.99, training accuracy is 0.281250, time 56.515 samples/sec
+epoch 9, total_step 143160, total loss is 12.96 , inference loss is 5.97, weight deacy loss is 6.99, training accuracy is 0.625000, time 56.516 samples/sec
+epoch 9, total_step 143180, total loss is 12.93 , inference loss is 5.95, weight deacy loss is 6.99, training accuracy is 0.531250, time 56.517 samples/sec
+epoch 9, total_step 143200, total loss is 14.51 , inference loss is 7.52, weight deacy loss is 6.99, training accuracy is 0.531250, time 56.669 samples/sec
+epoch 9, total_step 143220, total loss is 14.62 , inference loss is 7.63, weight deacy loss is 6.99, training accuracy is 0.437500, time 56.757 samples/sec
+epoch 9, total_step 143240, total loss is 14.08 , inference loss is 7.09, weight deacy loss is 6.99, training accuracy is 0.437500, time 56.832 samples/sec
+epoch 9, total_step 143260, total loss is 16.40 , inference loss is 9.41, weight deacy loss is 6.98, training accuracy is 0.437500, time 56.479 samples/sec
+epoch 9, total_step 143280, total loss is 15.85 , inference loss is 8.86, weight deacy loss is 6.98, training accuracy is 0.468750, time 56.584 samples/sec
+epoch 9, total_step 143300, total loss is 18.45 , inference loss is 11.46, weight deacy loss is 6.98, training accuracy is 0.281250, time 56.023 samples/sec
+epoch 9, total_step 143320, total loss is 15.23 , inference loss is 8.24, weight deacy loss is 6.98, training accuracy is 0.531250, time 57.093 samples/sec
+epoch 9, total_step 143340, total loss is 21.11 , inference loss is 14.12, weight deacy loss is 6.98, training accuracy is 0.312500, time 57.022 samples/sec
+epoch 9, total_step 143360, total loss is 14.70 , inference loss is 7.71, weight deacy loss is 6.98, training accuracy is 0.500000, time 56.803 samples/sec
+epoch 9, total_step 143380, total loss is 18.68 , inference loss is 11.69, weight deacy loss is 6.98, training accuracy is 0.312500, time 56.404 samples/sec
+epoch 9, total_step 143400, total loss is 13.93 , inference loss is 6.95, weight deacy loss is 6.98, training accuracy is 0.468750, time 56.642 samples/sec
+epoch 9, total_step 143420, total loss is 16.40 , inference loss is 9.42, weight deacy loss is 6.98, training accuracy is 0.531250, time 56.541 samples/sec
+epoch 9, total_step 143440, total loss is 16.81 , inference loss is 9.82, weight deacy loss is 6.98, training accuracy is 0.500000, time 56.684 samples/sec
+epoch 9, total_step 143460, total loss is 14.68 , inference loss is 7.70, weight deacy loss is 6.98, training accuracy is 0.500000, time 56.440 samples/sec
+epoch 9, total_step 143480, total loss is 17.55 , inference loss is 10.57, weight deacy loss is 6.98, training accuracy is 0.406250, time 56.657 samples/sec
+epoch 9, total_step 143500, total loss is 15.20 , inference loss is 8.22, weight deacy loss is 6.98, training accuracy is 0.406250, time 56.414 samples/sec
+epoch 9, total_step 143520, total loss is 14.96 , inference loss is 7.98, weight deacy loss is 6.98, training accuracy is 0.468750, time 56.317 samples/sec
+epoch 9, total_step 143540, total loss is 16.44 , inference loss is 9.46, weight deacy loss is 6.98, training accuracy is 0.437500, time 56.127 samples/sec
+epoch 9, total_step 143560, total loss is 14.21 , inference loss is 7.23, weight deacy loss is 6.98, training accuracy is 0.437500, time 56.924 samples/sec
+epoch 9, total_step 143580, total loss is 12.79 , inference loss is 5.81, weight deacy loss is 6.98, training accuracy is 0.406250, time 57.062 samples/sec
+epoch 9, total_step 143600, total loss is 15.08 , inference loss is 8.10, weight deacy loss is 6.98, training accuracy is 0.437500, time 56.467 samples/sec
+epoch 9, total_step 143620, total loss is 14.11 , inference loss is 7.13, weight deacy loss is 6.98, training accuracy is 0.531250, time 56.625 samples/sec
+epoch 9, total_step 143640, total loss is 16.52 , inference loss is 9.55, weight deacy loss is 6.98, training accuracy is 0.375000, time 56.982 samples/sec
+epoch 9, total_step 143660, total loss is 16.66 , inference loss is 9.69, weight deacy loss is 6.98, training accuracy is 0.500000, time 56.814 samples/sec
+epoch 9, total_step 143680, total loss is 12.89 , inference loss is 5.92, weight deacy loss is 6.98, training accuracy is 0.468750, time 56.474 samples/sec
+epoch 9, total_step 143700, total loss is 15.13 , inference loss is 8.15, weight deacy loss is 6.98, training accuracy is 0.437500, time 56.445 samples/sec
+epoch 9, total_step 143720, total loss is 16.99 , inference loss is 10.02, weight deacy loss is 6.98, training accuracy is 0.437500, time 57.656 samples/sec
+epoch 9, total_step 143740, total loss is 13.62 , inference loss is 6.65, weight deacy loss is 6.98, training accuracy is 0.562500, time 55.275 samples/sec
+epoch 9, total_step 143760, total loss is 13.56 , inference loss is 6.59, weight deacy loss is 6.98, training accuracy is 0.562500, time 55.566 samples/sec
+epoch 9, total_step 143780, total loss is 13.68 , inference loss is 6.71, weight deacy loss is 6.98, training accuracy is 0.468750, time 56.070 samples/sec
+epoch 9, total_step 143800, total loss is 15.86 , inference loss is 8.88, weight deacy loss is 6.98, training accuracy is 0.437500, time 56.965 samples/sec
+epoch 9, total_step 143820, total loss is 15.21 , inference loss is 8.23, weight deacy loss is 6.97, training accuracy is 0.468750, time 56.650 samples/sec
+epoch 9, total_step 143840, total loss is 13.49 , inference loss is 6.52, weight deacy loss is 6.97, training accuracy is 0.375000, time 55.943 samples/sec
+epoch 9, total_step 143860, total loss is 13.59 , inference loss is 6.61, weight deacy loss is 6.97, training accuracy is 0.468750, time 57.021 samples/sec
+epoch 9, total_step 143880, total loss is 15.44 , inference loss is 8.47, weight deacy loss is 6.97, training accuracy is 0.312500, time 55.758 samples/sec
+epoch 9, total_step 143900, total loss is 15.15 , inference loss is 8.18, weight deacy loss is 6.97, training accuracy is 0.500000, time 56.336 samples/sec
+epoch 9, total_step 143920, total loss is 14.22 , inference loss is 7.25, weight deacy loss is 6.97, training accuracy is 0.375000, time 56.668 samples/sec
+epoch 9, total_step 143940, total loss is 14.55 , inference loss is 7.58, weight deacy loss is 6.97, training accuracy is 0.406250, time 56.789 samples/sec
+epoch 9, total_step 143960, total loss is 17.28 , inference loss is 10.31, weight deacy loss is 6.97, training accuracy is 0.343750, time 56.730 samples/sec
+epoch 9, total_step 143980, total loss is 16.70 , inference loss is 9.72, weight deacy loss is 6.97, training accuracy is 0.375000, time 56.355 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.98215299999993
+best_threshold_index 125 0.9875925925925926
+best_threshold_index 117 0.9861111111111112
+best_threshold_index 117 0.9864814814814815
+best_threshold_index 116 0.9872222222222222
+best_threshold_index 125 0.9874074074074074
+best_threshold_index 125 0.9859259259259259
+best_threshold_index 117 0.9872222222222222
+best_threshold_index 117 0.9861111111111112
+best_threshold_index 125 0.9853703703703703
+best_threshold_index 117 0.9861111111111112
+[lfw][144000]XNorm: 21.624872
+[lfw][144000]Accuracy-Flip: 0.98417+-0.00659
+testing verification..
+(14000, 512)
+infer time 86.02158299999995
+best_threshold_index 117 0.986984126984127
+best_threshold_index 117 0.984920634920635
+best_threshold_index 117 0.9858730158730159
+best_threshold_index 120 0.984920634920635
+best_threshold_index 117 0.9866666666666667
+best_threshold_index 117 0.9855555555555555
+best_threshold_index 120 0.9866666666666667
+best_threshold_index 117 0.9855555555555555
+best_threshold_index 117 0.9852380952380952
+best_threshold_index 117 0.9868253968253968
+[cfp_ff][144000]XNorm: 20.995785
+[cfp_ff][144000]Accuracy-Flip: 0.98500+-0.00683
+testing verification..
+(14000, 512)
+infer time 85.88527299999993
+best_threshold_index 146 0.9265079365079365
+best_threshold_index 146 0.9263492063492064
+best_threshold_index 146 0.9249206349206349
+best_threshold_index 146 0.9261904761904762
+best_threshold_index 146 0.9277777777777778
+best_threshold_index 146 0.9253968253968254
+best_threshold_index 146 0.9242857142857143
+best_threshold_index 146 0.9250793650793651
+best_threshold_index 146 0.9263492063492064
+best_threshold_index 146 0.9271428571428572
+[cfp_fp][144000]XNorm: 21.740437
+[cfp_fp][144000]Accuracy-Flip: 0.92600+-0.00914
+testing verification..
+(12000, 512)
+infer time 73.75317000000004
+best_threshold_index 139 0.9016666666666666
+best_threshold_index 139 0.9018518518518519
+best_threshold_index 139 0.9062962962962963
+best_threshold_index 139 0.9048148148148148
+best_threshold_index 139 0.907962962962963
+best_threshold_index 139 0.905
+best_threshold_index 139 0.9057407407407407
+best_threshold_index 140 0.9081481481481481
+best_threshold_index 139 0.907962962962963
+best_threshold_index 139 0.9057407407407407
+[agedb_30][144000]XNorm: 22.850665
+[agedb_30][144000]Accuracy-Flip: 0.90483+-0.02049
+test accuracy is: 0.9841666666666666
+epoch 9, total_step 144000, total loss is 18.74 , inference loss is 11.77, weight deacy loss is 6.97, training accuracy is 0.312500, time 52.873 samples/sec
+epoch 9, total_step 144020, total loss is 16.36 , inference loss is 9.39, weight deacy loss is 6.97, training accuracy is 0.500000, time 55.504 samples/sec
+epoch 9, total_step 144040, total loss is 14.39 , inference loss is 7.42, weight deacy loss is 6.97, training accuracy is 0.468750, time 56.010 samples/sec
+epoch 9, total_step 144060, total loss is 16.40 , inference loss is 9.43, weight deacy loss is 6.97, training accuracy is 0.406250, time 56.281 samples/sec
+epoch 9, total_step 144080, total loss is 14.95 , inference loss is 7.98, weight deacy loss is 6.97, training accuracy is 0.281250, time 56.306 samples/sec
+epoch 9, total_step 144100, total loss is 16.74 , inference loss is 9.77, weight deacy loss is 6.97, training accuracy is 0.375000, time 56.905 samples/sec
+epoch 9, total_step 144120, total loss is 15.19 , inference loss is 8.22, weight deacy loss is 6.97, training accuracy is 0.343750, time 56.625 samples/sec
+epoch 9, total_step 144140, total loss is 15.83 , inference loss is 8.86, weight deacy loss is 6.97, training accuracy is 0.406250, time 56.341 samples/sec
+epoch 9, total_step 144160, total loss is 11.29 , inference loss is 4.32, weight deacy loss is 6.97, training accuracy is 0.562500, time 56.995 samples/sec
+epoch 9, total_step 144180, total loss is 15.32 , inference loss is 8.35, weight deacy loss is 6.97, training accuracy is 0.406250, time 56.287 samples/sec
+epoch 9, total_step 144200, total loss is 18.19 , inference loss is 11.22, weight deacy loss is 6.97, training accuracy is 0.343750, time 56.282 samples/sec
+epoch 9, total_step 144220, total loss is 15.76 , inference loss is 8.79, weight deacy loss is 6.97, training accuracy is 0.468750, time 56.506 samples/sec
+epoch 9, total_step 144240, total loss is 15.61 , inference loss is 8.65, weight deacy loss is 6.97, training accuracy is 0.531250, time 56.401 samples/sec
+epoch 9, total_step 144260, total loss is 12.78 , inference loss is 5.81, weight deacy loss is 6.97, training accuracy is 0.437500, time 56.341 samples/sec
+epoch 9, total_step 144280, total loss is 17.75 , inference loss is 10.79, weight deacy loss is 6.97, training accuracy is 0.406250, time 57.096 samples/sec
+epoch 9, total_step 144300, total loss is 14.16 , inference loss is 7.19, weight deacy loss is 6.97, training accuracy is 0.312500, time 56.697 samples/sec
+epoch 9, total_step 144320, total loss is 14.99 , inference loss is 8.02, weight deacy loss is 6.97, training accuracy is 0.437500, time 56.685 samples/sec
+epoch 9, total_step 144340, total loss is 13.56 , inference loss is 6.59, weight deacy loss is 6.97, training accuracy is 0.468750, time 57.078 samples/sec
+epoch 9, total_step 144360, total loss is 12.42 , inference loss is 5.46, weight deacy loss is 6.97, training accuracy is 0.468750, time 56.400 samples/sec
+epoch 9, total_step 144380, total loss is 15.10 , inference loss is 8.14, weight deacy loss is 6.97, training accuracy is 0.343750, time 55.871 samples/sec
+epoch 9, total_step 144400, total loss is 17.30 , inference loss is 10.33, weight deacy loss is 6.96, training accuracy is 0.437500, time 56.650 samples/sec
+epoch 9, total_step 144420, total loss is 13.31 , inference loss is 6.35, weight deacy loss is 6.96, training accuracy is 0.468750, time 56.402 samples/sec
+epoch 9, total_step 144440, total loss is 18.01 , inference loss is 11.05, weight deacy loss is 6.96, training accuracy is 0.250000, time 56.522 samples/sec
+epoch 9, total_step 144460, total loss is 16.37 , inference loss is 9.41, weight deacy loss is 6.96, training accuracy is 0.281250, time 56.101 samples/sec
+epoch 9, total_step 144480, total loss is 15.33 , inference loss is 8.36, weight deacy loss is 6.96, training accuracy is 0.406250, time 55.881 samples/sec
+epoch 9, total_step 144500, total loss is 19.26 , inference loss is 12.30, weight deacy loss is 6.96, training accuracy is 0.343750, time 57.501 samples/sec
+epoch 9, total_step 144520, total loss is 13.19 , inference loss is 6.23, weight deacy loss is 6.96, training accuracy is 0.625000, time 56.455 samples/sec
+epoch 9, total_step 144540, total loss is 13.73 , inference loss is 6.77, weight deacy loss is 6.96, training accuracy is 0.468750, time 56.490 samples/sec
+epoch 9, total_step 144560, total loss is 18.20 , inference loss is 11.23, weight deacy loss is 6.96, training accuracy is 0.468750, time 56.068 samples/sec
+epoch 9, total_step 144580, total loss is 13.17 , inference loss is 6.21, weight deacy loss is 6.96, training accuracy is 0.562500, time 56.529 samples/sec
+epoch 9, total_step 144600, total loss is 14.03 , inference loss is 7.07, weight deacy loss is 6.96, training accuracy is 0.437500, time 56.669 samples/sec
+epoch 9, total_step 144620, total loss is 17.47 , inference loss is 10.51, weight deacy loss is 6.96, training accuracy is 0.468750, time 56.281 samples/sec
+epoch 9, total_step 144640, total loss is 17.74 , inference loss is 10.78, weight deacy loss is 6.96, training accuracy is 0.281250, time 56.304 samples/sec
+epoch 9, total_step 144660, total loss is 17.55 , inference loss is 10.59, weight deacy loss is 6.96, training accuracy is 0.343750, time 56.336 samples/sec
+epoch 9, total_step 144680, total loss is 14.70 , inference loss is 7.74, weight deacy loss is 6.96, training accuracy is 0.406250, time 57.351 samples/sec
+epoch 9, total_step 144700, total loss is 17.85 , inference loss is 10.89, weight deacy loss is 6.96, training accuracy is 0.468750, time 56.552 samples/sec
+epoch 9, total_step 144720, total loss is 17.29 , inference loss is 10.33, weight deacy loss is 6.96, training accuracy is 0.281250, time 56.860 samples/sec
+epoch 9, total_step 144740, total loss is 17.41 , inference loss is 10.45, weight deacy loss is 6.96, training accuracy is 0.468750, time 56.780 samples/sec
+epoch 9, total_step 144760, total loss is 11.40 , inference loss is 4.45, weight deacy loss is 6.96, training accuracy is 0.562500, time 56.472 samples/sec
+epoch 9, total_step 144780, total loss is 19.36 , inference loss is 12.40, weight deacy loss is 6.96, training accuracy is 0.312500, time 55.892 samples/sec
+epoch 9, total_step 144800, total loss is 11.44 , inference loss is 4.48, weight deacy loss is 6.96, training accuracy is 0.562500, time 55.181 samples/sec
+epoch 9, total_step 144820, total loss is 14.63 , inference loss is 7.67, weight deacy loss is 6.96, training accuracy is 0.531250, time 56.643 samples/sec
+epoch 9, total_step 144840, total loss is 15.62 , inference loss is 8.66, weight deacy loss is 6.96, training accuracy is 0.437500, time 56.401 samples/sec
+epoch 9, total_step 144860, total loss is 14.35 , inference loss is 7.39, weight deacy loss is 6.96, training accuracy is 0.468750, time 57.684 samples/sec
+epoch 9, total_step 144880, total loss is 12.81 , inference loss is 5.86, weight deacy loss is 6.96, training accuracy is 0.500000, time 57.378 samples/sec
+epoch 9, total_step 144900, total loss is 16.36 , inference loss is 9.40, weight deacy loss is 6.96, training accuracy is 0.343750, time 55.857 samples/sec
+epoch 9, total_step 144920, total loss is 12.24 , inference loss is 5.28, weight deacy loss is 6.96, training accuracy is 0.562500, time 56.221 samples/sec
+epoch 9, total_step 144940, total loss is 17.09 , inference loss is 10.14, weight deacy loss is 6.96, training accuracy is 0.218750, time 56.747 samples/sec
+epoch 9, total_step 144960, total loss is 14.19 , inference loss is 7.23, weight deacy loss is 6.96, training accuracy is 0.437500, time 55.395 samples/sec
+epoch 9, total_step 144980, total loss is 14.50 , inference loss is 7.54, weight deacy loss is 6.96, training accuracy is 0.375000, time 55.828 samples/sec
+epoch 9, total_step 145000, total loss is 15.12 , inference loss is 8.16, weight deacy loss is 6.96, training accuracy is 0.406250, time 56.577 samples/sec
+epoch 9, total_step 145020, total loss is 15.95 , inference loss is 8.99, weight deacy loss is 6.95, training accuracy is 0.375000, time 56.460 samples/sec
+epoch 9, total_step 145040, total loss is 16.72 , inference loss is 9.76, weight deacy loss is 6.95, training accuracy is 0.312500, time 56.994 samples/sec
+epoch 9, total_step 145060, total loss is 12.51 , inference loss is 5.55, weight deacy loss is 6.95, training accuracy is 0.500000, time 56.431 samples/sec
+epoch 9, total_step 145080, total loss is 14.52 , inference loss is 7.57, weight deacy loss is 6.95, training accuracy is 0.531250, time 56.500 samples/sec
+epoch 9, total_step 145100, total loss is 14.78 , inference loss is 7.83, weight deacy loss is 6.95, training accuracy is 0.562500, time 56.439 samples/sec
+epoch 9, total_step 145120, total loss is 16.01 , inference loss is 9.06, weight deacy loss is 6.95, training accuracy is 0.406250, time 56.693 samples/sec
+epoch 9, total_step 145140, total loss is 14.65 , inference loss is 7.70, weight deacy loss is 6.95, training accuracy is 0.562500, time 56.517 samples/sec
+epoch 9, total_step 145160, total loss is 13.42 , inference loss is 6.47, weight deacy loss is 6.95, training accuracy is 0.406250, time 57.142 samples/sec
+epoch 9, total_step 145180, total loss is 15.14 , inference loss is 8.18, weight deacy loss is 6.95, training accuracy is 0.531250, time 57.290 samples/sec
+epoch 9, total_step 145200, total loss is 14.56 , inference loss is 7.61, weight deacy loss is 6.95, training accuracy is 0.500000, time 56.665 samples/sec
+epoch 9, total_step 145220, total loss is 16.02 , inference loss is 9.07, weight deacy loss is 6.95, training accuracy is 0.437500, time 56.087 samples/sec
+epoch 9, total_step 145240, total loss is 18.05 , inference loss is 11.10, weight deacy loss is 6.95, training accuracy is 0.343750, time 56.747 samples/sec
+epoch 9, total_step 145260, total loss is 16.21 , inference loss is 9.26, weight deacy loss is 6.95, training accuracy is 0.468750, time 56.447 samples/sec
+epoch 9, total_step 145280, total loss is 17.10 , inference loss is 10.15, weight deacy loss is 6.95, training accuracy is 0.437500, time 57.260 samples/sec
+epoch 9, total_step 145300, total loss is 14.71 , inference loss is 7.76, weight deacy loss is 6.95, training accuracy is 0.375000, time 54.791 samples/sec
+epoch 9, total_step 145320, total loss is 13.92 , inference loss is 6.97, weight deacy loss is 6.95, training accuracy is 0.406250, time 55.854 samples/sec
+epoch 9, total_step 145340, total loss is 15.59 , inference loss is 8.64, weight deacy loss is 6.95, training accuracy is 0.375000, time 56.582 samples/sec
+epoch 9, total_step 145360, total loss is 14.41 , inference loss is 7.46, weight deacy loss is 6.95, training accuracy is 0.312500, time 56.050 samples/sec
+epoch 9, total_step 145380, total loss is 13.18 , inference loss is 6.23, weight deacy loss is 6.95, training accuracy is 0.531250, time 56.660 samples/sec
+epoch 9, total_step 145400, total loss is 12.90 , inference loss is 5.95, weight deacy loss is 6.95, training accuracy is 0.468750, time 56.460 samples/sec
+epoch 9, total_step 145420, total loss is 18.91 , inference loss is 11.96, weight deacy loss is 6.95, training accuracy is 0.406250, time 56.366 samples/sec
+epoch 9, total_step 145440, total loss is 16.57 , inference loss is 9.63, weight deacy loss is 6.95, training accuracy is 0.500000, time 57.326 samples/sec
+epoch 9, total_step 145460, total loss is 16.18 , inference loss is 9.24, weight deacy loss is 6.95, training accuracy is 0.343750, time 56.729 samples/sec
+epoch 9, total_step 145480, total loss is 15.67 , inference loss is 8.73, weight deacy loss is 6.95, training accuracy is 0.437500, time 56.420 samples/sec
+epoch 9, total_step 145500, total loss is 19.26 , inference loss is 12.31, weight deacy loss is 6.95, training accuracy is 0.375000, time 56.009 samples/sec
+epoch 9, total_step 145520, total loss is 14.87 , inference loss is 7.93, weight deacy loss is 6.95, training accuracy is 0.406250, time 56.726 samples/sec
+epoch 9, total_step 145540, total loss is 11.75 , inference loss is 4.80, weight deacy loss is 6.95, training accuracy is 0.468750, time 57.114 samples/sec
+epoch 9, total_step 145560, total loss is 13.35 , inference loss is 6.40, weight deacy loss is 6.95, training accuracy is 0.468750, time 55.783 samples/sec
+epoch 9, total_step 145580, total loss is 12.58 , inference loss is 5.63, weight deacy loss is 6.95, training accuracy is 0.468750, time 55.319 samples/sec
+epoch 9, total_step 145600, total loss is 14.46 , inference loss is 7.51, weight deacy loss is 6.94, training accuracy is 0.500000, time 57.249 samples/sec
+epoch 9, total_step 145620, total loss is 18.67 , inference loss is 11.72, weight deacy loss is 6.94, training accuracy is 0.281250, time 57.045 samples/sec
+epoch 9, total_step 145640, total loss is 17.11 , inference loss is 10.16, weight deacy loss is 6.94, training accuracy is 0.406250, time 56.735 samples/sec
+epoch 9, total_step 145660, total loss is 11.20 , inference loss is 4.26, weight deacy loss is 6.94, training accuracy is 0.562500, time 56.596 samples/sec
+epoch 9, total_step 145680, total loss is 15.34 , inference loss is 8.40, weight deacy loss is 6.94, training accuracy is 0.500000, time 56.934 samples/sec
+epoch 9, total_step 145700, total loss is 16.74 , inference loss is 9.80, weight deacy loss is 6.94, training accuracy is 0.437500, time 56.460 samples/sec
+epoch 9, total_step 145720, total loss is 13.91 , inference loss is 6.96, weight deacy loss is 6.94, training accuracy is 0.468750, time 56.653 samples/sec
+epoch 9, total_step 145740, total loss is 12.21 , inference loss is 5.27, weight deacy loss is 6.94, training accuracy is 0.718750, time 57.339 samples/sec
+epoch 9, total_step 145760, total loss is 14.02 , inference loss is 7.08, weight deacy loss is 6.94, training accuracy is 0.500000, time 54.968 samples/sec
+epoch 9, total_step 145780, total loss is 13.89 , inference loss is 6.95, weight deacy loss is 6.94, training accuracy is 0.562500, time 56.741 samples/sec
+epoch 9, total_step 145800, total loss is 18.49 , inference loss is 11.55, weight deacy loss is 6.94, training accuracy is 0.281250, time 55.740 samples/sec
+epoch 9, total_step 145820, total loss is 16.29 , inference loss is 9.35, weight deacy loss is 6.94, training accuracy is 0.406250, time 56.863 samples/sec
+epoch 9, total_step 145840, total loss is 15.43 , inference loss is 8.49, weight deacy loss is 6.94, training accuracy is 0.406250, time 56.368 samples/sec
+epoch 9, total_step 145860, total loss is 18.02 , inference loss is 11.08, weight deacy loss is 6.94, training accuracy is 0.281250, time 56.094 samples/sec
+epoch 9, total_step 145880, total loss is 12.98 , inference loss is 6.04, weight deacy loss is 6.94, training accuracy is 0.562500, time 56.354 samples/sec
+epoch 9, total_step 145900, total loss is 16.19 , inference loss is 9.25, weight deacy loss is 6.94, training accuracy is 0.312500, time 56.668 samples/sec
+epoch 9, total_step 145920, total loss is 17.99 , inference loss is 11.05, weight deacy loss is 6.94, training accuracy is 0.312500, time 56.401 samples/sec
+epoch 9, total_step 145940, total loss is 18.39 , inference loss is 11.45, weight deacy loss is 6.94, training accuracy is 0.375000, time 56.230 samples/sec
+epoch 9, total_step 145960, total loss is 17.03 , inference loss is 10.09, weight deacy loss is 6.94, training accuracy is 0.437500, time 56.983 samples/sec
+epoch 9, total_step 145980, total loss is 18.20 , inference loss is 11.26, weight deacy loss is 6.94, training accuracy is 0.343750, time 55.926 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.11689400000009
+best_threshold_index 115 0.9868518518518519
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 108 0.9855555555555555
+best_threshold_index 114 0.9868518518518519
+best_threshold_index 117 0.9859259259259259
+best_threshold_index 121 0.985
+best_threshold_index 113 0.9857407407407407
+best_threshold_index 113 0.9848148148148148
+best_threshold_index 113 0.9844444444444445
+best_threshold_index 113 0.9842592592592593
+[lfw][146000]XNorm: 21.554732
+[lfw][146000]Accuracy-Flip: 0.98300+-0.00788
+testing verification..
+(14000, 512)
+infer time 86.32585200000001
+best_threshold_index 118 0.9879365079365079
+best_threshold_index 116 0.9873015873015873
+best_threshold_index 116 0.9866666666666667
+best_threshold_index 116 0.9871428571428571
+best_threshold_index 116 0.9873015873015873
+best_threshold_index 118 0.9865079365079366
+best_threshold_index 116 0.9876190476190476
+best_threshold_index 116 0.9868253968253968
+best_threshold_index 116 0.986984126984127
+best_threshold_index 116 0.9874603174603175
+[cfp_ff][146000]XNorm: 21.089279
+[cfp_ff][146000]Accuracy-Flip: 0.98657+-0.00384
+testing verification..
+(14000, 512)
+infer time 86.3164
+best_threshold_index 141 0.9250793650793651
+best_threshold_index 141 0.9253968253968254
+best_threshold_index 143 0.9241269841269841
+best_threshold_index 141 0.9255555555555556
+best_threshold_index 138 0.9276190476190476
+best_threshold_index 141 0.9244444444444444
+best_threshold_index 141 0.9234920634920635
+best_threshold_index 141 0.9247619047619048
+best_threshold_index 141 0.9249206349206349
+best_threshold_index 145 0.9258730158730158
+[cfp_fp][146000]XNorm: 21.848515
+[cfp_fp][146000]Accuracy-Flip: 0.92129+-0.01293
+testing verification..
+(12000, 512)
+infer time 73.95365399999996
+best_threshold_index 143 0.9037037037037037
+best_threshold_index 143 0.902962962962963
+best_threshold_index 143 0.9053703703703704
+best_threshold_index 141 0.9066666666666666
+best_threshold_index 141 0.909074074074074
+best_threshold_index 143 0.9033333333333333
+best_threshold_index 141 0.9044444444444445
+best_threshold_index 143 0.909074074074074
+best_threshold_index 143 0.9092592592592592
+best_threshold_index 143 0.9098148148148149
+[agedb_30][146000]XNorm: 22.848518
+[agedb_30][146000]Accuracy-Flip: 0.90500+-0.02358
+test accuracy is: 0.983
+epoch 9, total_step 146000, total loss is 13.97 , inference loss is 7.03, weight deacy loss is 6.94, training accuracy is 0.406250, time 54.243 samples/sec
+epoch 9, total_step 146020, total loss is 16.28 , inference loss is 9.34, weight deacy loss is 6.94, training accuracy is 0.468750, time 55.831 samples/sec
+epoch 9, total_step 146040, total loss is 13.82 , inference loss is 6.88, weight deacy loss is 6.94, training accuracy is 0.562500, time 56.701 samples/sec
+epoch 9, total_step 146060, total loss is 15.38 , inference loss is 8.44, weight deacy loss is 6.94, training accuracy is 0.531250, time 56.220 samples/sec
+epoch 9, total_step 146080, total loss is 13.16 , inference loss is 6.22, weight deacy loss is 6.94, training accuracy is 0.500000, time 56.405 samples/sec
+epoch 9, total_step 146100, total loss is 14.95 , inference loss is 8.01, weight deacy loss is 6.94, training accuracy is 0.250000, time 56.905 samples/sec
+epoch 9, total_step 146120, total loss is 18.57 , inference loss is 11.63, weight deacy loss is 6.94, training accuracy is 0.281250, time 57.019 samples/sec
+epoch 9, total_step 146140, total loss is 20.62 , inference loss is 13.69, weight deacy loss is 6.94, training accuracy is 0.281250, time 56.197 samples/sec
+epoch 9, total_step 146160, total loss is 15.61 , inference loss is 8.67, weight deacy loss is 6.93, training accuracy is 0.375000, time 56.673 samples/sec
+epoch 9, total_step 146180, total loss is 14.05 , inference loss is 7.12, weight deacy loss is 6.93, training accuracy is 0.406250, time 57.146 samples/sec
+epoch 9, total_step 146200, total loss is 13.15 , inference loss is 6.21, weight deacy loss is 6.93, training accuracy is 0.500000, time 56.977 samples/sec
+epoch 9, total_step 146220, total loss is 15.66 , inference loss is 8.73, weight deacy loss is 6.93, training accuracy is 0.312500, time 56.873 samples/sec
+epoch 9, total_step 146240, total loss is 11.17 , inference loss is 4.23, weight deacy loss is 6.93, training accuracy is 0.593750, time 57.652 samples/sec
+epoch 9, total_step 146260, total loss is 16.57 , inference loss is 9.64, weight deacy loss is 6.93, training accuracy is 0.375000, time 56.880 samples/sec
+epoch 9, total_step 146280, total loss is 12.99 , inference loss is 6.06, weight deacy loss is 6.93, training accuracy is 0.437500, time 56.814 samples/sec
+epoch 9, total_step 146300, total loss is 14.89 , inference loss is 7.96, weight deacy loss is 6.93, training accuracy is 0.343750, time 56.671 samples/sec
+epoch 9, total_step 146320, total loss is 12.69 , inference loss is 5.76, weight deacy loss is 6.93, training accuracy is 0.500000, time 56.628 samples/sec
+epoch 9, total_step 146340, total loss is 18.87 , inference loss is 11.94, weight deacy loss is 6.93, training accuracy is 0.343750, time 56.890 samples/sec
+epoch 9, total_step 146360, total loss is 12.12 , inference loss is 5.18, weight deacy loss is 6.93, training accuracy is 0.500000, time 56.615 samples/sec
+epoch 9, total_step 146380, total loss is 15.77 , inference loss is 8.84, weight deacy loss is 6.93, training accuracy is 0.437500, time 56.851 samples/sec
+epoch 9, total_step 146400, total loss is 17.48 , inference loss is 10.55, weight deacy loss is 6.93, training accuracy is 0.312500, time 56.416 samples/sec
+epoch 9, total_step 146420, total loss is 13.90 , inference loss is 6.97, weight deacy loss is 6.93, training accuracy is 0.593750, time 57.039 samples/sec
+epoch 9, total_step 146440, total loss is 14.76 , inference loss is 7.83, weight deacy loss is 6.93, training accuracy is 0.343750, time 55.536 samples/sec
+epoch 9, total_step 146460, total loss is 16.00 , inference loss is 9.07, weight deacy loss is 6.93, training accuracy is 0.343750, time 56.412 samples/sec
+epoch 9, total_step 146480, total loss is 16.05 , inference loss is 9.12, weight deacy loss is 6.93, training accuracy is 0.406250, time 56.797 samples/sec
+epoch 9, total_step 146500, total loss is 14.43 , inference loss is 7.50, weight deacy loss is 6.93, training accuracy is 0.375000, time 56.798 samples/sec
+epoch 9, total_step 146520, total loss is 15.57 , inference loss is 8.64, weight deacy loss is 6.93, training accuracy is 0.343750, time 55.836 samples/sec
+epoch 9, total_step 146540, total loss is 16.44 , inference loss is 9.51, weight deacy loss is 6.93, training accuracy is 0.375000, time 56.991 samples/sec
+epoch 9, total_step 146560, total loss is 17.68 , inference loss is 10.75, weight deacy loss is 6.93, training accuracy is 0.343750, time 56.343 samples/sec
+epoch 9, total_step 146580, total loss is 14.65 , inference loss is 7.72, weight deacy loss is 6.93, training accuracy is 0.281250, time 56.719 samples/sec
+epoch 9, total_step 146600, total loss is 14.08 , inference loss is 7.15, weight deacy loss is 6.93, training accuracy is 0.468750, time 56.179 samples/sec
+epoch 9, total_step 146620, total loss is 18.53 , inference loss is 11.60, weight deacy loss is 6.93, training accuracy is 0.437500, time 56.532 samples/sec
+epoch 9, total_step 146640, total loss is 16.83 , inference loss is 9.90, weight deacy loss is 6.93, training accuracy is 0.312500, time 55.706 samples/sec
+epoch 9, total_step 146660, total loss is 18.82 , inference loss is 11.89, weight deacy loss is 6.93, training accuracy is 0.312500, time 55.792 samples/sec
+epoch 9, total_step 146680, total loss is 16.00 , inference loss is 9.08, weight deacy loss is 6.93, training accuracy is 0.343750, time 56.640 samples/sec
+epoch 9, total_step 146700, total loss is 15.45 , inference loss is 8.52, weight deacy loss is 6.93, training accuracy is 0.312500, time 56.401 samples/sec
+epoch 9, total_step 146720, total loss is 15.32 , inference loss is 8.39, weight deacy loss is 6.93, training accuracy is 0.468750, time 56.680 samples/sec
+epoch 9, total_step 146740, total loss is 12.58 , inference loss is 5.65, weight deacy loss is 6.93, training accuracy is 0.562500, time 55.996 samples/sec
+epoch 9, total_step 146760, total loss is 14.63 , inference loss is 7.70, weight deacy loss is 6.93, training accuracy is 0.343750, time 56.636 samples/sec
+epoch 9, total_step 146780, total loss is 18.36 , inference loss is 11.43, weight deacy loss is 6.93, training accuracy is 0.250000, time 56.369 samples/sec
+epoch 9, total_step 146800, total loss is 15.55 , inference loss is 8.63, weight deacy loss is 6.92, training accuracy is 0.343750, time 56.823 samples/sec
+epoch 9, total_step 146820, total loss is 11.91 , inference loss is 4.98, weight deacy loss is 6.92, training accuracy is 0.531250, time 56.351 samples/sec
+epoch 9, total_step 146840, total loss is 15.38 , inference loss is 8.46, weight deacy loss is 6.92, training accuracy is 0.375000, time 56.481 samples/sec
+epoch 9, total_step 146860, total loss is 16.98 , inference loss is 10.06, weight deacy loss is 6.92, training accuracy is 0.218750, time 56.181 samples/sec
+epoch 9, total_step 146880, total loss is 14.44 , inference loss is 7.52, weight deacy loss is 6.92, training accuracy is 0.500000, time 55.918 samples/sec
+epoch 9, total_step 146900, total loss is 13.88 , inference loss is 6.96, weight deacy loss is 6.92, training accuracy is 0.437500, time 56.651 samples/sec
+epoch 9, total_step 146920, total loss is 14.17 , inference loss is 7.24, weight deacy loss is 6.92, training accuracy is 0.468750, time 55.879 samples/sec
+epoch 9, total_step 146940, total loss is 13.41 , inference loss is 6.49, weight deacy loss is 6.92, training accuracy is 0.625000, time 56.080 samples/sec
+epoch 9, total_step 146960, total loss is 14.87 , inference loss is 7.95, weight deacy loss is 6.92, training accuracy is 0.437500, time 56.472 samples/sec
+epoch 9, total_step 146980, total loss is 14.37 , inference loss is 7.45, weight deacy loss is 6.92, training accuracy is 0.500000, time 56.499 samples/sec
+epoch 9, total_step 147000, total loss is 17.94 , inference loss is 11.01, weight deacy loss is 6.92, training accuracy is 0.406250, time 56.196 samples/sec
+epoch 9, total_step 147020, total loss is 13.54 , inference loss is 6.62, weight deacy loss is 6.92, training accuracy is 0.500000, time 56.331 samples/sec
+epoch 9, total_step 147040, total loss is 14.85 , inference loss is 7.93, weight deacy loss is 6.92, training accuracy is 0.437500, time 56.296 samples/sec
+epoch 9, total_step 147060, total loss is 14.78 , inference loss is 7.85, weight deacy loss is 6.92, training accuracy is 0.375000, time 56.390 samples/sec
+epoch 9, total_step 147080, total loss is 15.57 , inference loss is 8.65, weight deacy loss is 6.92, training accuracy is 0.406250, time 56.743 samples/sec
+epoch 9, total_step 147100, total loss is 12.86 , inference loss is 5.94, weight deacy loss is 6.92, training accuracy is 0.343750, time 56.345 samples/sec
+epoch 9, total_step 147120, total loss is 17.13 , inference loss is 10.21, weight deacy loss is 6.92, training accuracy is 0.375000, time 56.640 samples/sec
+epoch 9, total_step 147140, total loss is 14.80 , inference loss is 7.88, weight deacy loss is 6.92, training accuracy is 0.406250, time 55.655 samples/sec
+epoch 9, total_step 147160, total loss is 15.11 , inference loss is 8.19, weight deacy loss is 6.92, training accuracy is 0.406250, time 56.140 samples/sec
+epoch 9, total_step 147180, total loss is 14.52 , inference loss is 7.60, weight deacy loss is 6.92, training accuracy is 0.406250, time 56.316 samples/sec
+epoch 9, total_step 147200, total loss is 13.13 , inference loss is 6.21, weight deacy loss is 6.92, training accuracy is 0.437500, time 56.124 samples/sec
+epoch 9, total_step 147220, total loss is 17.06 , inference loss is 10.14, weight deacy loss is 6.92, training accuracy is 0.468750, time 56.526 samples/sec
+epoch 9, total_step 147240, total loss is 12.75 , inference loss is 5.83, weight deacy loss is 6.92, training accuracy is 0.500000, time 56.946 samples/sec
+epoch 9, total_step 147260, total loss is 12.25 , inference loss is 5.33, weight deacy loss is 6.92, training accuracy is 0.406250, time 56.596 samples/sec
+epoch 9, total_step 147280, total loss is 20.20 , inference loss is 13.29, weight deacy loss is 6.92, training accuracy is 0.312500, time 56.281 samples/sec
+epoch 9, total_step 147300, total loss is 16.27 , inference loss is 9.36, weight deacy loss is 6.92, training accuracy is 0.406250, time 56.256 samples/sec
+epoch 9, total_step 147320, total loss is 12.90 , inference loss is 5.98, weight deacy loss is 6.92, training accuracy is 0.500000, time 55.506 samples/sec
+epoch 9, total_step 147340, total loss is 17.26 , inference loss is 10.34, weight deacy loss is 6.92, training accuracy is 0.281250, time 56.633 samples/sec
+epoch 9, total_step 147360, total loss is 17.39 , inference loss is 10.47, weight deacy loss is 6.92, training accuracy is 0.312500, time 56.681 samples/sec
+epoch 9, total_step 147380, total loss is 14.74 , inference loss is 7.82, weight deacy loss is 6.92, training accuracy is 0.312500, time 55.866 samples/sec
+epoch 9, total_step 147400, total loss is 18.76 , inference loss is 11.85, weight deacy loss is 6.92, training accuracy is 0.250000, time 57.329 samples/sec
+epoch 9, total_step 147420, total loss is 15.48 , inference loss is 8.56, weight deacy loss is 6.92, training accuracy is 0.468750, time 56.600 samples/sec
+epoch 9, total_step 147440, total loss is 16.59 , inference loss is 9.67, weight deacy loss is 6.92, training accuracy is 0.343750, time 56.192 samples/sec
+epoch 9, total_step 147460, total loss is 11.77 , inference loss is 4.86, weight deacy loss is 6.91, training accuracy is 0.593750, time 57.223 samples/sec
+epoch 9, total_step 147480, total loss is 12.40 , inference loss is 5.48, weight deacy loss is 6.91, training accuracy is 0.531250, time 56.321 samples/sec
+epoch 9, total_step 147500, total loss is 16.68 , inference loss is 9.76, weight deacy loss is 6.91, training accuracy is 0.406250, time 57.089 samples/sec
+epoch 9, total_step 147520, total loss is 16.48 , inference loss is 9.56, weight deacy loss is 6.91, training accuracy is 0.281250, time 54.902 samples/sec
+epoch 9, total_step 147540, total loss is 15.00 , inference loss is 8.08, weight deacy loss is 6.91, training accuracy is 0.468750, time 55.986 samples/sec
+epoch 9, total_step 147560, total loss is 17.74 , inference loss is 10.82, weight deacy loss is 6.91, training accuracy is 0.343750, time 56.555 samples/sec
+epoch 9, total_step 147580, total loss is 16.53 , inference loss is 9.61, weight deacy loss is 6.91, training accuracy is 0.281250, time 55.749 samples/sec
+epoch 9, total_step 147600, total loss is 14.76 , inference loss is 7.84, weight deacy loss is 6.91, training accuracy is 0.250000, time 55.760 samples/sec
+epoch 9, total_step 147620, total loss is 15.94 , inference loss is 9.02, weight deacy loss is 6.91, training accuracy is 0.343750, time 56.819 samples/sec
+epoch 9, total_step 147640, total loss is 19.33 , inference loss is 12.41, weight deacy loss is 6.91, training accuracy is 0.343750, time 55.884 samples/sec
+epoch 9, total_step 147660, total loss is 14.67 , inference loss is 7.76, weight deacy loss is 6.91, training accuracy is 0.375000, time 56.559 samples/sec
+epoch 9, total_step 147680, total loss is 14.26 , inference loss is 7.35, weight deacy loss is 6.91, training accuracy is 0.406250, time 57.232 samples/sec
+epoch 9, total_step 147700, total loss is 15.83 , inference loss is 8.92, weight deacy loss is 6.91, training accuracy is 0.375000, time 56.455 samples/sec
+epoch 9, total_step 147720, total loss is 15.49 , inference loss is 8.58, weight deacy loss is 6.91, training accuracy is 0.500000, time 57.027 samples/sec
+epoch 9, total_step 147740, total loss is 15.98 , inference loss is 9.07, weight deacy loss is 6.91, training accuracy is 0.437500, time 55.903 samples/sec
+epoch 9, total_step 147760, total loss is 14.31 , inference loss is 7.40, weight deacy loss is 6.91, training accuracy is 0.343750, time 55.556 samples/sec
+epoch 9, total_step 147780, total loss is 16.79 , inference loss is 9.88, weight deacy loss is 6.91, training accuracy is 0.437500, time 55.608 samples/sec
+epoch 9, total_step 147800, total loss is 12.20 , inference loss is 5.29, weight deacy loss is 6.91, training accuracy is 0.375000, time 55.931 samples/sec
+epoch 9, total_step 147820, total loss is 13.85 , inference loss is 6.94, weight deacy loss is 6.91, training accuracy is 0.468750, time 56.626 samples/sec
+epoch 9, total_step 147840, total loss is 16.78 , inference loss is 9.87, weight deacy loss is 6.91, training accuracy is 0.312500, time 55.799 samples/sec
+epoch 9, total_step 147860, total loss is 18.71 , inference loss is 11.80, weight deacy loss is 6.91, training accuracy is 0.375000, time 56.576 samples/sec
+epoch 9, total_step 147880, total loss is 17.83 , inference loss is 10.92, weight deacy loss is 6.91, training accuracy is 0.500000, time 56.925 samples/sec
+epoch 9, total_step 147900, total loss is 14.33 , inference loss is 7.42, weight deacy loss is 6.91, training accuracy is 0.468750, time 55.818 samples/sec
+epoch 9, total_step 147920, total loss is 14.03 , inference loss is 7.12, weight deacy loss is 6.91, training accuracy is 0.406250, time 56.744 samples/sec
+epoch 9, total_step 147940, total loss is 13.63 , inference loss is 6.73, weight deacy loss is 6.91, training accuracy is 0.312500, time 56.578 samples/sec
+epoch 9, total_step 147960, total loss is 16.43 , inference loss is 9.52, weight deacy loss is 6.91, training accuracy is 0.343750, time 56.911 samples/sec
+epoch 9, total_step 147980, total loss is 13.11 , inference loss is 6.20, weight deacy loss is 6.91, training accuracy is 0.562500, time 57.862 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.43057299999995
+best_threshold_index 109 0.9857407407407407
+best_threshold_index 109 0.9848148148148148
+best_threshold_index 108 0.9846296296296296
+best_threshold_index 109 0.9855555555555555
+best_threshold_index 109 0.9848148148148148
+best_threshold_index 113 0.9842592592592593
+best_threshold_index 109 0.9848148148148148
+best_threshold_index 107 0.9842592592592593
+best_threshold_index 109 0.9831481481481481
+best_threshold_index 113 0.9838888888888889
+[lfw][148000]XNorm: 23.143990
+[lfw][148000]Accuracy-Flip: 0.98283+-0.00619
+testing verification..
+(14000, 512)
+infer time 85.43691299999995
+best_threshold_index 121 0.9858730158730159
+best_threshold_index 113 0.9846031746031746
+best_threshold_index 120 0.9846031746031746
+best_threshold_index 112 0.9846031746031746
+best_threshold_index 118 0.9861904761904762
+best_threshold_index 113 0.9850793650793651
+best_threshold_index 120 0.9857142857142858
+best_threshold_index 120 0.9847619047619047
+best_threshold_index 120 0.984920634920635
+best_threshold_index 118 0.9853968253968254
+[cfp_ff][148000]XNorm: 22.211031
+[cfp_ff][148000]Accuracy-Flip: 0.98286+-0.00461
+testing verification..
+(14000, 512)
+infer time 85.59898200000013
+best_threshold_index 135 0.9231746031746032
+best_threshold_index 135 0.9238095238095239
+best_threshold_index 135 0.9217460317460318
+best_threshold_index 135 0.9271428571428572
+best_threshold_index 135 0.9280952380952381
+best_threshold_index 135 0.922063492063492
+best_threshold_index 135 0.9211111111111111
+best_threshold_index 135 0.9230158730158731
+best_threshold_index 135 0.9236507936507936
+best_threshold_index 135 0.9247619047619048
+[cfp_fp][148000]XNorm: 22.156060
+[cfp_fp][148000]Accuracy-Flip: 0.92386+-0.01930
+testing verification..
+(12000, 512)
+infer time 73.67102399999997
+best_threshold_index 131 0.9059259259259259
+best_threshold_index 130 0.9037037037037037
+best_threshold_index 130 0.9035185185185185
+best_threshold_index 130 0.9085185185185185
+best_threshold_index 130 0.9103703703703704
+best_threshold_index 131 0.9042592592592592
+best_threshold_index 130 0.9061111111111111
+best_threshold_index 131 0.9127777777777778
+best_threshold_index 130 0.9098148148148149
+best_threshold_index 130 0.9077777777777778
+[agedb_30][148000]XNorm: 23.747774
+[agedb_30][148000]Accuracy-Flip: 0.90517+-0.02675
+test accuracy is: 0.9828333333333333
+epoch 9, total_step 148000, total loss is 11.87 , inference loss is 4.97, weight deacy loss is 6.91, training accuracy is 0.500000, time 52.459 samples/sec
+epoch 9, total_step 148020, total loss is 13.52 , inference loss is 6.62, weight deacy loss is 6.91, training accuracy is 0.437500, time 56.520 samples/sec
+epoch 9, total_step 148040, total loss is 17.40 , inference loss is 10.50, weight deacy loss is 6.91, training accuracy is 0.343750, time 57.281 samples/sec
+epoch 9, total_step 148060, total loss is 14.44 , inference loss is 7.53, weight deacy loss is 6.90, training accuracy is 0.437500, time 56.534 samples/sec
+epoch 9, total_step 148080, total loss is 13.30 , inference loss is 6.39, weight deacy loss is 6.90, training accuracy is 0.468750, time 56.907 samples/sec
+epoch 9, total_step 148100, total loss is 10.72 , inference loss is 3.82, weight deacy loss is 6.90, training accuracy is 0.500000, time 56.768 samples/sec
+epoch 9, total_step 148120, total loss is 13.44 , inference loss is 6.54, weight deacy loss is 6.90, training accuracy is 0.406250, time 56.954 samples/sec
+epoch 9, total_step 148140, total loss is 16.07 , inference loss is 9.16, weight deacy loss is 6.90, training accuracy is 0.468750, time 56.636 samples/sec
+epoch 9, total_step 148160, total loss is 15.26 , inference loss is 8.36, weight deacy loss is 6.90, training accuracy is 0.406250, time 56.340 samples/sec
+epoch 9, total_step 148180, total loss is 15.58 , inference loss is 8.68, weight deacy loss is 6.90, training accuracy is 0.343750, time 56.404 samples/sec
+epoch 9, total_step 148200, total loss is 16.12 , inference loss is 9.22, weight deacy loss is 6.90, training accuracy is 0.375000, time 56.975 samples/sec
+epoch 9, total_step 148220, total loss is 16.08 , inference loss is 9.17, weight deacy loss is 6.90, training accuracy is 0.437500, time 57.054 samples/sec
+epoch 9, total_step 148240, total loss is 20.34 , inference loss is 13.44, weight deacy loss is 6.90, training accuracy is 0.281250, time 57.073 samples/sec
+epoch 9, total_step 148260, total loss is 13.42 , inference loss is 6.52, weight deacy loss is 6.90, training accuracy is 0.312500, time 56.717 samples/sec
+epoch 9, total_step 148280, total loss is 17.45 , inference loss is 10.54, weight deacy loss is 6.90, training accuracy is 0.312500, time 55.928 samples/sec
+epoch 9, total_step 148300, total loss is 12.03 , inference loss is 5.13, weight deacy loss is 6.90, training accuracy is 0.437500, time 56.998 samples/sec
+epoch 9, total_step 148320, total loss is 13.44 , inference loss is 6.54, weight deacy loss is 6.90, training accuracy is 0.406250, time 56.309 samples/sec
+epoch 9, total_step 148340, total loss is 15.66 , inference loss is 8.76, weight deacy loss is 6.90, training accuracy is 0.468750, time 56.823 samples/sec
+epoch 9, total_step 148360, total loss is 11.40 , inference loss is 4.50, weight deacy loss is 6.90, training accuracy is 0.593750, time 57.610 samples/sec
+epoch 9, total_step 148380, total loss is 13.31 , inference loss is 6.41, weight deacy loss is 6.90, training accuracy is 0.406250, time 56.131 samples/sec
+epoch 9, total_step 148400, total loss is 13.91 , inference loss is 7.01, weight deacy loss is 6.90, training accuracy is 0.500000, time 57.101 samples/sec
+epoch 9, total_step 148420, total loss is 13.49 , inference loss is 6.59, weight deacy loss is 6.90, training accuracy is 0.500000, time 56.361 samples/sec
+epoch 9, total_step 148440, total loss is 11.76 , inference loss is 4.86, weight deacy loss is 6.90, training accuracy is 0.593750, time 56.445 samples/sec
+epoch 9, total_step 148460, total loss is 13.18 , inference loss is 6.29, weight deacy loss is 6.90, training accuracy is 0.375000, time 55.956 samples/sec
+epoch 9, total_step 148480, total loss is 21.23 , inference loss is 14.33, weight deacy loss is 6.90, training accuracy is 0.312500, time 56.027 samples/sec
+epoch 9, total_step 148500, total loss is 13.78 , inference loss is 6.88, weight deacy loss is 6.90, training accuracy is 0.500000, time 56.239 samples/sec
+epoch 9, total_step 148520, total loss is 10.03 , inference loss is 3.13, weight deacy loss is 6.90, training accuracy is 0.531250, time 56.044 samples/sec
+epoch 9, total_step 148540, total loss is 12.96 , inference loss is 6.06, weight deacy loss is 6.90, training accuracy is 0.437500, time 57.165 samples/sec
+epoch 9, total_step 148560, total loss is 11.60 , inference loss is 4.71, weight deacy loss is 6.90, training accuracy is 0.500000, time 56.399 samples/sec
+epoch 9, total_step 148580, total loss is 16.79 , inference loss is 9.89, weight deacy loss is 6.90, training accuracy is 0.406250, time 56.684 samples/sec
+epoch 9, total_step 148600, total loss is 16.37 , inference loss is 9.47, weight deacy loss is 6.90, training accuracy is 0.312500, time 55.990 samples/sec
+epoch 9, total_step 148620, total loss is 16.35 , inference loss is 9.45, weight deacy loss is 6.90, training accuracy is 0.406250, time 56.715 samples/sec
+epoch 9, total_step 148640, total loss is 17.92 , inference loss is 11.03, weight deacy loss is 6.89, training accuracy is 0.312500, time 55.201 samples/sec
+epoch 9, total_step 148660, total loss is 17.96 , inference loss is 11.07, weight deacy loss is 6.89, training accuracy is 0.343750, time 56.292 samples/sec
+epoch 9, total_step 148680, total loss is 15.38 , inference loss is 8.49, weight deacy loss is 6.89, training accuracy is 0.500000, time 56.778 samples/sec
+epoch 9, total_step 148700, total loss is 18.71 , inference loss is 11.81, weight deacy loss is 6.89, training accuracy is 0.187500, time 56.563 samples/sec
+epoch 9, total_step 148720, total loss is 13.95 , inference loss is 7.06, weight deacy loss is 6.89, training accuracy is 0.500000, time 56.727 samples/sec
+epoch 9, total_step 148740, total loss is 13.99 , inference loss is 7.10, weight deacy loss is 6.89, training accuracy is 0.500000, time 56.340 samples/sec
+epoch 9, total_step 148760, total loss is 14.74 , inference loss is 7.85, weight deacy loss is 6.89, training accuracy is 0.500000, time 56.200 samples/sec
+epoch 9, total_step 148780, total loss is 14.79 , inference loss is 7.90, weight deacy loss is 6.89, training accuracy is 0.468750, time 57.105 samples/sec
+epoch 9, total_step 148800, total loss is 13.52 , inference loss is 6.62, weight deacy loss is 6.89, training accuracy is 0.437500, time 56.702 samples/sec
+epoch 9, total_step 148820, total loss is 16.45 , inference loss is 9.56, weight deacy loss is 6.89, training accuracy is 0.312500, time 55.819 samples/sec
+epoch 9, total_step 148840, total loss is 16.70 , inference loss is 9.81, weight deacy loss is 6.89, training accuracy is 0.250000, time 56.487 samples/sec
+epoch 9, total_step 148860, total loss is 17.60 , inference loss is 10.71, weight deacy loss is 6.89, training accuracy is 0.406250, time 56.809 samples/sec
+epoch 9, total_step 148880, total loss is 14.06 , inference loss is 7.17, weight deacy loss is 6.89, training accuracy is 0.406250, time 57.100 samples/sec
+epoch 9, total_step 148900, total loss is 14.11 , inference loss is 7.22, weight deacy loss is 6.89, training accuracy is 0.468750, time 56.593 samples/sec
+epoch 9, total_step 148920, total loss is 16.02 , inference loss is 9.13, weight deacy loss is 6.89, training accuracy is 0.281250, time 57.384 samples/sec
+epoch 9, total_step 148940, total loss is 16.38 , inference loss is 9.49, weight deacy loss is 6.89, training accuracy is 0.343750, time 55.827 samples/sec
+epoch 9, total_step 148960, total loss is 16.32 , inference loss is 9.43, weight deacy loss is 6.89, training accuracy is 0.468750, time 56.986 samples/sec
+epoch 9, total_step 148980, total loss is 14.75 , inference loss is 7.86, weight deacy loss is 6.89, training accuracy is 0.375000, time 56.006 samples/sec
+epoch 9, total_step 149000, total loss is 15.79 , inference loss is 8.90, weight deacy loss is 6.89, training accuracy is 0.343750, time 55.916 samples/sec
+epoch 9, total_step 149020, total loss is 14.00 , inference loss is 7.11, weight deacy loss is 6.89, training accuracy is 0.468750, time 56.273 samples/sec
+epoch 9, total_step 149040, total loss is 15.59 , inference loss is 8.71, weight deacy loss is 6.89, training accuracy is 0.375000, time 57.874 samples/sec
+epoch 9, total_step 149060, total loss is 13.87 , inference loss is 6.99, weight deacy loss is 6.89, training accuracy is 0.343750, time 56.847 samples/sec
+epoch 9, total_step 149080, total loss is 13.59 , inference loss is 6.70, weight deacy loss is 6.89, training accuracy is 0.468750, time 56.223 samples/sec
+epoch 9, total_step 149100, total loss is 15.31 , inference loss is 8.42, weight deacy loss is 6.89, training accuracy is 0.343750, time 56.581 samples/sec
+epoch 9, total_step 149120, total loss is 19.96 , inference loss is 13.07, weight deacy loss is 6.89, training accuracy is 0.250000, time 56.844 samples/sec
+epoch 9, total_step 149140, total loss is 15.12 , inference loss is 8.23, weight deacy loss is 6.89, training accuracy is 0.375000, time 57.130 samples/sec
+epoch 9, total_step 149160, total loss is 15.65 , inference loss is 8.77, weight deacy loss is 6.89, training accuracy is 0.468750, time 57.156 samples/sec
+epoch 9, total_step 149180, total loss is 15.45 , inference loss is 8.56, weight deacy loss is 6.89, training accuracy is 0.437500, time 56.185 samples/sec
+epoch 9, total_step 149200, total loss is 14.97 , inference loss is 8.09, weight deacy loss is 6.89, training accuracy is 0.375000, time 56.649 samples/sec
+epoch 9, total_step 149220, total loss is 17.34 , inference loss is 10.46, weight deacy loss is 6.89, training accuracy is 0.312500, time 57.080 samples/sec
+epoch 9, total_step 149240, total loss is 16.56 , inference loss is 9.67, weight deacy loss is 6.89, training accuracy is 0.312500, time 56.679 samples/sec
+epoch 9, total_step 149260, total loss is 12.50 , inference loss is 5.62, weight deacy loss is 6.89, training accuracy is 0.562500, time 56.830 samples/sec
+epoch 9, total_step 149280, total loss is 15.27 , inference loss is 8.38, weight deacy loss is 6.89, training accuracy is 0.312500, time 56.364 samples/sec
+epoch 9, total_step 149300, total loss is 14.74 , inference loss is 7.86, weight deacy loss is 6.88, training accuracy is 0.406250, time 56.632 samples/sec
+epoch 9, total_step 149320, total loss is 11.96 , inference loss is 5.07, weight deacy loss is 6.88, training accuracy is 0.500000, time 56.833 samples/sec
+epoch 9, total_step 149340, total loss is 12.74 , inference loss is 5.85, weight deacy loss is 6.88, training accuracy is 0.468750, time 56.706 samples/sec
+epoch 9, total_step 149360, total loss is 15.74 , inference loss is 8.85, weight deacy loss is 6.88, training accuracy is 0.343750, time 56.951 samples/sec
+epoch 9, total_step 149380, total loss is 17.12 , inference loss is 10.24, weight deacy loss is 6.88, training accuracy is 0.218750, time 56.290 samples/sec
+epoch 9, total_step 149400, total loss is 16.49 , inference loss is 9.61, weight deacy loss is 6.88, training accuracy is 0.250000, time 56.778 samples/sec
+epoch 9, total_step 149420, total loss is 17.11 , inference loss is 10.22, weight deacy loss is 6.88, training accuracy is 0.281250, time 56.430 samples/sec
+epoch 9, total_step 149440, total loss is 18.51 , inference loss is 11.62, weight deacy loss is 6.88, training accuracy is 0.312500, time 56.140 samples/sec
+epoch 9, total_step 149460, total loss is 15.26 , inference loss is 8.38, weight deacy loss is 6.88, training accuracy is 0.312500, time 57.278 samples/sec
+epoch 9, total_step 149480, total loss is 13.53 , inference loss is 6.64, weight deacy loss is 6.88, training accuracy is 0.593750, time 56.892 samples/sec
+epoch 9, total_step 149500, total loss is 18.37 , inference loss is 11.49, weight deacy loss is 6.88, training accuracy is 0.375000, time 56.667 samples/sec
+epoch 9, total_step 149520, total loss is 14.22 , inference loss is 7.34, weight deacy loss is 6.88, training accuracy is 0.375000, time 55.808 samples/sec
+epoch 9, total_step 149540, total loss is 16.46 , inference loss is 9.58, weight deacy loss is 6.88, training accuracy is 0.312500, time 56.292 samples/sec
+epoch 9, total_step 149560, total loss is 13.74 , inference loss is 6.86, weight deacy loss is 6.88, training accuracy is 0.343750, time 56.171 samples/sec
+epoch 9, total_step 149580, total loss is 14.49 , inference loss is 7.61, weight deacy loss is 6.88, training accuracy is 0.500000, time 55.914 samples/sec
+epoch 9, total_step 149600, total loss is 12.11 , inference loss is 5.23, weight deacy loss is 6.88, training accuracy is 0.406250, time 57.113 samples/sec
+epoch 9, total_step 149620, total loss is 13.50 , inference loss is 6.62, weight deacy loss is 6.88, training accuracy is 0.375000, time 57.707 samples/sec
+epoch 9, total_step 149640, total loss is 19.15 , inference loss is 12.28, weight deacy loss is 6.88, training accuracy is 0.281250, time 56.223 samples/sec
+epoch 9, total_step 149660, total loss is 11.44 , inference loss is 4.56, weight deacy loss is 6.88, training accuracy is 0.437500, time 56.712 samples/sec
+epoch 9, total_step 149680, total loss is 16.47 , inference loss is 9.59, weight deacy loss is 6.88, training accuracy is 0.250000, time 57.140 samples/sec
+epoch 9, total_step 149700, total loss is 16.59 , inference loss is 9.71, weight deacy loss is 6.88, training accuracy is 0.406250, time 56.048 samples/sec
+epoch 9, total_step 149720, total loss is 15.18 , inference loss is 8.30, weight deacy loss is 6.88, training accuracy is 0.343750, time 56.502 samples/sec
+epoch 9, total_step 149740, total loss is 16.61 , inference loss is 9.73, weight deacy loss is 6.88, training accuracy is 0.500000, time 56.951 samples/sec
+epoch 9, total_step 149760, total loss is 15.49 , inference loss is 8.62, weight deacy loss is 6.88, training accuracy is 0.312500, time 57.109 samples/sec
+epoch 9, total_step 149780, total loss is 17.40 , inference loss is 10.52, weight deacy loss is 6.88, training accuracy is 0.406250, time 56.787 samples/sec
+epoch 9, total_step 149800, total loss is 18.47 , inference loss is 11.59, weight deacy loss is 6.88, training accuracy is 0.250000, time 56.372 samples/sec
+epoch 9, total_step 149820, total loss is 17.96 , inference loss is 11.08, weight deacy loss is 6.88, training accuracy is 0.156250, time 56.907 samples/sec
+epoch 9, total_step 149840, total loss is 13.47 , inference loss is 6.60, weight deacy loss is 6.88, training accuracy is 0.375000, time 56.971 samples/sec
+epoch 9, total_step 149860, total loss is 15.93 , inference loss is 9.06, weight deacy loss is 6.88, training accuracy is 0.375000, time 55.105 samples/sec
+epoch 9, total_step 149880, total loss is 16.74 , inference loss is 9.87, weight deacy loss is 6.88, training accuracy is 0.187500, time 56.246 samples/sec
+epoch 9, total_step 149900, total loss is 17.98 , inference loss is 11.11, weight deacy loss is 6.87, training accuracy is 0.250000, time 56.857 samples/sec
+epoch 9, total_step 149920, total loss is 18.50 , inference loss is 11.62, weight deacy loss is 6.87, training accuracy is 0.156250, time 56.247 samples/sec
+epoch 9, total_step 149940, total loss is 16.57 , inference loss is 9.70, weight deacy loss is 6.87, training accuracy is 0.437500, time 56.279 samples/sec
+epoch 9, total_step 149960, total loss is 16.88 , inference loss is 10.01, weight deacy loss is 6.87, training accuracy is 0.343750, time 56.568 samples/sec
+epoch 9, total_step 149980, total loss is 14.22 , inference loss is 7.35, weight deacy loss is 6.87, training accuracy is 0.343750, time 56.423 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.84212599999994
+best_threshold_index 108 0.9857407407407407
+best_threshold_index 108 0.985
+best_threshold_index 108 0.9859259259259259
+best_threshold_index 108 0.9862962962962963
+best_threshold_index 107 0.9855555555555555
+best_threshold_index 109 0.9846296296296296
+best_threshold_index 109 0.985
+best_threshold_index 107 0.9846296296296296
+best_threshold_index 108 0.9833333333333333
+best_threshold_index 108 0.9842592592592593
+[lfw][150000]XNorm: 21.382974
+[lfw][150000]Accuracy-Flip: 0.98433+-0.00757
+testing verification..
+(14000, 512)
+infer time 86.13307799999994
+best_threshold_index 118 0.9873015873015873
+best_threshold_index 118 0.9871428571428571
+best_threshold_index 116 0.9863492063492063
+best_threshold_index 118 0.9865079365079366
+best_threshold_index 116 0.9874603174603175
+best_threshold_index 116 0.9858730158730159
+best_threshold_index 119 0.986984126984127
+best_threshold_index 116 0.9868253968253968
+best_threshold_index 119 0.986984126984127
+best_threshold_index 118 0.9874603174603175
+[cfp_ff][150000]XNorm: 20.965576
+[cfp_ff][150000]Accuracy-Flip: 0.98600+-0.00413
+testing verification..
+(14000, 512)
+infer time 86.36782399999996
+best_threshold_index 135 0.9192063492063492
+best_threshold_index 135 0.9196825396825397
+best_threshold_index 135 0.9179365079365079
+best_threshold_index 135 0.920952380952381
+best_threshold_index 135 0.9234920634920635
+best_threshold_index 135 0.9184126984126985
+best_threshold_index 135 0.9182539682539682
+best_threshold_index 135 0.9193650793650794
+best_threshold_index 135 0.9198412698412698
+best_threshold_index 138 0.9207936507936508
+[cfp_fp][150000]XNorm: 21.355078
+[cfp_fp][150000]Accuracy-Flip: 0.91843+-0.01449
+testing verification..
+(12000, 512)
+infer time 73.34858399999992
+best_threshold_index 139 0.9044444444444445
+best_threshold_index 139 0.9044444444444445
+best_threshold_index 139 0.905
+best_threshold_index 139 0.905
+best_threshold_index 139 0.9074074074074074
+best_threshold_index 139 0.902962962962963
+best_threshold_index 139 0.9070370370370371
+best_threshold_index 139 0.9096296296296297
+best_threshold_index 139 0.9068518518518518
+best_threshold_index 139 0.9072222222222223
+[agedb_30][150000]XNorm: 21.567974
+[agedb_30][150000]Accuracy-Flip: 0.90600+-0.01672
+test accuracy is: 0.9843333333333334
+epoch 9, total_step 150000, total loss is 20.00 , inference loss is 13.12, weight deacy loss is 6.87, training accuracy is 0.343750, time 54.033 samples/sec
+epoch 9, total_step 150020, total loss is 18.92 , inference loss is 12.05, weight deacy loss is 6.87, training accuracy is 0.343750, time 56.961 samples/sec
+epoch 9, total_step 150040, total loss is 11.91 , inference loss is 5.04, weight deacy loss is 6.87, training accuracy is 0.437500, time 56.785 samples/sec
+epoch 9, total_step 150060, total loss is 17.40 , inference loss is 10.53, weight deacy loss is 6.87, training accuracy is 0.281250, time 56.686 samples/sec
+epoch 9, total_step 150080, total loss is 17.62 , inference loss is 10.74, weight deacy loss is 6.87, training accuracy is 0.343750, time 55.331 samples/sec
+epoch 9, total_step 150100, total loss is 17.67 , inference loss is 10.80, weight deacy loss is 6.87, training accuracy is 0.312500, time 56.216 samples/sec
+epoch 9, total_step 150120, total loss is 15.82 , inference loss is 8.94, weight deacy loss is 6.87, training accuracy is 0.250000, time 57.468 samples/sec
+epoch 9, total_step 150140, total loss is 18.64 , inference loss is 11.77, weight deacy loss is 6.87, training accuracy is 0.343750, time 56.947 samples/sec
+epoch 9, total_step 150160, total loss is 15.65 , inference loss is 8.78, weight deacy loss is 6.87, training accuracy is 0.312500, time 57.082 samples/sec
+epoch 9, total_step 150180, total loss is 16.13 , inference loss is 9.26, weight deacy loss is 6.87, training accuracy is 0.500000, time 56.552 samples/sec
+epoch 9, total_step 150200, total loss is 17.66 , inference loss is 10.79, weight deacy loss is 6.87, training accuracy is 0.250000, time 56.170 samples/sec
+epoch 9, total_step 150220, total loss is 14.13 , inference loss is 7.26, weight deacy loss is 6.87, training accuracy is 0.281250, time 56.190 samples/sec
+epoch 9, total_step 150240, total loss is 14.47 , inference loss is 7.60, weight deacy loss is 6.87, training accuracy is 0.281250, time 55.876 samples/sec
+epoch 9, total_step 150260, total loss is 19.80 , inference loss is 12.93, weight deacy loss is 6.87, training accuracy is 0.218750, time 56.821 samples/sec
+epoch 9, total_step 150280, total loss is 12.21 , inference loss is 5.34, weight deacy loss is 6.87, training accuracy is 0.406250, time 56.720 samples/sec
+epoch 9, total_step 150300, total loss is 14.81 , inference loss is 7.94, weight deacy loss is 6.87, training accuracy is 0.406250, time 55.974 samples/sec
+epoch 9, total_step 150320, total loss is 18.63 , inference loss is 11.76, weight deacy loss is 6.87, training accuracy is 0.187500, time 56.372 samples/sec
+epoch 9, total_step 150340, total loss is 14.42 , inference loss is 7.55, weight deacy loss is 6.87, training accuracy is 0.406250, time 56.923 samples/sec
+epoch 9, total_step 150360, total loss is 14.42 , inference loss is 7.55, weight deacy loss is 6.87, training accuracy is 0.406250, time 56.949 samples/sec
+epoch 9, total_step 150380, total loss is 19.42 , inference loss is 12.55, weight deacy loss is 6.87, training accuracy is 0.437500, time 56.702 samples/sec
+epoch 9, total_step 150400, total loss is 14.72 , inference loss is 7.85, weight deacy loss is 6.87, training accuracy is 0.281250, time 56.771 samples/sec
+epoch 9, total_step 150420, total loss is 13.77 , inference loss is 6.90, weight deacy loss is 6.87, training accuracy is 0.406250, time 56.946 samples/sec
+epoch 9, total_step 150440, total loss is 14.68 , inference loss is 7.82, weight deacy loss is 6.87, training accuracy is 0.375000, time 56.691 samples/sec
+epoch 9, total_step 150460, total loss is 19.91 , inference loss is 13.04, weight deacy loss is 6.87, training accuracy is 0.281250, time 55.739 samples/sec
+epoch 9, total_step 150480, total loss is 15.04 , inference loss is 8.18, weight deacy loss is 6.87, training accuracy is 0.343750, time 56.258 samples/sec
+epoch 9, total_step 150500, total loss is 16.38 , inference loss is 9.52, weight deacy loss is 6.86, training accuracy is 0.312500, time 56.748 samples/sec
+epoch 9, total_step 150520, total loss is 15.55 , inference loss is 8.68, weight deacy loss is 6.86, training accuracy is 0.281250, time 56.376 samples/sec
+epoch 9, total_step 150540, total loss is 12.45 , inference loss is 5.59, weight deacy loss is 6.86, training accuracy is 0.343750, time 57.098 samples/sec
+epoch 9, total_step 150560, total loss is 14.58 , inference loss is 7.71, weight deacy loss is 6.86, training accuracy is 0.406250, time 56.720 samples/sec
+epoch 9, total_step 150580, total loss is 16.51 , inference loss is 9.65, weight deacy loss is 6.86, training accuracy is 0.437500, time 56.425 samples/sec
+epoch 9, total_step 150600, total loss is 16.91 , inference loss is 10.05, weight deacy loss is 6.86, training accuracy is 0.281250, time 55.065 samples/sec
+epoch 9, total_step 150620, total loss is 16.83 , inference loss is 9.96, weight deacy loss is 6.86, training accuracy is 0.343750, time 56.186 samples/sec
+epoch 9, total_step 150640, total loss is 17.81 , inference loss is 10.94, weight deacy loss is 6.86, training accuracy is 0.312500, time 56.925 samples/sec
+epoch 9, total_step 150660, total loss is 16.20 , inference loss is 9.34, weight deacy loss is 6.86, training accuracy is 0.312500, time 56.605 samples/sec
+epoch 9, total_step 150680, total loss is 13.90 , inference loss is 7.04, weight deacy loss is 6.86, training accuracy is 0.281250, time 56.109 samples/sec
+epoch 9, total_step 150700, total loss is 14.02 , inference loss is 7.15, weight deacy loss is 6.86, training accuracy is 0.406250, time 55.720 samples/sec
+epoch 9, total_step 150720, total loss is 17.54 , inference loss is 10.68, weight deacy loss is 6.86, training accuracy is 0.156250, time 56.817 samples/sec
+epoch 9, total_step 150740, total loss is 19.31 , inference loss is 12.45, weight deacy loss is 6.86, training accuracy is 0.218750, time 56.080 samples/sec
+epoch 9, total_step 150760, total loss is 14.50 , inference loss is 7.64, weight deacy loss is 6.86, training accuracy is 0.312500, time 56.164 samples/sec
+epoch 9, total_step 150780, total loss is 19.59 , inference loss is 12.73, weight deacy loss is 6.86, training accuracy is 0.250000, time 57.026 samples/sec
+epoch 9, total_step 150800, total loss is 17.94 , inference loss is 11.08, weight deacy loss is 6.86, training accuracy is 0.156250, time 56.313 samples/sec
+epoch 9, total_step 150820, total loss is 16.04 , inference loss is 9.18, weight deacy loss is 6.86, training accuracy is 0.343750, time 56.388 samples/sec
+epoch 9, total_step 150840, total loss is 16.52 , inference loss is 9.66, weight deacy loss is 6.86, training accuracy is 0.281250, time 56.585 samples/sec
+epoch 9, total_step 150860, total loss is 18.53 , inference loss is 11.67, weight deacy loss is 6.86, training accuracy is 0.281250, time 56.473 samples/sec
+epoch 9, total_step 150880, total loss is 14.91 , inference loss is 8.05, weight deacy loss is 6.86, training accuracy is 0.531250, time 56.690 samples/sec
+epoch 9, total_step 150900, total loss is 15.53 , inference loss is 8.67, weight deacy loss is 6.86, training accuracy is 0.375000, time 56.213 samples/sec
+epoch 9, total_step 150920, total loss is 21.57 , inference loss is 14.71, weight deacy loss is 6.86, training accuracy is 0.187500, time 53.904 samples/sec
+epoch 9, total_step 150940, total loss is 21.41 , inference loss is 14.55, weight deacy loss is 6.86, training accuracy is 0.312500, time 56.937 samples/sec
+epoch 9, total_step 150960, total loss is 19.76 , inference loss is 12.90, weight deacy loss is 6.86, training accuracy is 0.312500, time 56.708 samples/sec
+epoch 9, total_step 150980, total loss is 23.55 , inference loss is 16.69, weight deacy loss is 6.86, training accuracy is 0.281250, time 56.140 samples/sec
+epoch 9, total_step 151000, total loss is 17.88 , inference loss is 11.03, weight deacy loss is 6.86, training accuracy is 0.343750, time 56.198 samples/sec
+epoch 9, total_step 151020, total loss is 14.27 , inference loss is 7.41, weight deacy loss is 6.86, training accuracy is 0.281250, time 56.725 samples/sec
+epoch 9, total_step 151040, total loss is 16.52 , inference loss is 9.66, weight deacy loss is 6.86, training accuracy is 0.343750, time 57.166 samples/sec
+epoch 9, total_step 151060, total loss is 14.28 , inference loss is 7.43, weight deacy loss is 6.85, training accuracy is 0.312500, time 56.716 samples/sec
+epoch 9, total_step 151080, total loss is 19.52 , inference loss is 12.67, weight deacy loss is 6.85, training accuracy is 0.312500, time 56.419 samples/sec
+epoch 9, total_step 151100, total loss is 12.20 , inference loss is 5.35, weight deacy loss is 6.85, training accuracy is 0.406250, time 56.497 samples/sec
+epoch 9, total_step 151120, total loss is 19.54 , inference loss is 12.69, weight deacy loss is 6.85, training accuracy is 0.218750, time 56.653 samples/sec
+epoch 9, total_step 151140, total loss is 15.59 , inference loss is 8.73, weight deacy loss is 6.85, training accuracy is 0.156250, time 55.978 samples/sec
+epoch 9, total_step 151160, total loss is 16.43 , inference loss is 9.57, weight deacy loss is 6.85, training accuracy is 0.406250, time 56.661 samples/sec
+epoch 9, total_step 151180, total loss is 14.18 , inference loss is 7.33, weight deacy loss is 6.85, training accuracy is 0.375000, time 56.304 samples/sec
+epoch 9, total_step 151200, total loss is 18.14 , inference loss is 11.28, weight deacy loss is 6.85, training accuracy is 0.312500, time 56.162 samples/sec
+epoch 9, total_step 151220, total loss is 14.56 , inference loss is 7.71, weight deacy loss is 6.85, training accuracy is 0.437500, time 56.164 samples/sec
+epoch 9, total_step 151240, total loss is 18.70 , inference loss is 11.85, weight deacy loss is 6.85, training accuracy is 0.187500, time 57.556 samples/sec
+epoch 9, total_step 151260, total loss is 17.69 , inference loss is 10.84, weight deacy loss is 6.85, training accuracy is 0.156250, time 57.121 samples/sec
+epoch 9, total_step 151280, total loss is 16.41 , inference loss is 9.55, weight deacy loss is 6.85, training accuracy is 0.437500, time 55.331 samples/sec
+epoch 9, total_step 151300, total loss is 19.56 , inference loss is 12.71, weight deacy loss is 6.85, training accuracy is 0.343750, time 57.187 samples/sec
+epoch 9, total_step 151320, total loss is 14.20 , inference loss is 7.35, weight deacy loss is 6.85, training accuracy is 0.437500, time 56.849 samples/sec
+epoch 9, total_step 151340, total loss is 14.33 , inference loss is 7.48, weight deacy loss is 6.85, training accuracy is 0.312500, time 56.306 samples/sec
+epoch 9, total_step 151360, total loss is 17.38 , inference loss is 10.53, weight deacy loss is 6.85, training accuracy is 0.281250, time 56.037 samples/sec
+epoch 9, total_step 151380, total loss is 20.14 , inference loss is 13.29, weight deacy loss is 6.85, training accuracy is 0.281250, time 56.940 samples/sec
+epoch 9, total_step 151400, total loss is 13.22 , inference loss is 6.38, weight deacy loss is 6.85, training accuracy is 0.375000, time 56.708 samples/sec
+epoch 9, total_step 151420, total loss is 13.68 , inference loss is 6.83, weight deacy loss is 6.85, training accuracy is 0.500000, time 55.754 samples/sec
+epoch 9, total_step 151440, total loss is 15.38 , inference loss is 8.53, weight deacy loss is 6.85, training accuracy is 0.312500, time 56.694 samples/sec
+epoch 9, total_step 151460, total loss is 16.68 , inference loss is 9.83, weight deacy loss is 6.85, training accuracy is 0.312500, time 56.387 samples/sec
+epoch 9, total_step 151480, total loss is 12.38 , inference loss is 5.54, weight deacy loss is 6.85, training accuracy is 0.437500, time 55.828 samples/sec
+epoch 9, total_step 151500, total loss is 12.00 , inference loss is 5.15, weight deacy loss is 6.85, training accuracy is 0.437500, time 56.290 samples/sec
+epoch 9, total_step 151520, total loss is 14.19 , inference loss is 7.34, weight deacy loss is 6.85, training accuracy is 0.406250, time 56.923 samples/sec
+epoch 9, total_step 151540, total loss is 14.59 , inference loss is 7.74, weight deacy loss is 6.85, training accuracy is 0.281250, time 55.259 samples/sec
+epoch 9, total_step 151560, total loss is 16.56 , inference loss is 9.71, weight deacy loss is 6.85, training accuracy is 0.312500, time 55.442 samples/sec
+epoch 9, total_step 151580, total loss is 14.92 , inference loss is 8.08, weight deacy loss is 6.85, training accuracy is 0.343750, time 56.722 samples/sec
+epoch 9, total_step 151600, total loss is 19.17 , inference loss is 12.33, weight deacy loss is 6.85, training accuracy is 0.187500, time 56.726 samples/sec
+epoch 9, total_step 151620, total loss is 14.72 , inference loss is 7.87, weight deacy loss is 6.84, training accuracy is 0.343750, time 56.685 samples/sec
+epoch 9, total_step 151640, total loss is 18.15 , inference loss is 11.30, weight deacy loss is 6.84, training accuracy is 0.250000, time 56.224 samples/sec
+epoch 9, total_step 151660, total loss is 16.42 , inference loss is 9.58, weight deacy loss is 6.84, training accuracy is 0.281250, time 56.970 samples/sec
+epoch 9, total_step 151680, total loss is 15.56 , inference loss is 8.71, weight deacy loss is 6.84, training accuracy is 0.281250, time 55.933 samples/sec
+epoch 9, total_step 151700, total loss is 22.44 , inference loss is 15.60, weight deacy loss is 6.84, training accuracy is 0.218750, time 56.244 samples/sec
+epoch 9, total_step 151720, total loss is 16.20 , inference loss is 9.35, weight deacy loss is 6.84, training accuracy is 0.312500, time 57.301 samples/sec
+epoch 9, total_step 151740, total loss is 16.76 , inference loss is 9.92, weight deacy loss is 6.84, training accuracy is 0.218750, time 56.379 samples/sec
+epoch 9, total_step 151760, total loss is 21.08 , inference loss is 14.24, weight deacy loss is 6.84, training accuracy is 0.281250, time 56.777 samples/sec
+epoch 9, total_step 151780, total loss is 18.49 , inference loss is 11.65, weight deacy loss is 6.84, training accuracy is 0.250000, time 56.355 samples/sec
+epoch 9, total_step 151800, total loss is 12.28 , inference loss is 5.44, weight deacy loss is 6.84, training accuracy is 0.312500, time 56.703 samples/sec
+epoch 9, total_step 151820, total loss is 13.38 , inference loss is 6.54, weight deacy loss is 6.84, training accuracy is 0.406250, time 56.630 samples/sec
+epoch 9, total_step 151840, total loss is 18.03 , inference loss is 11.19, weight deacy loss is 6.84, training accuracy is 0.312500, time 56.417 samples/sec
+epoch 9, total_step 151860, total loss is 14.73 , inference loss is 7.89, weight deacy loss is 6.84, training accuracy is 0.406250, time 56.621 samples/sec
+epoch 9, total_step 151880, total loss is 16.13 , inference loss is 9.29, weight deacy loss is 6.84, training accuracy is 0.250000, time 56.503 samples/sec
+epoch 9, total_step 151900, total loss is 16.15 , inference loss is 9.31, weight deacy loss is 6.84, training accuracy is 0.218750, time 56.405 samples/sec
+epoch 9, total_step 151920, total loss is 15.09 , inference loss is 8.25, weight deacy loss is 6.84, training accuracy is 0.250000, time 56.659 samples/sec
+epoch 9, total_step 151940, total loss is 15.20 , inference loss is 8.36, weight deacy loss is 6.84, training accuracy is 0.531250, time 56.173 samples/sec
+epoch 9, total_step 151960, total loss is 14.42 , inference loss is 7.58, weight deacy loss is 6.84, training accuracy is 0.343750, time 56.508 samples/sec
+epoch 9, total_step 151980, total loss is 19.82 , inference loss is 12.98, weight deacy loss is 6.84, training accuracy is 0.312500, time 56.266 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.40948599999993
+best_threshold_index 108 0.9864814814814815
+best_threshold_index 111 0.9851851851851852
+best_threshold_index 111 0.9859259259259259
+best_threshold_index 111 0.9864814814814815
+best_threshold_index 111 0.9857407407407407
+best_threshold_index 111 0.9844444444444445
+best_threshold_index 111 0.9855555555555555
+best_threshold_index 111 0.9853703703703703
+best_threshold_index 111 0.9837037037037037
+best_threshold_index 111 0.9844444444444445
+[lfw][152000]XNorm: 22.861843
+[lfw][152000]Accuracy-Flip: 0.98467+-0.00884
+testing verification..
+(14000, 512)
+infer time 85.21958100000005
+best_threshold_index 116 0.9855555555555555
+best_threshold_index 122 0.9855555555555555
+best_threshold_index 122 0.9852380952380952
+best_threshold_index 121 0.9858730158730159
+best_threshold_index 121 0.9865079365079366
+best_threshold_index 120 0.9850793650793651
+best_threshold_index 122 0.9861904761904762
+best_threshold_index 122 0.9852380952380952
+best_threshold_index 122 0.9855555555555555
+best_threshold_index 122 0.9865079365079366
+[cfp_ff][152000]XNorm: 22.559153
+[cfp_ff][152000]Accuracy-Flip: 0.98429+-0.00461
+testing verification..
+(14000, 512)
+infer time 86.25256499999998
+best_threshold_index 135 0.9203174603174603
+best_threshold_index 135 0.9223809523809524
+best_threshold_index 135 0.9201587301587302
+best_threshold_index 135 0.9233333333333333
+best_threshold_index 135 0.9255555555555556
+best_threshold_index 135 0.9201587301587302
+best_threshold_index 135 0.919047619047619
+best_threshold_index 135 0.9215873015873016
+best_threshold_index 135 0.9215873015873016
+best_threshold_index 139 0.9219047619047619
+[cfp_fp][152000]XNorm: 22.500615
+[cfp_fp][152000]Accuracy-Flip: 0.92071+-0.01620
+testing verification..
+(12000, 512)
+infer time 73.03782699999995
+best_threshold_index 140 0.8970370370370371
+best_threshold_index 138 0.8957407407407407
+best_threshold_index 140 0.8992592592592593
+best_threshold_index 138 0.9
+best_threshold_index 140 0.902962962962963
+best_threshold_index 140 0.8968518518518519
+best_threshold_index 140 0.8985185185185185
+best_threshold_index 144 0.9031481481481481
+best_threshold_index 140 0.9011111111111111
+best_threshold_index 140 0.9016666666666666
+[agedb_30][152000]XNorm: 23.223230
+[agedb_30][152000]Accuracy-Flip: 0.89650+-0.02287
+test accuracy is: 0.9846666666666668
+epoch 9, total_step 152000, total loss is 17.35 , inference loss is 10.51, weight deacy loss is 6.84, training accuracy is 0.343750, time 54.507 samples/sec
+epoch 9, total_step 152020, total loss is 20.67 , inference loss is 13.83, weight deacy loss is 6.84, training accuracy is 0.218750, time 56.361 samples/sec
+epoch 9, total_step 152040, total loss is 13.40 , inference loss is 6.56, weight deacy loss is 6.84, training accuracy is 0.406250, time 56.918 samples/sec
+epoch 9, total_step 152060, total loss is 18.64 , inference loss is 11.81, weight deacy loss is 6.84, training accuracy is 0.406250, time 57.119 samples/sec
+epoch 9, total_step 152080, total loss is 18.75 , inference loss is 11.91, weight deacy loss is 6.84, training accuracy is 0.218750, time 56.484 samples/sec
+epoch 9, total_step 152100, total loss is 16.47 , inference loss is 9.64, weight deacy loss is 6.84, training accuracy is 0.343750, time 55.997 samples/sec
+epoch 9, total_step 152120, total loss is 13.82 , inference loss is 6.99, weight deacy loss is 6.84, training accuracy is 0.343750, time 56.958 samples/sec
+epoch 9, total_step 152140, total loss is 18.26 , inference loss is 11.43, weight deacy loss is 6.84, training accuracy is 0.218750, time 57.138 samples/sec
+epoch 9, total_step 152160, total loss is 17.05 , inference loss is 10.22, weight deacy loss is 6.84, training accuracy is 0.343750, time 57.716 samples/sec
+epoch 9, total_step 152180, total loss is 16.81 , inference loss is 9.97, weight deacy loss is 6.84, training accuracy is 0.437500, time 56.298 samples/sec
+epoch 9, total_step 152200, total loss is 20.81 , inference loss is 13.97, weight deacy loss is 6.83, training accuracy is 0.218750, time 56.078 samples/sec
+epoch 9, total_step 152220, total loss is 15.77 , inference loss is 8.94, weight deacy loss is 6.83, training accuracy is 0.281250, time 56.625 samples/sec
+epoch 9, total_step 152240, total loss is 17.45 , inference loss is 10.62, weight deacy loss is 6.83, training accuracy is 0.218750, time 56.736 samples/sec
+epoch 9, total_step 152260, total loss is 18.18 , inference loss is 11.35, weight deacy loss is 6.83, training accuracy is 0.406250, time 56.639 samples/sec
+epoch 9, total_step 152280, total loss is 16.74 , inference loss is 9.91, weight deacy loss is 6.83, training accuracy is 0.375000, time 56.693 samples/sec
+epoch 9, total_step 152300, total loss is 18.40 , inference loss is 11.57, weight deacy loss is 6.83, training accuracy is 0.218750, time 57.185 samples/sec
+epoch 9, total_step 152320, total loss is 16.69 , inference loss is 9.86, weight deacy loss is 6.83, training accuracy is 0.375000, time 56.633 samples/sec
+epoch 9, total_step 152340, total loss is 16.38 , inference loss is 9.55, weight deacy loss is 6.83, training accuracy is 0.343750, time 56.797 samples/sec
+epoch 9, total_step 152360, total loss is 14.86 , inference loss is 8.03, weight deacy loss is 6.83, training accuracy is 0.343750, time 55.917 samples/sec
+epoch 9, total_step 152380, total loss is 17.65 , inference loss is 10.82, weight deacy loss is 6.83, training accuracy is 0.312500, time 56.163 samples/sec
+epoch 9, total_step 152400, total loss is 12.42 , inference loss is 5.59, weight deacy loss is 6.83, training accuracy is 0.406250, time 56.260 samples/sec
+epoch 9, total_step 152420, total loss is 16.76 , inference loss is 9.93, weight deacy loss is 6.83, training accuracy is 0.250000, time 57.448 samples/sec
+epoch 9, total_step 152440, total loss is 14.65 , inference loss is 7.81, weight deacy loss is 6.83, training accuracy is 0.250000, time 57.177 samples/sec
+epoch 9, total_step 152460, total loss is 17.94 , inference loss is 11.11, weight deacy loss is 6.83, training accuracy is 0.218750, time 56.182 samples/sec
+epoch 9, total_step 152480, total loss is 13.86 , inference loss is 7.03, weight deacy loss is 6.83, training accuracy is 0.406250, time 57.340 samples/sec
+epoch 9, total_step 152500, total loss is 18.48 , inference loss is 11.66, weight deacy loss is 6.83, training accuracy is 0.281250, time 55.946 samples/sec
+epoch 9, total_step 152520, total loss is 12.35 , inference loss is 5.52, weight deacy loss is 6.83, training accuracy is 0.468750, time 56.039 samples/sec
+epoch 9, total_step 152540, total loss is 17.11 , inference loss is 10.28, weight deacy loss is 6.83, training accuracy is 0.187500, time 56.647 samples/sec
+epoch 9, total_step 152560, total loss is 17.83 , inference loss is 11.00, weight deacy loss is 6.83, training accuracy is 0.187500, time 56.418 samples/sec
+epoch 9, total_step 152580, total loss is 14.98 , inference loss is 8.15, weight deacy loss is 6.83, training accuracy is 0.375000, time 56.188 samples/sec
+epoch 9, total_step 152600, total loss is 18.96 , inference loss is 12.14, weight deacy loss is 6.83, training accuracy is 0.218750, time 57.202 samples/sec
+epoch 9, total_step 152620, total loss is 19.25 , inference loss is 12.42, weight deacy loss is 6.83, training accuracy is 0.281250, time 56.498 samples/sec
+epoch 9, total_step 152640, total loss is 15.02 , inference loss is 8.19, weight deacy loss is 6.83, training accuracy is 0.406250, time 56.758 samples/sec
+epoch 9, total_step 152660, total loss is 18.93 , inference loss is 12.10, weight deacy loss is 6.83, training accuracy is 0.218750, time 57.471 samples/sec
+epoch 9, total_step 152680, total loss is 17.02 , inference loss is 10.20, weight deacy loss is 6.83, training accuracy is 0.312500, time 56.246 samples/sec
+epoch 9, total_step 152700, total loss is 16.63 , inference loss is 9.80, weight deacy loss is 6.83, training accuracy is 0.281250, time 56.672 samples/sec
+epoch 9, total_step 152720, total loss is 15.49 , inference loss is 8.67, weight deacy loss is 6.83, training accuracy is 0.343750, time 56.735 samples/sec
+epoch 9, total_step 152740, total loss is 13.42 , inference loss is 6.59, weight deacy loss is 6.83, training accuracy is 0.281250, time 55.993 samples/sec
+epoch 9, total_step 152760, total loss is 20.05 , inference loss is 13.22, weight deacy loss is 6.82, training accuracy is 0.156250, time 56.460 samples/sec
+epoch 9, total_step 152780, total loss is 14.68 , inference loss is 7.85, weight deacy loss is 6.82, training accuracy is 0.406250, time 56.105 samples/sec
+epoch 9, total_step 152800, total loss is 14.79 , inference loss is 7.97, weight deacy loss is 6.82, training accuracy is 0.343750, time 56.231 samples/sec
+epoch 9, total_step 152820, total loss is 16.77 , inference loss is 9.95, weight deacy loss is 6.82, training accuracy is 0.312500, time 56.357 samples/sec
+epoch 9, total_step 152840, total loss is 15.52 , inference loss is 8.70, weight deacy loss is 6.82, training accuracy is 0.375000, time 56.578 samples/sec
+epoch 9, total_step 152860, total loss is 19.48 , inference loss is 12.66, weight deacy loss is 6.82, training accuracy is 0.250000, time 56.035 samples/sec
+epoch 9, total_step 152880, total loss is 15.96 , inference loss is 9.14, weight deacy loss is 6.82, training accuracy is 0.156250, time 56.009 samples/sec
+epoch 9, total_step 152900, total loss is 18.03 , inference loss is 11.20, weight deacy loss is 6.82, training accuracy is 0.156250, time 56.212 samples/sec
+epoch 9, total_step 152920, total loss is 14.50 , inference loss is 7.68, weight deacy loss is 6.82, training accuracy is 0.250000, time 56.806 samples/sec
+epoch 9, total_step 152940, total loss is 18.58 , inference loss is 11.76, weight deacy loss is 6.82, training accuracy is 0.281250, time 56.979 samples/sec
+epoch 9, total_step 152960, total loss is 15.97 , inference loss is 9.14, weight deacy loss is 6.82, training accuracy is 0.250000, time 56.244 samples/sec
+epoch 9, total_step 152980, total loss is 15.23 , inference loss is 8.41, weight deacy loss is 6.82, training accuracy is 0.250000, time 56.225 samples/sec
+epoch 9, total_step 153000, total loss is 20.25 , inference loss is 13.43, weight deacy loss is 6.82, training accuracy is 0.187500, time 56.483 samples/sec
+epoch 9, total_step 153020, total loss is 18.36 , inference loss is 11.54, weight deacy loss is 6.82, training accuracy is 0.281250, time 57.122 samples/sec
+epoch 9, total_step 153040, total loss is 15.48 , inference loss is 8.66, weight deacy loss is 6.82, training accuracy is 0.375000, time 56.514 samples/sec
+epoch 9, total_step 153060, total loss is 17.64 , inference loss is 10.82, weight deacy loss is 6.82, training accuracy is 0.156250, time 55.910 samples/sec
+epoch 9, total_step 153080, total loss is 16.97 , inference loss is 10.15, weight deacy loss is 6.82, training accuracy is 0.218750, time 56.623 samples/sec
+epoch 9, total_step 153100, total loss is 19.10 , inference loss is 12.28, weight deacy loss is 6.82, training accuracy is 0.250000, time 56.994 samples/sec
+epoch 9, total_step 153120, total loss is 18.03 , inference loss is 11.21, weight deacy loss is 6.82, training accuracy is 0.156250, time 56.230 samples/sec
+epoch 9, total_step 153140, total loss is 18.13 , inference loss is 11.31, weight deacy loss is 6.82, training accuracy is 0.437500, time 56.618 samples/sec
+epoch 9, total_step 153160, total loss is 19.05 , inference loss is 12.23, weight deacy loss is 6.82, training accuracy is 0.218750, time 56.624 samples/sec
+epoch 9, total_step 153180, total loss is 16.06 , inference loss is 9.24, weight deacy loss is 6.82, training accuracy is 0.312500, time 54.116 samples/sec
+epoch 9, total_step 153200, total loss is 19.66 , inference loss is 12.84, weight deacy loss is 6.82, training accuracy is 0.250000, time 54.730 samples/sec
+epoch 9, total_step 153220, total loss is 16.89 , inference loss is 10.07, weight deacy loss is 6.82, training accuracy is 0.250000, time 56.313 samples/sec
+epoch 9, total_step 153240, total loss is 16.15 , inference loss is 9.34, weight deacy loss is 6.82, training accuracy is 0.250000, time 56.606 samples/sec
+epoch 9, total_step 153260, total loss is 14.73 , inference loss is 7.91, weight deacy loss is 6.82, training accuracy is 0.468750, time 56.625 samples/sec
+epoch 9, total_step 153280, total loss is 15.99 , inference loss is 9.17, weight deacy loss is 6.82, training accuracy is 0.375000, time 56.825 samples/sec
+epoch 9, total_step 153300, total loss is 18.96 , inference loss is 12.15, weight deacy loss is 6.82, training accuracy is 0.281250, time 55.331 samples/sec
+
+
+
+
+End of epoch 9
+epoch 10, total_step 153320, total loss is 25.93 , inference loss is 19.11, weight deacy loss is 6.82, training accuracy is 0.000000, time 53.733 samples/sec
+epoch 10, total_step 153340, total loss is 15.39 , inference loss is 8.58, weight deacy loss is 6.82, training accuracy is 0.406250, time 56.864 samples/sec
+epoch 10, total_step 153360, total loss is 13.85 , inference loss is 7.04, weight deacy loss is 6.82, training accuracy is 0.500000, time 56.789 samples/sec
+epoch 10, total_step 153380, total loss is 12.71 , inference loss is 5.89, weight deacy loss is 6.81, training accuracy is 0.687500, time 57.043 samples/sec
+epoch 10, total_step 153400, total loss is 15.06 , inference loss is 8.25, weight deacy loss is 6.81, training accuracy is 0.531250, time 56.800 samples/sec
+epoch 10, total_step 153420, total loss is 11.34 , inference loss is 4.53, weight deacy loss is 6.81, training accuracy is 0.781250, time 55.899 samples/sec
+epoch 10, total_step 153440, total loss is 10.22 , inference loss is 3.41, weight deacy loss is 6.81, training accuracy is 0.687500, time 56.872 samples/sec
+epoch 10, total_step 153460, total loss is 14.43 , inference loss is 7.62, weight deacy loss is 6.81, training accuracy is 0.531250, time 55.797 samples/sec
+epoch 10, total_step 153480, total loss is 13.93 , inference loss is 7.12, weight deacy loss is 6.81, training accuracy is 0.500000, time 56.605 samples/sec
+epoch 10, total_step 153500, total loss is 11.54 , inference loss is 4.73, weight deacy loss is 6.81, training accuracy is 0.625000, time 55.736 samples/sec
+epoch 10, total_step 153520, total loss is 11.53 , inference loss is 4.71, weight deacy loss is 6.81, training accuracy is 0.625000, time 56.798 samples/sec
+epoch 10, total_step 153540, total loss is 12.97 , inference loss is 6.16, weight deacy loss is 6.81, training accuracy is 0.750000, time 56.807 samples/sec
+epoch 10, total_step 153560, total loss is 8.94 , inference loss is 2.13, weight deacy loss is 6.81, training accuracy is 0.750000, time 56.013 samples/sec
+epoch 10, total_step 153580, total loss is 10.84 , inference loss is 4.03, weight deacy loss is 6.81, training accuracy is 0.656250, time 56.839 samples/sec
+epoch 10, total_step 153600, total loss is 11.81 , inference loss is 5.00, weight deacy loss is 6.81, training accuracy is 0.781250, time 56.114 samples/sec
+epoch 10, total_step 153620, total loss is 13.08 , inference loss is 6.27, weight deacy loss is 6.81, training accuracy is 0.468750, time 55.802 samples/sec
+epoch 10, total_step 153640, total loss is 15.18 , inference loss is 8.37, weight deacy loss is 6.81, training accuracy is 0.562500, time 56.223 samples/sec
+epoch 10, total_step 153660, total loss is 15.29 , inference loss is 8.48, weight deacy loss is 6.81, training accuracy is 0.437500, time 56.657 samples/sec
+epoch 10, total_step 153680, total loss is 12.85 , inference loss is 6.04, weight deacy loss is 6.81, training accuracy is 0.562500, time 56.295 samples/sec
+epoch 10, total_step 153700, total loss is 12.95 , inference loss is 6.14, weight deacy loss is 6.81, training accuracy is 0.656250, time 56.437 samples/sec
+epoch 10, total_step 153720, total loss is 14.96 , inference loss is 8.15, weight deacy loss is 6.81, training accuracy is 0.562500, time 57.166 samples/sec
+epoch 10, total_step 153740, total loss is 12.29 , inference loss is 5.48, weight deacy loss is 6.81, training accuracy is 0.593750, time 56.263 samples/sec
+epoch 10, total_step 153760, total loss is 11.06 , inference loss is 4.25, weight deacy loss is 6.81, training accuracy is 0.687500, time 57.185 samples/sec
+epoch 10, total_step 153780, total loss is 12.67 , inference loss is 5.86, weight deacy loss is 6.81, training accuracy is 0.625000, time 55.420 samples/sec
+epoch 10, total_step 153800, total loss is 14.66 , inference loss is 7.85, weight deacy loss is 6.81, training accuracy is 0.593750, time 56.784 samples/sec
+epoch 10, total_step 153820, total loss is 10.29 , inference loss is 3.48, weight deacy loss is 6.81, training accuracy is 0.687500, time 56.533 samples/sec
+epoch 10, total_step 153840, total loss is 12.83 , inference loss is 6.02, weight deacy loss is 6.81, training accuracy is 0.625000, time 57.323 samples/sec
+epoch 10, total_step 153860, total loss is 14.96 , inference loss is 8.15, weight deacy loss is 6.81, training accuracy is 0.468750, time 56.535 samples/sec
+epoch 10, total_step 153880, total loss is 10.78 , inference loss is 3.98, weight deacy loss is 6.81, training accuracy is 0.718750, time 56.532 samples/sec
+epoch 10, total_step 153900, total loss is 13.33 , inference loss is 6.52, weight deacy loss is 6.81, training accuracy is 0.531250, time 54.809 samples/sec
+epoch 10, total_step 153920, total loss is 11.12 , inference loss is 4.31, weight deacy loss is 6.81, training accuracy is 0.625000, time 56.420 samples/sec
+epoch 10, total_step 153940, total loss is 12.42 , inference loss is 5.61, weight deacy loss is 6.81, training accuracy is 0.593750, time 56.980 samples/sec
+epoch 10, total_step 153960, total loss is 14.50 , inference loss is 7.69, weight deacy loss is 6.81, training accuracy is 0.562500, time 55.823 samples/sec
+epoch 10, total_step 153980, total loss is 10.21 , inference loss is 3.40, weight deacy loss is 6.81, training accuracy is 0.781250, time 56.153 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.93744500000003
+best_threshold_index 113 0.9837037037037037
+best_threshold_index 111 0.9835185185185186
+best_threshold_index 113 0.9835185185185186
+best_threshold_index 113 0.9838888888888889
+best_threshold_index 117 0.9840740740740741
+best_threshold_index 113 0.9833333333333333
+best_threshold_index 117 0.9837037037037037
+best_threshold_index 113 0.9842592592592593
+best_threshold_index 113 0.9827777777777778
+best_threshold_index 113 0.9827777777777778
+[lfw][154000]XNorm: 19.084292
+[lfw][154000]Accuracy-Flip: 0.98233+-0.00455
+testing verification..
+(14000, 512)
+infer time 85.999045
+best_threshold_index 109 0.9806349206349206
+best_threshold_index 108 0.9787301587301588
+best_threshold_index 108 0.9792063492063492
+best_threshold_index 108 0.979047619047619
+best_threshold_index 108 0.9788888888888889
+best_threshold_index 108 0.9784126984126984
+best_threshold_index 108 0.9796825396825397
+best_threshold_index 108 0.9785714285714285
+best_threshold_index 108 0.9785714285714285
+best_threshold_index 107 0.9801587301587301
+[cfp_ff][154000]XNorm: 18.875987
+[cfp_ff][154000]Accuracy-Flip: 0.97829+-0.00705
+testing verification..
+(14000, 512)
+infer time 86.08605599999996
+best_threshold_index 130 0.9077777777777778
+best_threshold_index 130 0.9084126984126984
+best_threshold_index 130 0.9077777777777778
+best_threshold_index 130 0.9082539682539682
+best_threshold_index 130 0.9114285714285715
+best_threshold_index 130 0.9063492063492063
+best_threshold_index 129 0.9057142857142857
+best_threshold_index 130 0.9061904761904762
+best_threshold_index 130 0.9074603174603174
+best_threshold_index 130 0.9077777777777778
+[cfp_fp][154000]XNorm: 20.784029
+[cfp_fp][154000]Accuracy-Flip: 0.90743+-0.01324
+testing verification..
+(12000, 512)
+infer time 72.93862800000001
+best_threshold_index 134 0.9001851851851852
+best_threshold_index 134 0.8992592592592593
+best_threshold_index 134 0.9012962962962963
+best_threshold_index 134 0.9012962962962963
+best_threshold_index 134 0.9062962962962963
+best_threshold_index 134 0.8998148148148148
+best_threshold_index 134 0.9027777777777778
+best_threshold_index 134 0.9044444444444445
+best_threshold_index 134 0.9044444444444445
+best_threshold_index 134 0.9035185185185185
+[agedb_30][154000]XNorm: 20.573847
+[agedb_30][154000]Accuracy-Flip: 0.90233+-0.01988
+test accuracy is: 0.9823333333333333
+epoch 10, total_step 154000, total loss is 14.80 , inference loss is 7.99, weight deacy loss is 6.81, training accuracy is 0.468750, time 53.234 samples/sec
+epoch 10, total_step 154020, total loss is 13.03 , inference loss is 6.22, weight deacy loss is 6.81, training accuracy is 0.531250, time 56.676 samples/sec
+epoch 10, total_step 154040, total loss is 18.96 , inference loss is 12.15, weight deacy loss is 6.81, training accuracy is 0.375000, time 56.646 samples/sec
+epoch 10, total_step 154060, total loss is 13.62 , inference loss is 6.81, weight deacy loss is 6.81, training accuracy is 0.531250, time 55.579 samples/sec
+epoch 10, total_step 154080, total loss is 17.21 , inference loss is 10.40, weight deacy loss is 6.81, training accuracy is 0.500000, time 56.750 samples/sec
+epoch 10, total_step 154100, total loss is 15.03 , inference loss is 8.23, weight deacy loss is 6.81, training accuracy is 0.500000, time 57.150 samples/sec
+epoch 10, total_step 154120, total loss is 11.17 , inference loss is 4.36, weight deacy loss is 6.81, training accuracy is 0.531250, time 56.926 samples/sec
+epoch 10, total_step 154140, total loss is 13.13 , inference loss is 6.32, weight deacy loss is 6.81, training accuracy is 0.625000, time 57.250 samples/sec
+epoch 10, total_step 154160, total loss is 11.03 , inference loss is 4.23, weight deacy loss is 6.81, training accuracy is 0.562500, time 56.406 samples/sec
+epoch 10, total_step 154180, total loss is 12.07 , inference loss is 5.27, weight deacy loss is 6.81, training accuracy is 0.718750, time 57.123 samples/sec
+epoch 10, total_step 154200, total loss is 16.82 , inference loss is 10.01, weight deacy loss is 6.81, training accuracy is 0.437500, time 56.306 samples/sec
+epoch 10, total_step 154220, total loss is 12.73 , inference loss is 5.92, weight deacy loss is 6.81, training accuracy is 0.468750, time 56.635 samples/sec
+epoch 10, total_step 154240, total loss is 12.08 , inference loss is 5.27, weight deacy loss is 6.81, training accuracy is 0.625000, time 56.317 samples/sec
+epoch 10, total_step 154260, total loss is 12.37 , inference loss is 5.56, weight deacy loss is 6.81, training accuracy is 0.562500, time 57.136 samples/sec
+epoch 10, total_step 154280, total loss is 13.37 , inference loss is 6.56, weight deacy loss is 6.81, training accuracy is 0.500000, time 55.826 samples/sec
+epoch 10, total_step 154300, total loss is 16.72 , inference loss is 9.91, weight deacy loss is 6.81, training accuracy is 0.656250, time 56.512 samples/sec
+epoch 10, total_step 154320, total loss is 15.81 , inference loss is 9.01, weight deacy loss is 6.81, training accuracy is 0.500000, time 56.533 samples/sec
+epoch 10, total_step 154340, total loss is 14.43 , inference loss is 7.63, weight deacy loss is 6.81, training accuracy is 0.437500, time 56.587 samples/sec
+epoch 10, total_step 154360, total loss is 11.85 , inference loss is 5.04, weight deacy loss is 6.81, training accuracy is 0.531250, time 56.602 samples/sec
+epoch 10, total_step 154380, total loss is 17.39 , inference loss is 10.59, weight deacy loss is 6.81, training accuracy is 0.468750, time 55.423 samples/sec
+epoch 10, total_step 154400, total loss is 13.88 , inference loss is 7.07, weight deacy loss is 6.81, training accuracy is 0.593750, time 56.309 samples/sec
+epoch 10, total_step 154420, total loss is 15.59 , inference loss is 8.78, weight deacy loss is 6.81, training accuracy is 0.343750, time 56.206 samples/sec
+epoch 10, total_step 154440, total loss is 11.70 , inference loss is 4.89, weight deacy loss is 6.81, training accuracy is 0.562500, time 56.281 samples/sec
+epoch 10, total_step 154460, total loss is 12.86 , inference loss is 6.05, weight deacy loss is 6.81, training accuracy is 0.468750, time 56.774 samples/sec
+epoch 10, total_step 154480, total loss is 13.73 , inference loss is 6.93, weight deacy loss is 6.81, training accuracy is 0.500000, time 56.916 samples/sec
+epoch 10, total_step 154500, total loss is 10.64 , inference loss is 3.83, weight deacy loss is 6.81, training accuracy is 0.625000, time 55.994 samples/sec
+epoch 10, total_step 154520, total loss is 13.14 , inference loss is 6.34, weight deacy loss is 6.81, training accuracy is 0.562500, time 55.927 samples/sec
+epoch 10, total_step 154540, total loss is 16.42 , inference loss is 9.61, weight deacy loss is 6.81, training accuracy is 0.437500, time 56.964 samples/sec
+epoch 10, total_step 154560, total loss is 12.91 , inference loss is 6.11, weight deacy loss is 6.81, training accuracy is 0.468750, time 56.667 samples/sec
+epoch 10, total_step 154580, total loss is 14.03 , inference loss is 7.22, weight deacy loss is 6.81, training accuracy is 0.406250, time 57.578 samples/sec
+epoch 10, total_step 154600, total loss is 12.16 , inference loss is 5.35, weight deacy loss is 6.81, training accuracy is 0.437500, time 56.042 samples/sec
+epoch 10, total_step 154620, total loss is 12.35 , inference loss is 5.54, weight deacy loss is 6.81, training accuracy is 0.593750, time 56.846 samples/sec
+epoch 10, total_step 154640, total loss is 15.04 , inference loss is 8.24, weight deacy loss is 6.81, training accuracy is 0.468750, time 56.186 samples/sec
+epoch 10, total_step 154660, total loss is 10.57 , inference loss is 3.76, weight deacy loss is 6.81, training accuracy is 0.593750, time 56.231 samples/sec
+epoch 10, total_step 154680, total loss is 15.55 , inference loss is 8.75, weight deacy loss is 6.81, training accuracy is 0.406250, time 56.662 samples/sec
+epoch 10, total_step 154700, total loss is 11.75 , inference loss is 4.95, weight deacy loss is 6.81, training accuracy is 0.500000, time 56.892 samples/sec
+epoch 10, total_step 154720, total loss is 12.57 , inference loss is 5.76, weight deacy loss is 6.81, training accuracy is 0.531250, time 56.988 samples/sec
+epoch 10, total_step 154740, total loss is 16.71 , inference loss is 9.90, weight deacy loss is 6.81, training accuracy is 0.375000, time 57.355 samples/sec
+epoch 10, total_step 154760, total loss is 14.68 , inference loss is 7.87, weight deacy loss is 6.81, training accuracy is 0.406250, time 56.492 samples/sec
+epoch 10, total_step 154780, total loss is 14.06 , inference loss is 7.26, weight deacy loss is 6.81, training accuracy is 0.562500, time 56.610 samples/sec
+epoch 10, total_step 154800, total loss is 13.63 , inference loss is 6.82, weight deacy loss is 6.81, training accuracy is 0.593750, time 56.774 samples/sec
+epoch 10, total_step 154820, total loss is 12.18 , inference loss is 5.37, weight deacy loss is 6.81, training accuracy is 0.406250, time 56.559 samples/sec
+epoch 10, total_step 154840, total loss is 14.39 , inference loss is 7.59, weight deacy loss is 6.81, training accuracy is 0.500000, time 56.523 samples/sec
+epoch 10, total_step 154860, total loss is 18.66 , inference loss is 11.86, weight deacy loss is 6.81, training accuracy is 0.406250, time 57.078 samples/sec
+epoch 10, total_step 154880, total loss is 14.44 , inference loss is 7.63, weight deacy loss is 6.81, training accuracy is 0.468750, time 55.749 samples/sec
+epoch 10, total_step 154900, total loss is 9.40 , inference loss is 2.60, weight deacy loss is 6.81, training accuracy is 0.625000, time 56.690 samples/sec
+epoch 10, total_step 154920, total loss is 9.77 , inference loss is 2.96, weight deacy loss is 6.81, training accuracy is 0.781250, time 57.170 samples/sec
+epoch 10, total_step 154940, total loss is 15.74 , inference loss is 8.93, weight deacy loss is 6.81, training accuracy is 0.468750, time 56.164 samples/sec
+epoch 10, total_step 154960, total loss is 11.58 , inference loss is 4.77, weight deacy loss is 6.81, training accuracy is 0.562500, time 57.007 samples/sec
+epoch 10, total_step 154980, total loss is 12.15 , inference loss is 5.34, weight deacy loss is 6.81, training accuracy is 0.562500, time 56.949 samples/sec
+epoch 10, total_step 155000, total loss is 17.86 , inference loss is 11.06, weight deacy loss is 6.80, training accuracy is 0.437500, time 57.003 samples/sec
+epoch 10, total_step 155020, total loss is 16.20 , inference loss is 9.40, weight deacy loss is 6.80, training accuracy is 0.437500, time 56.523 samples/sec
+epoch 10, total_step 155040, total loss is 15.05 , inference loss is 8.25, weight deacy loss is 6.80, training accuracy is 0.343750, time 56.448 samples/sec
+epoch 10, total_step 155060, total loss is 14.09 , inference loss is 7.29, weight deacy loss is 6.80, training accuracy is 0.500000, time 56.246 samples/sec
+epoch 10, total_step 155080, total loss is 12.35 , inference loss is 5.55, weight deacy loss is 6.80, training accuracy is 0.500000, time 56.419 samples/sec
+epoch 10, total_step 155100, total loss is 10.85 , inference loss is 4.04, weight deacy loss is 6.80, training accuracy is 0.656250, time 56.076 samples/sec
+epoch 10, total_step 155120, total loss is 14.84 , inference loss is 8.03, weight deacy loss is 6.80, training accuracy is 0.468750, time 56.738 samples/sec
+epoch 10, total_step 155140, total loss is 15.06 , inference loss is 8.25, weight deacy loss is 6.80, training accuracy is 0.406250, time 57.461 samples/sec
+epoch 10, total_step 155160, total loss is 12.75 , inference loss is 5.94, weight deacy loss is 6.80, training accuracy is 0.562500, time 56.420 samples/sec
+epoch 10, total_step 155180, total loss is 12.32 , inference loss is 5.51, weight deacy loss is 6.80, training accuracy is 0.468750, time 56.302 samples/sec
+epoch 10, total_step 155200, total loss is 13.63 , inference loss is 6.82, weight deacy loss is 6.80, training accuracy is 0.468750, time 56.467 samples/sec
+epoch 10, total_step 155220, total loss is 17.19 , inference loss is 10.38, weight deacy loss is 6.80, training accuracy is 0.406250, time 55.927 samples/sec
+epoch 10, total_step 155240, total loss is 14.28 , inference loss is 7.48, weight deacy loss is 6.80, training accuracy is 0.468750, time 57.447 samples/sec
+epoch 10, total_step 155260, total loss is 10.42 , inference loss is 3.61, weight deacy loss is 6.80, training accuracy is 0.593750, time 57.023 samples/sec
+epoch 10, total_step 155280, total loss is 14.01 , inference loss is 7.20, weight deacy loss is 6.80, training accuracy is 0.625000, time 55.911 samples/sec
+epoch 10, total_step 155300, total loss is 11.13 , inference loss is 4.33, weight deacy loss is 6.80, training accuracy is 0.500000, time 56.615 samples/sec
+epoch 10, total_step 155320, total loss is 14.08 , inference loss is 7.27, weight deacy loss is 6.80, training accuracy is 0.500000, time 56.468 samples/sec
+epoch 10, total_step 155340, total loss is 14.30 , inference loss is 7.50, weight deacy loss is 6.80, training accuracy is 0.406250, time 57.101 samples/sec
+epoch 10, total_step 155360, total loss is 13.82 , inference loss is 7.02, weight deacy loss is 6.80, training accuracy is 0.562500, time 56.635 samples/sec
+epoch 10, total_step 155380, total loss is 13.09 , inference loss is 6.29, weight deacy loss is 6.80, training accuracy is 0.562500, time 56.716 samples/sec
+epoch 10, total_step 155400, total loss is 11.27 , inference loss is 4.47, weight deacy loss is 6.80, training accuracy is 0.531250, time 55.812 samples/sec
+epoch 10, total_step 155420, total loss is 15.16 , inference loss is 8.36, weight deacy loss is 6.80, training accuracy is 0.468750, time 56.415 samples/sec
+epoch 10, total_step 155440, total loss is 11.91 , inference loss is 5.11, weight deacy loss is 6.80, training accuracy is 0.718750, time 56.312 samples/sec
+epoch 10, total_step 155460, total loss is 14.64 , inference loss is 7.84, weight deacy loss is 6.80, training accuracy is 0.437500, time 55.456 samples/sec
+epoch 10, total_step 155480, total loss is 13.13 , inference loss is 6.33, weight deacy loss is 6.80, training accuracy is 0.468750, time 56.594 samples/sec
+epoch 10, total_step 155500, total loss is 12.80 , inference loss is 6.00, weight deacy loss is 6.80, training accuracy is 0.562500, time 56.096 samples/sec
+epoch 10, total_step 155520, total loss is 13.09 , inference loss is 6.29, weight deacy loss is 6.80, training accuracy is 0.625000, time 56.747 samples/sec
+epoch 10, total_step 155540, total loss is 11.09 , inference loss is 4.29, weight deacy loss is 6.80, training accuracy is 0.625000, time 56.316 samples/sec
+epoch 10, total_step 155560, total loss is 14.14 , inference loss is 7.34, weight deacy loss is 6.80, training accuracy is 0.437500, time 56.649 samples/sec
+epoch 10, total_step 155580, total loss is 13.53 , inference loss is 6.73, weight deacy loss is 6.80, training accuracy is 0.437500, time 55.302 samples/sec
+epoch 10, total_step 155600, total loss is 12.36 , inference loss is 5.56, weight deacy loss is 6.80, training accuracy is 0.625000, time 56.909 samples/sec
+epoch 10, total_step 155620, total loss is 11.67 , inference loss is 4.87, weight deacy loss is 6.80, training accuracy is 0.562500, time 56.439 samples/sec
+epoch 10, total_step 155640, total loss is 12.15 , inference loss is 5.35, weight deacy loss is 6.80, training accuracy is 0.625000, time 55.979 samples/sec
+epoch 10, total_step 155660, total loss is 11.85 , inference loss is 5.06, weight deacy loss is 6.80, training accuracy is 0.531250, time 57.169 samples/sec
+epoch 10, total_step 155680, total loss is 12.35 , inference loss is 5.55, weight deacy loss is 6.80, training accuracy is 0.500000, time 57.122 samples/sec
+epoch 10, total_step 155700, total loss is 11.84 , inference loss is 5.04, weight deacy loss is 6.80, training accuracy is 0.781250, time 56.555 samples/sec
+epoch 10, total_step 155720, total loss is 15.88 , inference loss is 9.09, weight deacy loss is 6.80, training accuracy is 0.500000, time 56.936 samples/sec
+epoch 10, total_step 155740, total loss is 12.13 , inference loss is 5.33, weight deacy loss is 6.80, training accuracy is 0.625000, time 56.555 samples/sec
+epoch 10, total_step 155760, total loss is 10.77 , inference loss is 3.97, weight deacy loss is 6.80, training accuracy is 0.687500, time 56.289 samples/sec
+epoch 10, total_step 155780, total loss is 10.72 , inference loss is 3.93, weight deacy loss is 6.80, training accuracy is 0.562500, time 56.501 samples/sec
+epoch 10, total_step 155800, total loss is 12.20 , inference loss is 5.41, weight deacy loss is 6.80, training accuracy is 0.593750, time 56.862 samples/sec
+epoch 10, total_step 155820, total loss is 11.75 , inference loss is 4.96, weight deacy loss is 6.80, training accuracy is 0.562500, time 56.198 samples/sec
+epoch 10, total_step 155840, total loss is 10.38 , inference loss is 3.58, weight deacy loss is 6.80, training accuracy is 0.593750, time 56.396 samples/sec
+epoch 10, total_step 155860, total loss is 13.29 , inference loss is 6.50, weight deacy loss is 6.80, training accuracy is 0.625000, time 55.917 samples/sec
+epoch 10, total_step 155880, total loss is 13.56 , inference loss is 6.76, weight deacy loss is 6.80, training accuracy is 0.562500, time 57.057 samples/sec
+epoch 10, total_step 155900, total loss is 13.84 , inference loss is 7.05, weight deacy loss is 6.80, training accuracy is 0.562500, time 56.707 samples/sec
+epoch 10, total_step 155920, total loss is 16.64 , inference loss is 9.85, weight deacy loss is 6.79, training accuracy is 0.625000, time 56.188 samples/sec
+epoch 10, total_step 155940, total loss is 12.16 , inference loss is 5.37, weight deacy loss is 6.79, training accuracy is 0.625000, time 55.355 samples/sec
+epoch 10, total_step 155960, total loss is 12.27 , inference loss is 5.47, weight deacy loss is 6.79, training accuracy is 0.625000, time 56.623 samples/sec
+epoch 10, total_step 155980, total loss is 11.95 , inference loss is 5.15, weight deacy loss is 6.79, training accuracy is 0.625000, time 56.817 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.71922800000009
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 116 0.987037037037037
+best_threshold_index 116 0.9861111111111112
+best_threshold_index 115 0.9875925925925926
+best_threshold_index 115 0.987037037037037
+best_threshold_index 117 0.9859259259259259
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9859259259259259
+best_threshold_index 117 0.9859259259259259
+best_threshold_index 115 0.9853703703703703
+[lfw][156000]XNorm: 20.861145
+[lfw][156000]Accuracy-Flip: 0.98583+-0.00684
+testing verification..
+(14000, 512)
+infer time 84.79597799999995
+best_threshold_index 121 0.984920634920635
+best_threshold_index 117 0.9841269841269841
+best_threshold_index 120 0.9839682539682539
+best_threshold_index 117 0.9836507936507937
+best_threshold_index 117 0.9834920634920635
+best_threshold_index 117 0.9833333333333333
+best_threshold_index 117 0.9846031746031746
+best_threshold_index 120 0.9836507936507937
+best_threshold_index 120 0.9831746031746031
+best_threshold_index 117 0.9847619047619047
+[cfp_ff][156000]XNorm: 19.791688
+[cfp_ff][156000]Accuracy-Flip: 0.98214+-0.00554
+testing verification..
+(14000, 512)
+infer time 85.52437199999983
+best_threshold_index 140 0.9222222222222223
+best_threshold_index 143 0.9219047619047619
+best_threshold_index 143 0.9223809523809524
+best_threshold_index 140 0.9234920634920635
+best_threshold_index 143 0.9242857142857143
+best_threshold_index 143 0.9207936507936508
+best_threshold_index 143 0.9201587301587302
+best_threshold_index 140 0.9198412698412698
+best_threshold_index 140 0.9214285714285714
+best_threshold_index 143 0.920952380952381
+[cfp_fp][156000]XNorm: 20.718280
+[cfp_fp][156000]Accuracy-Flip: 0.92029+-0.01214
+testing verification..
+(12000, 512)
+infer time 73.36534900000002
+best_threshold_index 143 0.8961111111111111
+best_threshold_index 146 0.8966666666666666
+best_threshold_index 147 0.8959259259259259
+best_threshold_index 146 0.8962962962962963
+best_threshold_index 143 0.9022222222222223
+best_threshold_index 144 0.8964814814814814
+best_threshold_index 141 0.8974074074074074
+best_threshold_index 147 0.9012962962962963
+best_threshold_index 145 0.9005555555555556
+best_threshold_index 141 0.8962962962962963
+[agedb_30][156000]XNorm: 21.921795
+[agedb_30][156000]Accuracy-Flip: 0.89350+-0.01998
+test accuracy is: 0.9858333333333332
+epoch 10, total_step 156000, total loss is 13.62 , inference loss is 6.83, weight deacy loss is 6.79, training accuracy is 0.593750, time 54.111 samples/sec
+epoch 10, total_step 156020, total loss is 10.06 , inference loss is 3.27, weight deacy loss is 6.79, training accuracy is 0.625000, time 56.302 samples/sec
+epoch 10, total_step 156040, total loss is 11.48 , inference loss is 4.69, weight deacy loss is 6.79, training accuracy is 0.531250, time 56.452 samples/sec
+epoch 10, total_step 156060, total loss is 12.95 , inference loss is 6.16, weight deacy loss is 6.79, training accuracy is 0.593750, time 55.921 samples/sec
+epoch 10, total_step 156080, total loss is 9.92 , inference loss is 3.13, weight deacy loss is 6.79, training accuracy is 0.656250, time 56.164 samples/sec
+epoch 10, total_step 156100, total loss is 16.86 , inference loss is 10.07, weight deacy loss is 6.79, training accuracy is 0.375000, time 56.881 samples/sec
+epoch 10, total_step 156120, total loss is 12.29 , inference loss is 5.50, weight deacy loss is 6.79, training accuracy is 0.593750, time 56.327 samples/sec
+epoch 10, total_step 156140, total loss is 13.44 , inference loss is 6.65, weight deacy loss is 6.79, training accuracy is 0.593750, time 57.040 samples/sec
+epoch 10, total_step 156160, total loss is 10.57 , inference loss is 3.77, weight deacy loss is 6.79, training accuracy is 0.562500, time 56.894 samples/sec
+epoch 10, total_step 156180, total loss is 12.72 , inference loss is 5.93, weight deacy loss is 6.79, training accuracy is 0.562500, time 56.794 samples/sec
+epoch 10, total_step 156200, total loss is 11.03 , inference loss is 4.24, weight deacy loss is 6.79, training accuracy is 0.687500, time 56.436 samples/sec
+epoch 10, total_step 156220, total loss is 8.26 , inference loss is 1.47, weight deacy loss is 6.79, training accuracy is 0.781250, time 57.422 samples/sec
+epoch 10, total_step 156240, total loss is 13.22 , inference loss is 6.43, weight deacy loss is 6.79, training accuracy is 0.500000, time 56.524 samples/sec
+epoch 10, total_step 156260, total loss is 12.16 , inference loss is 5.37, weight deacy loss is 6.79, training accuracy is 0.625000, time 56.740 samples/sec
+epoch 10, total_step 156280, total loss is 13.59 , inference loss is 6.80, weight deacy loss is 6.79, training accuracy is 0.593750, time 55.676 samples/sec
+epoch 10, total_step 156300, total loss is 10.17 , inference loss is 3.38, weight deacy loss is 6.79, training accuracy is 0.593750, time 56.696 samples/sec
+epoch 10, total_step 156320, total loss is 14.85 , inference loss is 8.06, weight deacy loss is 6.79, training accuracy is 0.500000, time 56.604 samples/sec
+epoch 10, total_step 156340, total loss is 10.31 , inference loss is 3.52, weight deacy loss is 6.79, training accuracy is 0.718750, time 56.683 samples/sec
+epoch 10, total_step 156360, total loss is 16.83 , inference loss is 10.04, weight deacy loss is 6.79, training accuracy is 0.531250, time 56.778 samples/sec
+epoch 10, total_step 156380, total loss is 10.54 , inference loss is 3.75, weight deacy loss is 6.79, training accuracy is 0.687500, time 56.863 samples/sec
+epoch 10, total_step 156400, total loss is 12.07 , inference loss is 5.29, weight deacy loss is 6.79, training accuracy is 0.468750, time 56.099 samples/sec
+epoch 10, total_step 156420, total loss is 10.87 , inference loss is 4.08, weight deacy loss is 6.79, training accuracy is 0.687500, time 56.233 samples/sec
+epoch 10, total_step 156440, total loss is 12.25 , inference loss is 5.47, weight deacy loss is 6.79, training accuracy is 0.531250, time 56.935 samples/sec
+epoch 10, total_step 156460, total loss is 16.48 , inference loss is 9.70, weight deacy loss is 6.79, training accuracy is 0.468750, time 56.356 samples/sec
+epoch 10, total_step 156480, total loss is 12.31 , inference loss is 5.53, weight deacy loss is 6.79, training accuracy is 0.531250, time 56.633 samples/sec
+epoch 10, total_step 156500, total loss is 9.62 , inference loss is 2.83, weight deacy loss is 6.79, training accuracy is 0.656250, time 55.657 samples/sec
+epoch 10, total_step 156520, total loss is 13.92 , inference loss is 7.13, weight deacy loss is 6.79, training accuracy is 0.500000, time 57.152 samples/sec
+epoch 10, total_step 156540, total loss is 15.64 , inference loss is 8.85, weight deacy loss is 6.79, training accuracy is 0.468750, time 56.715 samples/sec
+epoch 10, total_step 156560, total loss is 13.46 , inference loss is 6.68, weight deacy loss is 6.79, training accuracy is 0.468750, time 56.267 samples/sec
+epoch 10, total_step 156580, total loss is 10.61 , inference loss is 3.83, weight deacy loss is 6.79, training accuracy is 0.593750, time 56.656 samples/sec
+epoch 10, total_step 156600, total loss is 11.74 , inference loss is 4.95, weight deacy loss is 6.79, training accuracy is 0.468750, time 56.368 samples/sec
+epoch 10, total_step 156620, total loss is 14.68 , inference loss is 7.90, weight deacy loss is 6.79, training accuracy is 0.468750, time 56.792 samples/sec
+epoch 10, total_step 156640, total loss is 15.31 , inference loss is 8.53, weight deacy loss is 6.79, training accuracy is 0.343750, time 56.358 samples/sec
+epoch 10, total_step 156660, total loss is 14.53 , inference loss is 7.75, weight deacy loss is 6.79, training accuracy is 0.375000, time 56.917 samples/sec
+epoch 10, total_step 156680, total loss is 15.43 , inference loss is 8.64, weight deacy loss is 6.79, training accuracy is 0.531250, time 56.415 samples/sec
+epoch 10, total_step 156700, total loss is 12.75 , inference loss is 5.96, weight deacy loss is 6.79, training accuracy is 0.406250, time 56.493 samples/sec
+epoch 10, total_step 156720, total loss is 11.63 , inference loss is 4.85, weight deacy loss is 6.78, training accuracy is 0.656250, time 56.751 samples/sec
+epoch 10, total_step 156740, total loss is 14.24 , inference loss is 7.46, weight deacy loss is 6.78, training accuracy is 0.468750, time 56.666 samples/sec
+epoch 10, total_step 156760, total loss is 13.16 , inference loss is 6.38, weight deacy loss is 6.78, training accuracy is 0.468750, time 56.782 samples/sec
+epoch 10, total_step 156780, total loss is 11.21 , inference loss is 4.43, weight deacy loss is 6.78, training accuracy is 0.500000, time 56.382 samples/sec
+epoch 10, total_step 156800, total loss is 17.77 , inference loss is 10.98, weight deacy loss is 6.78, training accuracy is 0.343750, time 56.073 samples/sec
+epoch 10, total_step 156820, total loss is 15.43 , inference loss is 8.65, weight deacy loss is 6.78, training accuracy is 0.562500, time 56.731 samples/sec
+epoch 10, total_step 156840, total loss is 12.35 , inference loss is 5.57, weight deacy loss is 6.78, training accuracy is 0.531250, time 55.979 samples/sec
+epoch 10, total_step 156860, total loss is 13.73 , inference loss is 6.95, weight deacy loss is 6.78, training accuracy is 0.406250, time 56.457 samples/sec
+epoch 10, total_step 156880, total loss is 16.74 , inference loss is 9.95, weight deacy loss is 6.78, training accuracy is 0.406250, time 57.421 samples/sec
+epoch 10, total_step 156900, total loss is 14.63 , inference loss is 7.85, weight deacy loss is 6.78, training accuracy is 0.531250, time 56.217 samples/sec
+epoch 10, total_step 156920, total loss is 16.23 , inference loss is 9.45, weight deacy loss is 6.78, training accuracy is 0.406250, time 56.624 samples/sec
+epoch 10, total_step 156940, total loss is 13.30 , inference loss is 6.52, weight deacy loss is 6.78, training accuracy is 0.562500, time 56.816 samples/sec
+epoch 10, total_step 156960, total loss is 15.01 , inference loss is 8.22, weight deacy loss is 6.78, training accuracy is 0.500000, time 56.186 samples/sec
+epoch 10, total_step 156980, total loss is 14.17 , inference loss is 7.39, weight deacy loss is 6.78, training accuracy is 0.250000, time 56.390 samples/sec
+epoch 10, total_step 157000, total loss is 13.92 , inference loss is 7.14, weight deacy loss is 6.78, training accuracy is 0.375000, time 56.651 samples/sec
+epoch 10, total_step 157020, total loss is 21.30 , inference loss is 14.52, weight deacy loss is 6.78, training accuracy is 0.312500, time 56.305 samples/sec
+epoch 10, total_step 157040, total loss is 20.47 , inference loss is 13.69, weight deacy loss is 6.78, training accuracy is 0.406250, time 57.176 samples/sec
+epoch 10, total_step 157060, total loss is 20.82 , inference loss is 14.03, weight deacy loss is 6.78, training accuracy is 0.312500, time 56.846 samples/sec
+epoch 10, total_step 157080, total loss is 15.63 , inference loss is 8.85, weight deacy loss is 6.78, training accuracy is 0.531250, time 56.841 samples/sec
+epoch 10, total_step 157100, total loss is 12.62 , inference loss is 5.83, weight deacy loss is 6.78, training accuracy is 0.500000, time 56.584 samples/sec
+epoch 10, total_step 157120, total loss is 14.56 , inference loss is 7.78, weight deacy loss is 6.78, training accuracy is 0.531250, time 56.167 samples/sec
+epoch 10, total_step 157140, total loss is 15.87 , inference loss is 9.09, weight deacy loss is 6.78, training accuracy is 0.500000, time 56.584 samples/sec
+epoch 10, total_step 157160, total loss is 12.39 , inference loss is 5.61, weight deacy loss is 6.78, training accuracy is 0.531250, time 55.818 samples/sec
+epoch 10, total_step 157180, total loss is 12.51 , inference loss is 5.72, weight deacy loss is 6.78, training accuracy is 0.500000, time 57.521 samples/sec
+epoch 10, total_step 157200, total loss is 15.35 , inference loss is 8.57, weight deacy loss is 6.78, training accuracy is 0.437500, time 56.646 samples/sec
+epoch 10, total_step 157220, total loss is 15.78 , inference loss is 9.00, weight deacy loss is 6.78, training accuracy is 0.312500, time 56.156 samples/sec
+epoch 10, total_step 157240, total loss is 16.46 , inference loss is 9.68, weight deacy loss is 6.78, training accuracy is 0.468750, time 55.949 samples/sec
+epoch 10, total_step 157260, total loss is 11.93 , inference loss is 5.15, weight deacy loss is 6.78, training accuracy is 0.593750, time 56.507 samples/sec
+epoch 10, total_step 157280, total loss is 15.90 , inference loss is 9.12, weight deacy loss is 6.78, training accuracy is 0.375000, time 56.429 samples/sec
+epoch 10, total_step 157300, total loss is 18.44 , inference loss is 11.66, weight deacy loss is 6.78, training accuracy is 0.437500, time 57.722 samples/sec
+epoch 10, total_step 157320, total loss is 16.75 , inference loss is 9.97, weight deacy loss is 6.78, training accuracy is 0.281250, time 57.531 samples/sec
+epoch 10, total_step 157340, total loss is 16.11 , inference loss is 9.34, weight deacy loss is 6.78, training accuracy is 0.437500, time 56.584 samples/sec
+epoch 10, total_step 157360, total loss is 20.63 , inference loss is 13.85, weight deacy loss is 6.78, training accuracy is 0.375000, time 56.386 samples/sec
+epoch 10, total_step 157380, total loss is 12.46 , inference loss is 5.68, weight deacy loss is 6.78, training accuracy is 0.531250, time 56.747 samples/sec
+epoch 10, total_step 157400, total loss is 12.86 , inference loss is 6.08, weight deacy loss is 6.78, training accuracy is 0.562500, time 56.548 samples/sec
+epoch 10, total_step 157420, total loss is 19.39 , inference loss is 12.61, weight deacy loss is 6.78, training accuracy is 0.375000, time 56.190 samples/sec
+epoch 10, total_step 157440, total loss is 12.71 , inference loss is 5.93, weight deacy loss is 6.78, training accuracy is 0.468750, time 55.882 samples/sec
+epoch 10, total_step 157460, total loss is 12.35 , inference loss is 5.57, weight deacy loss is 6.78, training accuracy is 0.531250, time 56.528 samples/sec
+epoch 10, total_step 157480, total loss is 14.60 , inference loss is 7.82, weight deacy loss is 6.78, training accuracy is 0.625000, time 56.699 samples/sec
+epoch 10, total_step 157500, total loss is 16.98 , inference loss is 10.20, weight deacy loss is 6.78, training accuracy is 0.406250, time 56.415 samples/sec
+epoch 10, total_step 157520, total loss is 13.97 , inference loss is 7.19, weight deacy loss is 6.78, training accuracy is 0.406250, time 56.463 samples/sec
+epoch 10, total_step 157540, total loss is 11.23 , inference loss is 4.45, weight deacy loss is 6.78, training accuracy is 0.562500, time 56.532 samples/sec
+epoch 10, total_step 157560, total loss is 13.45 , inference loss is 6.68, weight deacy loss is 6.78, training accuracy is 0.593750, time 56.422 samples/sec
+epoch 10, total_step 157580, total loss is 15.72 , inference loss is 8.94, weight deacy loss is 6.78, training accuracy is 0.437500, time 56.673 samples/sec
+epoch 10, total_step 157600, total loss is 13.48 , inference loss is 6.70, weight deacy loss is 6.78, training accuracy is 0.437500, time 56.118 samples/sec
+epoch 10, total_step 157620, total loss is 15.73 , inference loss is 8.95, weight deacy loss is 6.78, training accuracy is 0.437500, time 56.487 samples/sec
+epoch 10, total_step 157640, total loss is 19.70 , inference loss is 12.93, weight deacy loss is 6.78, training accuracy is 0.250000, time 56.054 samples/sec
+epoch 10, total_step 157660, total loss is 17.88 , inference loss is 11.11, weight deacy loss is 6.78, training accuracy is 0.500000, time 56.837 samples/sec
+epoch 10, total_step 157680, total loss is 17.04 , inference loss is 10.27, weight deacy loss is 6.78, training accuracy is 0.343750, time 56.602 samples/sec
+epoch 10, total_step 157700, total loss is 16.12 , inference loss is 9.35, weight deacy loss is 6.78, training accuracy is 0.343750, time 56.496 samples/sec
+epoch 10, total_step 157720, total loss is 12.37 , inference loss is 5.59, weight deacy loss is 6.78, training accuracy is 0.500000, time 56.709 samples/sec
+epoch 10, total_step 157740, total loss is 15.21 , inference loss is 8.44, weight deacy loss is 6.77, training accuracy is 0.375000, time 56.563 samples/sec
+epoch 10, total_step 157760, total loss is 14.59 , inference loss is 7.82, weight deacy loss is 6.77, training accuracy is 0.406250, time 57.065 samples/sec
+epoch 10, total_step 157780, total loss is 16.33 , inference loss is 9.56, weight deacy loss is 6.77, training accuracy is 0.343750, time 56.106 samples/sec
+epoch 10, total_step 157800, total loss is 15.75 , inference loss is 8.98, weight deacy loss is 6.77, training accuracy is 0.218750, time 56.292 samples/sec
+epoch 10, total_step 157820, total loss is 18.42 , inference loss is 11.64, weight deacy loss is 6.77, training accuracy is 0.468750, time 56.530 samples/sec
+epoch 10, total_step 157840, total loss is 13.03 , inference loss is 6.25, weight deacy loss is 6.77, training accuracy is 0.531250, time 57.247 samples/sec
+epoch 10, total_step 157860, total loss is 13.74 , inference loss is 6.97, weight deacy loss is 6.77, training accuracy is 0.468750, time 57.566 samples/sec
+epoch 10, total_step 157880, total loss is 15.94 , inference loss is 9.17, weight deacy loss is 6.77, training accuracy is 0.281250, time 56.027 samples/sec
+epoch 10, total_step 157900, total loss is 12.82 , inference loss is 6.04, weight deacy loss is 6.77, training accuracy is 0.593750, time 55.650 samples/sec
+epoch 10, total_step 157920, total loss is 11.47 , inference loss is 4.69, weight deacy loss is 6.77, training accuracy is 0.531250, time 56.420 samples/sec
+epoch 10, total_step 157940, total loss is 15.38 , inference loss is 8.60, weight deacy loss is 6.77, training accuracy is 0.437500, time 56.271 samples/sec
+epoch 10, total_step 157960, total loss is 16.42 , inference loss is 9.65, weight deacy loss is 6.77, training accuracy is 0.375000, time 56.988 samples/sec
+epoch 10, total_step 157980, total loss is 13.57 , inference loss is 6.80, weight deacy loss is 6.77, training accuracy is 0.375000, time 56.663 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.46802299999996
+best_threshold_index 115 0.9879629629629629
+best_threshold_index 116 0.9874074074074074
+best_threshold_index 116 0.9868518518518519
+best_threshold_index 115 0.9872222222222222
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 116 0.9862962962962963
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 112 0.9868518518518519
+best_threshold_index 115 0.9862962962962963
+best_threshold_index 115 0.9861111111111112
+[lfw][158000]XNorm: 18.552655
+[lfw][158000]Accuracy-Flip: 0.98617+-0.00578
+testing verification..
+(14000, 512)
+infer time 86.20467799999994
+best_threshold_index 120 0.9879365079365079
+best_threshold_index 120 0.9866666666666667
+best_threshold_index 120 0.9873015873015873
+best_threshold_index 120 0.9873015873015873
+best_threshold_index 120 0.9879365079365079
+best_threshold_index 120 0.986984126984127
+best_threshold_index 120 0.9876190476190476
+best_threshold_index 120 0.9861904761904762
+best_threshold_index 120 0.9863492063492063
+best_threshold_index 120 0.9885714285714285
+[cfp_ff][158000]XNorm: 17.980905
+[cfp_ff][158000]Accuracy-Flip: 0.98729+-0.00647
+testing verification..
+(14000, 512)
+infer time 85.67096999999988
+best_threshold_index 134 0.9223809523809524
+best_threshold_index 136 0.9230158730158731
+best_threshold_index 136 0.922063492063492
+best_threshold_index 136 0.9244444444444444
+best_threshold_index 136 0.9255555555555556
+best_threshold_index 136 0.9207936507936508
+best_threshold_index 134 0.9212698412698412
+best_threshold_index 136 0.9223809523809524
+best_threshold_index 136 0.9233333333333333
+best_threshold_index 136 0.9236507936507936
+[cfp_fp][158000]XNorm: 19.151126
+[cfp_fp][158000]Accuracy-Flip: 0.92171+-0.01169
+testing verification..
+(12000, 512)
+infer time 73.83162200000002
+best_threshold_index 139 0.8992592592592593
+best_threshold_index 139 0.8966666666666666
+best_threshold_index 139 0.899074074074074
+best_threshold_index 139 0.8994444444444445
+best_threshold_index 139 0.9031481481481481
+best_threshold_index 139 0.8968518518518519
+best_threshold_index 139 0.9007407407407407
+best_threshold_index 139 0.9061111111111111
+best_threshold_index 139 0.9024074074074074
+best_threshold_index 139 0.902962962962963
+[agedb_30][158000]XNorm: 20.224172
+[agedb_30][158000]Accuracy-Flip: 0.90067+-0.02556
+test accuracy is: 0.9861666666666666
+epoch 10, total_step 158000, total loss is 10.61 , inference loss is 3.84, weight deacy loss is 6.77, training accuracy is 0.531250, time 54.272 samples/sec
+epoch 10, total_step 158020, total loss is 17.03 , inference loss is 10.26, weight deacy loss is 6.77, training accuracy is 0.218750, time 56.569 samples/sec
+epoch 10, total_step 158040, total loss is 11.67 , inference loss is 4.89, weight deacy loss is 6.77, training accuracy is 0.625000, time 57.661 samples/sec
+epoch 10, total_step 158060, total loss is 18.60 , inference loss is 11.83, weight deacy loss is 6.77, training accuracy is 0.406250, time 57.395 samples/sec
+epoch 10, total_step 158080, total loss is 15.93 , inference loss is 9.16, weight deacy loss is 6.77, training accuracy is 0.437500, time 56.625 samples/sec
+epoch 10, total_step 158100, total loss is 15.62 , inference loss is 8.85, weight deacy loss is 6.77, training accuracy is 0.281250, time 56.390 samples/sec
+epoch 10, total_step 158120, total loss is 17.12 , inference loss is 10.35, weight deacy loss is 6.77, training accuracy is 0.437500, time 56.582 samples/sec
+epoch 10, total_step 158140, total loss is 14.75 , inference loss is 7.98, weight deacy loss is 6.77, training accuracy is 0.468750, time 56.287 samples/sec
+epoch 10, total_step 158160, total loss is 14.67 , inference loss is 7.90, weight deacy loss is 6.77, training accuracy is 0.437500, time 56.760 samples/sec
+epoch 10, total_step 158180, total loss is 13.34 , inference loss is 6.57, weight deacy loss is 6.77, training accuracy is 0.500000, time 56.273 samples/sec
+epoch 10, total_step 158200, total loss is 14.73 , inference loss is 7.96, weight deacy loss is 6.77, training accuracy is 0.406250, time 56.925 samples/sec
+epoch 10, total_step 158220, total loss is 14.69 , inference loss is 7.92, weight deacy loss is 6.77, training accuracy is 0.468750, time 57.069 samples/sec
+epoch 10, total_step 158240, total loss is 13.79 , inference loss is 7.02, weight deacy loss is 6.77, training accuracy is 0.468750, time 55.506 samples/sec
+epoch 10, total_step 158260, total loss is 20.26 , inference loss is 13.49, weight deacy loss is 6.77, training accuracy is 0.406250, time 57.053 samples/sec
+epoch 10, total_step 158280, total loss is 21.16 , inference loss is 14.39, weight deacy loss is 6.77, training accuracy is 0.312500, time 57.195 samples/sec
+epoch 10, total_step 158300, total loss is 14.57 , inference loss is 7.80, weight deacy loss is 6.77, training accuracy is 0.531250, time 56.713 samples/sec
+epoch 10, total_step 158320, total loss is 18.87 , inference loss is 12.10, weight deacy loss is 6.77, training accuracy is 0.437500, time 56.452 samples/sec
+epoch 10, total_step 158340, total loss is 15.62 , inference loss is 8.85, weight deacy loss is 6.77, training accuracy is 0.437500, time 57.065 samples/sec
+epoch 10, total_step 158360, total loss is 14.00 , inference loss is 7.23, weight deacy loss is 6.77, training accuracy is 0.437500, time 57.041 samples/sec
+epoch 10, total_step 158380, total loss is 17.43 , inference loss is 10.67, weight deacy loss is 6.77, training accuracy is 0.281250, time 55.950 samples/sec
+epoch 10, total_step 158400, total loss is 12.37 , inference loss is 5.60, weight deacy loss is 6.77, training accuracy is 0.468750, time 56.885 samples/sec
+epoch 10, total_step 158420, total loss is 13.57 , inference loss is 6.81, weight deacy loss is 6.77, training accuracy is 0.406250, time 56.397 samples/sec
+epoch 10, total_step 158440, total loss is 14.33 , inference loss is 7.57, weight deacy loss is 6.77, training accuracy is 0.406250, time 56.697 samples/sec
+epoch 10, total_step 158460, total loss is 17.32 , inference loss is 10.56, weight deacy loss is 6.77, training accuracy is 0.500000, time 56.194 samples/sec
+epoch 10, total_step 158480, total loss is 18.02 , inference loss is 11.25, weight deacy loss is 6.77, training accuracy is 0.437500, time 56.345 samples/sec
+epoch 10, total_step 158500, total loss is 14.76 , inference loss is 8.00, weight deacy loss is 6.77, training accuracy is 0.437500, time 56.530 samples/sec
+epoch 10, total_step 158520, total loss is 12.47 , inference loss is 5.70, weight deacy loss is 6.77, training accuracy is 0.437500, time 56.351 samples/sec
+epoch 10, total_step 158540, total loss is 17.00 , inference loss is 10.23, weight deacy loss is 6.77, training accuracy is 0.375000, time 56.757 samples/sec
+epoch 10, total_step 158560, total loss is 14.10 , inference loss is 7.33, weight deacy loss is 6.77, training accuracy is 0.687500, time 56.897 samples/sec
+epoch 10, total_step 158580, total loss is 13.08 , inference loss is 6.32, weight deacy loss is 6.77, training accuracy is 0.562500, time 56.753 samples/sec
+epoch 10, total_step 158600, total loss is 10.24 , inference loss is 3.47, weight deacy loss is 6.77, training accuracy is 0.562500, time 55.791 samples/sec
+epoch 10, total_step 158620, total loss is 13.35 , inference loss is 6.58, weight deacy loss is 6.76, training accuracy is 0.500000, time 56.779 samples/sec
+epoch 10, total_step 158640, total loss is 16.33 , inference loss is 9.57, weight deacy loss is 6.76, training accuracy is 0.406250, time 56.538 samples/sec
+epoch 10, total_step 158660, total loss is 18.32 , inference loss is 11.56, weight deacy loss is 6.76, training accuracy is 0.375000, time 56.672 samples/sec
+epoch 10, total_step 158680, total loss is 12.12 , inference loss is 5.35, weight deacy loss is 6.76, training accuracy is 0.468750, time 56.790 samples/sec
+epoch 10, total_step 158700, total loss is 15.29 , inference loss is 8.53, weight deacy loss is 6.76, training accuracy is 0.468750, time 56.246 samples/sec
+epoch 10, total_step 158720, total loss is 13.77 , inference loss is 7.01, weight deacy loss is 6.76, training accuracy is 0.437500, time 56.526 samples/sec
+epoch 10, total_step 158740, total loss is 15.60 , inference loss is 8.84, weight deacy loss is 6.76, training accuracy is 0.437500, time 57.009 samples/sec
+epoch 10, total_step 158760, total loss is 15.09 , inference loss is 8.33, weight deacy loss is 6.76, training accuracy is 0.406250, time 56.862 samples/sec
+epoch 10, total_step 158780, total loss is 14.08 , inference loss is 7.32, weight deacy loss is 6.76, training accuracy is 0.500000, time 57.231 samples/sec
+epoch 10, total_step 158800, total loss is 14.97 , inference loss is 8.21, weight deacy loss is 6.76, training accuracy is 0.343750, time 56.618 samples/sec
+epoch 10, total_step 158820, total loss is 15.21 , inference loss is 8.45, weight deacy loss is 6.76, training accuracy is 0.531250, time 56.515 samples/sec
+epoch 10, total_step 158840, total loss is 18.03 , inference loss is 11.27, weight deacy loss is 6.76, training accuracy is 0.343750, time 56.614 samples/sec
+epoch 10, total_step 158860, total loss is 17.89 , inference loss is 11.13, weight deacy loss is 6.76, training accuracy is 0.375000, time 56.169 samples/sec
+epoch 10, total_step 158880, total loss is 11.42 , inference loss is 4.66, weight deacy loss is 6.76, training accuracy is 0.437500, time 56.685 samples/sec
+epoch 10, total_step 158900, total loss is 18.12 , inference loss is 11.36, weight deacy loss is 6.76, training accuracy is 0.281250, time 56.036 samples/sec
+epoch 10, total_step 158920, total loss is 18.46 , inference loss is 11.69, weight deacy loss is 6.76, training accuracy is 0.312500, time 56.088 samples/sec
+epoch 10, total_step 158940, total loss is 13.39 , inference loss is 6.63, weight deacy loss is 6.76, training accuracy is 0.468750, time 55.717 samples/sec
+epoch 10, total_step 158960, total loss is 14.02 , inference loss is 7.26, weight deacy loss is 6.76, training accuracy is 0.437500, time 56.478 samples/sec
+epoch 10, total_step 158980, total loss is 13.41 , inference loss is 6.65, weight deacy loss is 6.76, training accuracy is 0.468750, time 56.780 samples/sec
+epoch 10, total_step 159000, total loss is 13.24 , inference loss is 6.48, weight deacy loss is 6.76, training accuracy is 0.468750, time 56.354 samples/sec
+epoch 10, total_step 159020, total loss is 14.86 , inference loss is 8.10, weight deacy loss is 6.76, training accuracy is 0.468750, time 56.179 samples/sec
+epoch 10, total_step 159040, total loss is 12.12 , inference loss is 5.36, weight deacy loss is 6.76, training accuracy is 0.531250, time 56.264 samples/sec
+epoch 10, total_step 159060, total loss is 13.41 , inference loss is 6.66, weight deacy loss is 6.76, training accuracy is 0.437500, time 56.045 samples/sec
+epoch 10, total_step 159080, total loss is 14.62 , inference loss is 7.86, weight deacy loss is 6.76, training accuracy is 0.562500, time 56.735 samples/sec
+epoch 10, total_step 159100, total loss is 12.33 , inference loss is 5.57, weight deacy loss is 6.76, training accuracy is 0.593750, time 56.508 samples/sec
+epoch 10, total_step 159120, total loss is 18.55 , inference loss is 11.80, weight deacy loss is 6.76, training accuracy is 0.218750, time 56.864 samples/sec
+epoch 10, total_step 159140, total loss is 14.35 , inference loss is 7.59, weight deacy loss is 6.76, training accuracy is 0.437500, time 56.188 samples/sec
+epoch 10, total_step 159160, total loss is 14.43 , inference loss is 7.67, weight deacy loss is 6.76, training accuracy is 0.437500, time 56.566 samples/sec
+epoch 10, total_step 159180, total loss is 18.83 , inference loss is 12.07, weight deacy loss is 6.76, training accuracy is 0.437500, time 56.342 samples/sec
+epoch 10, total_step 159200, total loss is 12.75 , inference loss is 5.99, weight deacy loss is 6.76, training accuracy is 0.531250, time 57.222 samples/sec
+epoch 10, total_step 159220, total loss is 11.32 , inference loss is 4.56, weight deacy loss is 6.76, training accuracy is 0.562500, time 55.440 samples/sec
+epoch 10, total_step 159240, total loss is 12.46 , inference loss is 5.71, weight deacy loss is 6.76, training accuracy is 0.531250, time 56.755 samples/sec
+epoch 10, total_step 159260, total loss is 17.14 , inference loss is 10.38, weight deacy loss is 6.76, training accuracy is 0.281250, time 56.106 samples/sec
+epoch 10, total_step 159280, total loss is 14.43 , inference loss is 7.67, weight deacy loss is 6.76, training accuracy is 0.468750, time 56.698 samples/sec
+epoch 10, total_step 159300, total loss is 15.73 , inference loss is 8.98, weight deacy loss is 6.76, training accuracy is 0.593750, time 56.715 samples/sec
+epoch 10, total_step 159320, total loss is 12.47 , inference loss is 5.72, weight deacy loss is 6.76, training accuracy is 0.562500, time 56.062 samples/sec
+epoch 10, total_step 159340, total loss is 12.55 , inference loss is 5.79, weight deacy loss is 6.76, training accuracy is 0.531250, time 54.901 samples/sec
+epoch 10, total_step 159360, total loss is 13.80 , inference loss is 7.04, weight deacy loss is 6.76, training accuracy is 0.500000, time 56.470 samples/sec
+epoch 10, total_step 159380, total loss is 15.59 , inference loss is 8.84, weight deacy loss is 6.76, training accuracy is 0.406250, time 56.170 samples/sec
+epoch 10, total_step 159400, total loss is 13.50 , inference loss is 6.75, weight deacy loss is 6.75, training accuracy is 0.500000, time 56.920 samples/sec
+epoch 10, total_step 159420, total loss is 15.47 , inference loss is 8.72, weight deacy loss is 6.75, training accuracy is 0.437500, time 56.507 samples/sec
+epoch 10, total_step 159440, total loss is 15.14 , inference loss is 8.38, weight deacy loss is 6.75, training accuracy is 0.250000, time 57.317 samples/sec
+epoch 10, total_step 159460, total loss is 14.51 , inference loss is 7.76, weight deacy loss is 6.75, training accuracy is 0.437500, time 57.275 samples/sec
+epoch 10, total_step 159480, total loss is 16.24 , inference loss is 9.48, weight deacy loss is 6.75, training accuracy is 0.406250, time 56.501 samples/sec
+epoch 10, total_step 159500, total loss is 13.10 , inference loss is 6.35, weight deacy loss is 6.75, training accuracy is 0.437500, time 57.082 samples/sec
+epoch 10, total_step 159520, total loss is 14.73 , inference loss is 7.98, weight deacy loss is 6.75, training accuracy is 0.437500, time 58.264 samples/sec
+epoch 10, total_step 159540, total loss is 13.27 , inference loss is 6.52, weight deacy loss is 6.75, training accuracy is 0.468750, time 56.090 samples/sec
+epoch 10, total_step 159560, total loss is 11.44 , inference loss is 4.68, weight deacy loss is 6.75, training accuracy is 0.625000, time 56.452 samples/sec
+epoch 10, total_step 159580, total loss is 17.87 , inference loss is 11.12, weight deacy loss is 6.75, training accuracy is 0.437500, time 56.085 samples/sec
+epoch 10, total_step 159600, total loss is 17.45 , inference loss is 10.70, weight deacy loss is 6.75, training accuracy is 0.500000, time 55.794 samples/sec
+epoch 10, total_step 159620, total loss is 14.32 , inference loss is 7.57, weight deacy loss is 6.75, training accuracy is 0.468750, time 56.594 samples/sec
+epoch 10, total_step 159640, total loss is 10.85 , inference loss is 4.10, weight deacy loss is 6.75, training accuracy is 0.593750, time 56.767 samples/sec
+epoch 10, total_step 159660, total loss is 12.65 , inference loss is 5.90, weight deacy loss is 6.75, training accuracy is 0.437500, time 56.884 samples/sec
+epoch 10, total_step 159680, total loss is 11.52 , inference loss is 4.77, weight deacy loss is 6.75, training accuracy is 0.562500, time 56.057 samples/sec
+epoch 10, total_step 159700, total loss is 12.14 , inference loss is 5.39, weight deacy loss is 6.75, training accuracy is 0.656250, time 55.437 samples/sec
+epoch 10, total_step 159720, total loss is 15.72 , inference loss is 8.97, weight deacy loss is 6.75, training accuracy is 0.437500, time 56.667 samples/sec
+epoch 10, total_step 159740, total loss is 12.65 , inference loss is 5.90, weight deacy loss is 6.75, training accuracy is 0.468750, time 56.622 samples/sec
+epoch 10, total_step 159760, total loss is 12.98 , inference loss is 6.23, weight deacy loss is 6.75, training accuracy is 0.468750, time 56.496 samples/sec
+epoch 10, total_step 159780, total loss is 15.36 , inference loss is 8.61, weight deacy loss is 6.75, training accuracy is 0.437500, time 56.592 samples/sec
+epoch 10, total_step 159800, total loss is 15.45 , inference loss is 8.70, weight deacy loss is 6.75, training accuracy is 0.375000, time 56.074 samples/sec
+epoch 10, total_step 159820, total loss is 13.91 , inference loss is 7.16, weight deacy loss is 6.75, training accuracy is 0.468750, time 56.724 samples/sec
+epoch 10, total_step 159840, total loss is 13.23 , inference loss is 6.48, weight deacy loss is 6.75, training accuracy is 0.625000, time 55.819 samples/sec
+epoch 10, total_step 159860, total loss is 14.19 , inference loss is 7.44, weight deacy loss is 6.75, training accuracy is 0.468750, time 55.789 samples/sec
+epoch 10, total_step 159880, total loss is 12.24 , inference loss is 5.49, weight deacy loss is 6.75, training accuracy is 0.687500, time 56.206 samples/sec
+epoch 10, total_step 159900, total loss is 18.75 , inference loss is 12.00, weight deacy loss is 6.75, training accuracy is 0.218750, time 56.066 samples/sec
+epoch 10, total_step 159920, total loss is 12.59 , inference loss is 5.84, weight deacy loss is 6.75, training accuracy is 0.531250, time 55.573 samples/sec
+epoch 10, total_step 159940, total loss is 13.32 , inference loss is 6.58, weight deacy loss is 6.75, training accuracy is 0.437500, time 56.864 samples/sec
+epoch 10, total_step 159960, total loss is 13.76 , inference loss is 7.01, weight deacy loss is 6.75, training accuracy is 0.437500, time 55.815 samples/sec
+epoch 10, total_step 159980, total loss is 12.83 , inference loss is 6.09, weight deacy loss is 6.75, training accuracy is 0.437500, time 55.555 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.33337600000006
+best_threshold_index 112 0.987037037037037
+best_threshold_index 114 0.9855555555555555
+best_threshold_index 112 0.9859259259259259
+best_threshold_index 112 0.9864814814814815
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 112 0.9851851851851852
+best_threshold_index 112 0.987037037037037
+best_threshold_index 112 0.9862962962962963
+best_threshold_index 114 0.9853703703703703
+best_threshold_index 114 0.9851851851851852
+[lfw][160000]XNorm: 22.359023
+[lfw][160000]Accuracy-Flip: 0.98517+-0.00639
+testing verification..
+(14000, 512)
+infer time 84.87002399999994
+best_threshold_index 112 0.9888888888888889
+best_threshold_index 113 0.9874603174603175
+best_threshold_index 114 0.9879365079365079
+best_threshold_index 113 0.9876190476190476
+best_threshold_index 115 0.9885714285714285
+best_threshold_index 113 0.9876190476190476
+best_threshold_index 113 0.9884126984126984
+best_threshold_index 112 0.9873015873015873
+best_threshold_index 113 0.9879365079365079
+best_threshold_index 113 0.9885714285714285
+[cfp_ff][160000]XNorm: 21.873670
+[cfp_ff][160000]Accuracy-Flip: 0.98714+-0.00487
+testing verification..
+(14000, 512)
+infer time 85.618553
+best_threshold_index 132 0.9246031746031746
+best_threshold_index 133 0.9242857142857143
+best_threshold_index 135 0.9223809523809524
+best_threshold_index 135 0.9250793650793651
+best_threshold_index 133 0.9274603174603174
+best_threshold_index 134 0.9206349206349206
+best_threshold_index 135 0.9219047619047619
+best_threshold_index 134 0.9234920634920635
+best_threshold_index 134 0.9238095238095239
+best_threshold_index 134 0.9242857142857143
+[cfp_fp][160000]XNorm: 22.678964
+[cfp_fp][160000]Accuracy-Flip: 0.92243+-0.01655
+testing verification..
+(12000, 512)
+infer time 72.81033400000005
+best_threshold_index 135 0.9061111111111111
+best_threshold_index 135 0.9046296296296297
+best_threshold_index 135 0.9087037037037037
+best_threshold_index 134 0.9094444444444445
+best_threshold_index 134 0.9142592592592592
+best_threshold_index 134 0.9062962962962963
+best_threshold_index 135 0.9094444444444445
+best_threshold_index 135 0.9133333333333333
+best_threshold_index 134 0.9103703703703704
+best_threshold_index 133 0.9098148148148149
+[agedb_30][160000]XNorm: 23.116582
+[agedb_30][160000]Accuracy-Flip: 0.90750+-0.02631
+test accuracy is: 0.9851666666666669
+epoch 10, total_step 160000, total loss is 12.81 , inference loss is 6.06, weight deacy loss is 6.75, training accuracy is 0.562500, time 53.910 samples/sec
+epoch 10, total_step 160020, total loss is 16.55 , inference loss is 9.80, weight deacy loss is 6.75, training accuracy is 0.437500, time 56.862 samples/sec
+epoch 10, total_step 160040, total loss is 13.68 , inference loss is 6.93, weight deacy loss is 6.75, training accuracy is 0.531250, time 56.842 samples/sec
+epoch 10, total_step 160060, total loss is 14.05 , inference loss is 7.31, weight deacy loss is 6.75, training accuracy is 0.500000, time 56.839 samples/sec
+epoch 10, total_step 160080, total loss is 18.98 , inference loss is 12.24, weight deacy loss is 6.75, training accuracy is 0.343750, time 56.728 samples/sec
+epoch 10, total_step 160100, total loss is 14.87 , inference loss is 8.12, weight deacy loss is 6.75, training accuracy is 0.375000, time 55.708 samples/sec
+epoch 10, total_step 160120, total loss is 15.06 , inference loss is 8.32, weight deacy loss is 6.75, training accuracy is 0.406250, time 56.972 samples/sec
+epoch 10, total_step 160140, total loss is 14.90 , inference loss is 8.15, weight deacy loss is 6.75, training accuracy is 0.406250, time 57.124 samples/sec
+epoch 10, total_step 160160, total loss is 10.84 , inference loss is 4.09, weight deacy loss is 6.74, training accuracy is 0.500000, time 56.584 samples/sec
+epoch 10, total_step 160180, total loss is 12.84 , inference loss is 6.10, weight deacy loss is 6.74, training accuracy is 0.468750, time 55.823 samples/sec
+epoch 10, total_step 160200, total loss is 18.30 , inference loss is 11.55, weight deacy loss is 6.74, training accuracy is 0.406250, time 56.421 samples/sec
+epoch 10, total_step 160220, total loss is 13.69 , inference loss is 6.95, weight deacy loss is 6.74, training accuracy is 0.500000, time 56.886 samples/sec
+epoch 10, total_step 160240, total loss is 15.72 , inference loss is 8.98, weight deacy loss is 6.74, training accuracy is 0.343750, time 56.911 samples/sec
+epoch 10, total_step 160260, total loss is 15.96 , inference loss is 9.21, weight deacy loss is 6.74, training accuracy is 0.312500, time 56.960 samples/sec
+epoch 10, total_step 160280, total loss is 13.51 , inference loss is 6.77, weight deacy loss is 6.74, training accuracy is 0.406250, time 56.360 samples/sec
+epoch 10, total_step 160300, total loss is 18.16 , inference loss is 11.41, weight deacy loss is 6.74, training accuracy is 0.250000, time 56.808 samples/sec
+epoch 10, total_step 160320, total loss is 12.50 , inference loss is 5.76, weight deacy loss is 6.74, training accuracy is 0.468750, time 56.691 samples/sec
+epoch 10, total_step 160340, total loss is 13.26 , inference loss is 6.52, weight deacy loss is 6.74, training accuracy is 0.625000, time 56.927 samples/sec
+epoch 10, total_step 160360, total loss is 13.68 , inference loss is 6.94, weight deacy loss is 6.74, training accuracy is 0.437500, time 55.607 samples/sec
+epoch 10, total_step 160380, total loss is 10.77 , inference loss is 4.03, weight deacy loss is 6.74, training accuracy is 0.562500, time 55.989 samples/sec
+epoch 10, total_step 160400, total loss is 14.70 , inference loss is 7.95, weight deacy loss is 6.74, training accuracy is 0.437500, time 56.385 samples/sec
+epoch 10, total_step 160420, total loss is 17.73 , inference loss is 10.98, weight deacy loss is 6.74, training accuracy is 0.437500, time 55.748 samples/sec
+epoch 10, total_step 160440, total loss is 19.05 , inference loss is 12.31, weight deacy loss is 6.74, training accuracy is 0.375000, time 56.546 samples/sec
+epoch 10, total_step 160460, total loss is 17.21 , inference loss is 10.47, weight deacy loss is 6.74, training accuracy is 0.406250, time 56.562 samples/sec
+epoch 10, total_step 160480, total loss is 20.51 , inference loss is 13.77, weight deacy loss is 6.74, training accuracy is 0.187500, time 56.499 samples/sec
+epoch 10, total_step 160500, total loss is 15.21 , inference loss is 8.46, weight deacy loss is 6.74, training accuracy is 0.437500, time 56.239 samples/sec
+epoch 10, total_step 160520, total loss is 13.08 , inference loss is 6.34, weight deacy loss is 6.74, training accuracy is 0.562500, time 56.172 samples/sec
+epoch 10, total_step 160540, total loss is 14.02 , inference loss is 7.28, weight deacy loss is 6.74, training accuracy is 0.500000, time 56.813 samples/sec
+epoch 10, total_step 160560, total loss is 13.47 , inference loss is 6.73, weight deacy loss is 6.74, training accuracy is 0.375000, time 56.508 samples/sec
+epoch 10, total_step 160580, total loss is 13.02 , inference loss is 6.28, weight deacy loss is 6.74, training accuracy is 0.468750, time 56.005 samples/sec
+epoch 10, total_step 160600, total loss is 20.70 , inference loss is 13.96, weight deacy loss is 6.74, training accuracy is 0.156250, time 56.488 samples/sec
+epoch 10, total_step 160620, total loss is 12.74 , inference loss is 6.00, weight deacy loss is 6.74, training accuracy is 0.375000, time 56.152 samples/sec
+epoch 10, total_step 160640, total loss is 14.88 , inference loss is 8.14, weight deacy loss is 6.74, training accuracy is 0.375000, time 55.915 samples/sec
+epoch 10, total_step 160660, total loss is 18.35 , inference loss is 11.61, weight deacy loss is 6.74, training accuracy is 0.437500, time 57.135 samples/sec
+epoch 10, total_step 160680, total loss is 10.98 , inference loss is 4.24, weight deacy loss is 6.74, training accuracy is 0.593750, time 57.614 samples/sec
+epoch 10, total_step 160700, total loss is 16.24 , inference loss is 9.50, weight deacy loss is 6.74, training accuracy is 0.500000, time 56.301 samples/sec
+epoch 10, total_step 160720, total loss is 16.00 , inference loss is 9.26, weight deacy loss is 6.74, training accuracy is 0.531250, time 56.591 samples/sec
+epoch 10, total_step 160740, total loss is 14.28 , inference loss is 7.55, weight deacy loss is 6.74, training accuracy is 0.562500, time 56.707 samples/sec
+epoch 10, total_step 160760, total loss is 17.96 , inference loss is 11.22, weight deacy loss is 6.74, training accuracy is 0.437500, time 56.668 samples/sec
+epoch 10, total_step 160780, total loss is 12.80 , inference loss is 6.07, weight deacy loss is 6.74, training accuracy is 0.468750, time 56.212 samples/sec
+epoch 10, total_step 160800, total loss is 14.98 , inference loss is 8.25, weight deacy loss is 6.74, training accuracy is 0.375000, time 56.659 samples/sec
+epoch 10, total_step 160820, total loss is 18.78 , inference loss is 12.04, weight deacy loss is 6.74, training accuracy is 0.218750, time 56.634 samples/sec
+epoch 10, total_step 160840, total loss is 12.71 , inference loss is 5.97, weight deacy loss is 6.74, training accuracy is 0.625000, time 56.374 samples/sec
+epoch 10, total_step 160860, total loss is 14.10 , inference loss is 7.37, weight deacy loss is 6.74, training accuracy is 0.593750, time 56.465 samples/sec
+epoch 10, total_step 160880, total loss is 14.55 , inference loss is 7.81, weight deacy loss is 6.74, training accuracy is 0.500000, time 56.821 samples/sec
+epoch 10, total_step 160900, total loss is 13.73 , inference loss is 6.99, weight deacy loss is 6.74, training accuracy is 0.343750, time 56.071 samples/sec
+epoch 10, total_step 160920, total loss is 12.44 , inference loss is 5.71, weight deacy loss is 6.74, training accuracy is 0.500000, time 56.293 samples/sec
+epoch 10, total_step 160940, total loss is 16.37 , inference loss is 9.63, weight deacy loss is 6.74, training accuracy is 0.531250, time 56.944 samples/sec
+epoch 10, total_step 160960, total loss is 14.52 , inference loss is 7.79, weight deacy loss is 6.73, training accuracy is 0.406250, time 55.588 samples/sec
+epoch 10, total_step 160980, total loss is 13.46 , inference loss is 6.72, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.476 samples/sec
+epoch 10, total_step 161000, total loss is 13.86 , inference loss is 7.13, weight deacy loss is 6.73, training accuracy is 0.531250, time 56.967 samples/sec
+epoch 10, total_step 161020, total loss is 11.26 , inference loss is 4.53, weight deacy loss is 6.73, training accuracy is 0.468750, time 57.063 samples/sec
+epoch 10, total_step 161040, total loss is 16.77 , inference loss is 10.03, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.215 samples/sec
+epoch 10, total_step 161060, total loss is 19.68 , inference loss is 12.95, weight deacy loss is 6.73, training accuracy is 0.312500, time 56.128 samples/sec
+epoch 10, total_step 161080, total loss is 17.69 , inference loss is 10.96, weight deacy loss is 6.73, training accuracy is 0.218750, time 56.665 samples/sec
+epoch 10, total_step 161100, total loss is 13.87 , inference loss is 7.14, weight deacy loss is 6.73, training accuracy is 0.687500, time 56.498 samples/sec
+epoch 10, total_step 161120, total loss is 16.44 , inference loss is 9.71, weight deacy loss is 6.73, training accuracy is 0.500000, time 55.467 samples/sec
+epoch 10, total_step 161140, total loss is 11.53 , inference loss is 4.80, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.463 samples/sec
+epoch 10, total_step 161160, total loss is 16.63 , inference loss is 9.90, weight deacy loss is 6.73, training accuracy is 0.468750, time 55.794 samples/sec
+epoch 10, total_step 161180, total loss is 12.33 , inference loss is 5.60, weight deacy loss is 6.73, training accuracy is 0.500000, time 55.682 samples/sec
+epoch 10, total_step 161200, total loss is 14.04 , inference loss is 7.30, weight deacy loss is 6.73, training accuracy is 0.437500, time 55.575 samples/sec
+epoch 10, total_step 161220, total loss is 17.44 , inference loss is 10.71, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.117 samples/sec
+epoch 10, total_step 161240, total loss is 16.93 , inference loss is 10.20, weight deacy loss is 6.73, training accuracy is 0.312500, time 56.785 samples/sec
+epoch 10, total_step 161260, total loss is 12.33 , inference loss is 5.60, weight deacy loss is 6.73, training accuracy is 0.531250, time 56.784 samples/sec
+epoch 10, total_step 161280, total loss is 19.35 , inference loss is 12.62, weight deacy loss is 6.73, training accuracy is 0.281250, time 56.493 samples/sec
+epoch 10, total_step 161300, total loss is 14.53 , inference loss is 7.80, weight deacy loss is 6.73, training accuracy is 0.500000, time 56.383 samples/sec
+epoch 10, total_step 161320, total loss is 14.26 , inference loss is 7.53, weight deacy loss is 6.73, training accuracy is 0.406250, time 56.353 samples/sec
+epoch 10, total_step 161340, total loss is 14.86 , inference loss is 8.13, weight deacy loss is 6.73, training accuracy is 0.375000, time 56.452 samples/sec
+epoch 10, total_step 161360, total loss is 14.59 , inference loss is 7.86, weight deacy loss is 6.73, training accuracy is 0.437500, time 56.443 samples/sec
+epoch 10, total_step 161380, total loss is 13.23 , inference loss is 6.50, weight deacy loss is 6.73, training accuracy is 0.500000, time 56.895 samples/sec
+epoch 10, total_step 161400, total loss is 15.43 , inference loss is 8.70, weight deacy loss is 6.73, training accuracy is 0.406250, time 57.290 samples/sec
+epoch 10, total_step 161420, total loss is 15.53 , inference loss is 8.80, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.321 samples/sec
+epoch 10, total_step 161440, total loss is 14.23 , inference loss is 7.51, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.360 samples/sec
+epoch 10, total_step 161460, total loss is 13.97 , inference loss is 7.24, weight deacy loss is 6.73, training accuracy is 0.500000, time 56.677 samples/sec
+epoch 10, total_step 161480, total loss is 15.06 , inference loss is 8.34, weight deacy loss is 6.73, training accuracy is 0.500000, time 56.631 samples/sec
+epoch 10, total_step 161500, total loss is 15.87 , inference loss is 9.15, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.460 samples/sec
+epoch 10, total_step 161520, total loss is 14.02 , inference loss is 7.30, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.790 samples/sec
+epoch 10, total_step 161540, total loss is 12.96 , inference loss is 6.24, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.736 samples/sec
+epoch 10, total_step 161560, total loss is 14.63 , inference loss is 7.90, weight deacy loss is 6.73, training accuracy is 0.437500, time 56.407 samples/sec
+epoch 10, total_step 161580, total loss is 11.49 , inference loss is 4.77, weight deacy loss is 6.73, training accuracy is 0.468750, time 56.080 samples/sec
+epoch 10, total_step 161600, total loss is 17.09 , inference loss is 10.36, weight deacy loss is 6.73, training accuracy is 0.437500, time 56.108 samples/sec
+epoch 10, total_step 161620, total loss is 13.10 , inference loss is 6.38, weight deacy loss is 6.73, training accuracy is 0.500000, time 57.133 samples/sec
+epoch 10, total_step 161640, total loss is 16.32 , inference loss is 9.60, weight deacy loss is 6.73, training accuracy is 0.250000, time 57.148 samples/sec
+epoch 10, total_step 161660, total loss is 16.92 , inference loss is 10.19, weight deacy loss is 6.73, training accuracy is 0.281250, time 57.106 samples/sec
+epoch 10, total_step 161680, total loss is 15.09 , inference loss is 8.36, weight deacy loss is 6.73, training accuracy is 0.437500, time 56.847 samples/sec
+epoch 10, total_step 161700, total loss is 16.30 , inference loss is 9.57, weight deacy loss is 6.73, training accuracy is 0.312500, time 56.746 samples/sec
+epoch 10, total_step 161720, total loss is 15.65 , inference loss is 8.92, weight deacy loss is 6.73, training accuracy is 0.375000, time 56.609 samples/sec
+epoch 10, total_step 161740, total loss is 13.11 , inference loss is 6.39, weight deacy loss is 6.73, training accuracy is 0.406250, time 57.237 samples/sec
+epoch 10, total_step 161760, total loss is 15.00 , inference loss is 8.27, weight deacy loss is 6.72, training accuracy is 0.468750, time 55.523 samples/sec
+epoch 10, total_step 161780, total loss is 13.94 , inference loss is 7.22, weight deacy loss is 6.72, training accuracy is 0.375000, time 56.669 samples/sec
+epoch 10, total_step 161800, total loss is 11.88 , inference loss is 5.16, weight deacy loss is 6.72, training accuracy is 0.593750, time 56.202 samples/sec
+epoch 10, total_step 161820, total loss is 18.59 , inference loss is 11.86, weight deacy loss is 6.72, training accuracy is 0.312500, time 56.396 samples/sec
+epoch 10, total_step 161840, total loss is 14.21 , inference loss is 7.49, weight deacy loss is 6.72, training accuracy is 0.500000, time 55.490 samples/sec
+epoch 10, total_step 161860, total loss is 11.32 , inference loss is 4.60, weight deacy loss is 6.72, training accuracy is 0.500000, time 55.801 samples/sec
+epoch 10, total_step 161880, total loss is 14.17 , inference loss is 7.45, weight deacy loss is 6.72, training accuracy is 0.468750, time 56.032 samples/sec
+epoch 10, total_step 161900, total loss is 13.73 , inference loss is 7.00, weight deacy loss is 6.72, training accuracy is 0.468750, time 56.373 samples/sec
+epoch 10, total_step 161920, total loss is 11.31 , inference loss is 4.59, weight deacy loss is 6.72, training accuracy is 0.562500, time 56.105 samples/sec
+epoch 10, total_step 161940, total loss is 13.81 , inference loss is 7.09, weight deacy loss is 6.72, training accuracy is 0.531250, time 56.127 samples/sec
+epoch 10, total_step 161960, total loss is 10.67 , inference loss is 3.94, weight deacy loss is 6.72, training accuracy is 0.312500, time 55.826 samples/sec
+epoch 10, total_step 161980, total loss is 12.88 , inference loss is 6.16, weight deacy loss is 6.72, training accuracy is 0.343750, time 56.162 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.26480899999997
+best_threshold_index 123 0.9905555555555555
+best_threshold_index 123 0.9896296296296296
+best_threshold_index 123 0.99
+best_threshold_index 124 0.9903703703703703
+best_threshold_index 123 0.9909259259259259
+best_threshold_index 123 0.9896296296296296
+best_threshold_index 123 0.9905555555555555
+best_threshold_index 123 0.99
+best_threshold_index 123 0.9888888888888889
+best_threshold_index 123 0.9896296296296296
+[lfw][162000]XNorm: 21.078154
+[lfw][162000]Accuracy-Flip: 0.98967+-0.00526
+testing verification..
+(14000, 512)
+infer time 85.4815839999999
+best_threshold_index 121 0.9877777777777778
+best_threshold_index 125 0.986031746031746
+best_threshold_index 125 0.9865079365079366
+best_threshold_index 125 0.986031746031746
+best_threshold_index 125 0.9868253968253968
+best_threshold_index 125 0.9858730158730159
+best_threshold_index 125 0.9868253968253968
+best_threshold_index 125 0.986031746031746
+best_threshold_index 125 0.9861904761904762
+best_threshold_index 123 0.9865079365079366
+[cfp_ff][162000]XNorm: 20.945657
+[cfp_ff][162000]Accuracy-Flip: 0.98529+-0.00642
+testing verification..
+(14000, 512)
+infer time 85.5317130000001
+best_threshold_index 138 0.9312698412698412
+best_threshold_index 138 0.9304761904761905
+best_threshold_index 138 0.9296825396825397
+best_threshold_index 138 0.9314285714285714
+best_threshold_index 138 0.9334920634920635
+best_threshold_index 138 0.9292063492063493
+best_threshold_index 140 0.9285714285714286
+best_threshold_index 138 0.9288888888888889
+best_threshold_index 138 0.9292063492063493
+best_threshold_index 138 0.9293650793650794
+[cfp_fp][162000]XNorm: 20.928852
+[cfp_fp][162000]Accuracy-Flip: 0.92971+-0.01271
+testing verification..
+(12000, 512)
+infer time 72.61535300000007
+best_threshold_index 138 0.9068518518518518
+best_threshold_index 138 0.9068518518518518
+best_threshold_index 138 0.9092592592592592
+best_threshold_index 137 0.9092592592592592
+best_threshold_index 138 0.9118518518518518
+best_threshold_index 138 0.9055555555555556
+best_threshold_index 138 0.9098148148148149
+best_threshold_index 138 0.9103703703703704
+best_threshold_index 137 0.914074074074074
+best_threshold_index 138 0.9081481481481481
+[agedb_30][162000]XNorm: 22.260597
+[agedb_30][162000]Accuracy-Flip: 0.90817+-0.02238
+test accuracy is: 0.9896666666666667
+epoch 10, total_step 162000, total loss is 16.67 , inference loss is 9.94, weight deacy loss is 6.72, training accuracy is 0.406250, time 53.086 samples/sec
+epoch 10, total_step 162020, total loss is 12.13 , inference loss is 5.41, weight deacy loss is 6.72, training accuracy is 0.625000, time 57.042 samples/sec
+epoch 10, total_step 162040, total loss is 14.53 , inference loss is 7.81, weight deacy loss is 6.72, training accuracy is 0.468750, time 56.800 samples/sec
+epoch 10, total_step 162060, total loss is 11.50 , inference loss is 4.77, weight deacy loss is 6.72, training accuracy is 0.468750, time 57.445 samples/sec
+epoch 10, total_step 162080, total loss is 18.61 , inference loss is 11.89, weight deacy loss is 6.72, training accuracy is 0.250000, time 56.102 samples/sec
+epoch 10, total_step 162100, total loss is 14.55 , inference loss is 7.83, weight deacy loss is 6.72, training accuracy is 0.437500, time 56.119 samples/sec
+epoch 10, total_step 162120, total loss is 12.44 , inference loss is 5.71, weight deacy loss is 6.72, training accuracy is 0.468750, time 56.152 samples/sec
+epoch 10, total_step 162140, total loss is 13.58 , inference loss is 6.86, weight deacy loss is 6.72, training accuracy is 0.406250, time 56.934 samples/sec
+epoch 10, total_step 162160, total loss is 15.90 , inference loss is 9.17, weight deacy loss is 6.72, training accuracy is 0.437500, time 56.531 samples/sec
+epoch 10, total_step 162180, total loss is 16.47 , inference loss is 9.75, weight deacy loss is 6.72, training accuracy is 0.187500, time 55.889 samples/sec
+epoch 10, total_step 162200, total loss is 13.74 , inference loss is 7.01, weight deacy loss is 6.72, training accuracy is 0.406250, time 56.906 samples/sec
+epoch 10, total_step 162220, total loss is 16.93 , inference loss is 10.21, weight deacy loss is 6.72, training accuracy is 0.343750, time 56.392 samples/sec
+epoch 10, total_step 162240, total loss is 17.25 , inference loss is 10.53, weight deacy loss is 6.72, training accuracy is 0.437500, time 55.987 samples/sec
+epoch 10, total_step 162260, total loss is 12.84 , inference loss is 6.12, weight deacy loss is 6.72, training accuracy is 0.437500, time 56.788 samples/sec
+epoch 10, total_step 162280, total loss is 12.99 , inference loss is 6.27, weight deacy loss is 6.72, training accuracy is 0.437500, time 55.308 samples/sec
+epoch 10, total_step 162300, total loss is 15.91 , inference loss is 9.19, weight deacy loss is 6.72, training accuracy is 0.375000, time 56.035 samples/sec
+epoch 10, total_step 162320, total loss is 16.47 , inference loss is 9.75, weight deacy loss is 6.72, training accuracy is 0.437500, time 56.104 samples/sec
+epoch 10, total_step 162340, total loss is 13.96 , inference loss is 7.24, weight deacy loss is 6.72, training accuracy is 0.500000, time 56.880 samples/sec
+epoch 10, total_step 162360, total loss is 18.72 , inference loss is 12.00, weight deacy loss is 6.72, training accuracy is 0.312500, time 56.563 samples/sec
+epoch 10, total_step 162380, total loss is 13.78 , inference loss is 7.06, weight deacy loss is 6.72, training accuracy is 0.593750, time 56.739 samples/sec
+epoch 10, total_step 162400, total loss is 13.30 , inference loss is 6.58, weight deacy loss is 6.72, training accuracy is 0.531250, time 56.450 samples/sec
+epoch 10, total_step 162420, total loss is 10.62 , inference loss is 3.90, weight deacy loss is 6.72, training accuracy is 0.531250, time 56.554 samples/sec
+epoch 10, total_step 162440, total loss is 11.45 , inference loss is 4.73, weight deacy loss is 6.72, training accuracy is 0.531250, time 56.294 samples/sec
+epoch 10, total_step 162460, total loss is 16.43 , inference loss is 9.72, weight deacy loss is 6.72, training accuracy is 0.468750, time 55.851 samples/sec
+epoch 10, total_step 162480, total loss is 13.52 , inference loss is 6.80, weight deacy loss is 6.72, training accuracy is 0.437500, time 56.298 samples/sec
+epoch 10, total_step 162500, total loss is 13.76 , inference loss is 7.04, weight deacy loss is 6.72, training accuracy is 0.281250, time 56.199 samples/sec
+epoch 10, total_step 162520, total loss is 10.19 , inference loss is 3.47, weight deacy loss is 6.72, training accuracy is 0.625000, time 56.631 samples/sec
+epoch 10, total_step 162540, total loss is 14.36 , inference loss is 7.64, weight deacy loss is 6.72, training accuracy is 0.406250, time 56.765 samples/sec
+epoch 10, total_step 162560, total loss is 17.07 , inference loss is 10.35, weight deacy loss is 6.72, training accuracy is 0.375000, time 57.675 samples/sec
+epoch 10, total_step 162580, total loss is 15.82 , inference loss is 9.10, weight deacy loss is 6.72, training accuracy is 0.437500, time 56.461 samples/sec
+epoch 10, total_step 162600, total loss is 12.28 , inference loss is 5.56, weight deacy loss is 6.72, training accuracy is 0.437500, time 54.531 samples/sec
+epoch 10, total_step 162620, total loss is 15.91 , inference loss is 9.19, weight deacy loss is 6.72, training accuracy is 0.406250, time 56.056 samples/sec
+epoch 10, total_step 162640, total loss is 15.89 , inference loss is 9.18, weight deacy loss is 6.72, training accuracy is 0.406250, time 56.296 samples/sec
+epoch 10, total_step 162660, total loss is 13.84 , inference loss is 7.12, weight deacy loss is 6.72, training accuracy is 0.437500, time 56.151 samples/sec
+epoch 10, total_step 162680, total loss is 16.55 , inference loss is 9.84, weight deacy loss is 6.72, training accuracy is 0.468750, time 57.022 samples/sec
+epoch 10, total_step 162700, total loss is 15.43 , inference loss is 8.72, weight deacy loss is 6.72, training accuracy is 0.406250, time 55.929 samples/sec
+epoch 10, total_step 162720, total loss is 15.42 , inference loss is 8.71, weight deacy loss is 6.72, training accuracy is 0.437500, time 56.556 samples/sec
+epoch 10, total_step 162740, total loss is 12.56 , inference loss is 5.84, weight deacy loss is 6.72, training accuracy is 0.562500, time 56.433 samples/sec
+epoch 10, total_step 162760, total loss is 13.80 , inference loss is 7.08, weight deacy loss is 6.71, training accuracy is 0.406250, time 56.550 samples/sec
+epoch 10, total_step 162780, total loss is 12.67 , inference loss is 5.96, weight deacy loss is 6.71, training accuracy is 0.437500, time 56.818 samples/sec
+epoch 10, total_step 162800, total loss is 12.29 , inference loss is 5.57, weight deacy loss is 6.71, training accuracy is 0.500000, time 56.351 samples/sec
+epoch 10, total_step 162820, total loss is 15.44 , inference loss is 8.72, weight deacy loss is 6.71, training accuracy is 0.468750, time 55.522 samples/sec
+epoch 10, total_step 162840, total loss is 12.90 , inference loss is 6.18, weight deacy loss is 6.71, training accuracy is 0.281250, time 56.385 samples/sec
+epoch 10, total_step 162860, total loss is 15.49 , inference loss is 8.77, weight deacy loss is 6.71, training accuracy is 0.406250, time 56.582 samples/sec
+epoch 10, total_step 162880, total loss is 14.65 , inference loss is 7.93, weight deacy loss is 6.71, training accuracy is 0.437500, time 57.437 samples/sec
+epoch 10, total_step 162900, total loss is 14.03 , inference loss is 7.32, weight deacy loss is 6.71, training accuracy is 0.437500, time 56.678 samples/sec
+epoch 10, total_step 162920, total loss is 11.94 , inference loss is 5.23, weight deacy loss is 6.71, training accuracy is 0.531250, time 56.550 samples/sec
+epoch 10, total_step 162940, total loss is 15.59 , inference loss is 8.87, weight deacy loss is 6.71, training accuracy is 0.281250, time 56.653 samples/sec
+epoch 10, total_step 162960, total loss is 15.94 , inference loss is 9.22, weight deacy loss is 6.71, training accuracy is 0.375000, time 56.470 samples/sec
+epoch 10, total_step 162980, total loss is 12.39 , inference loss is 5.68, weight deacy loss is 6.71, training accuracy is 0.437500, time 57.189 samples/sec
+epoch 10, total_step 163000, total loss is 16.64 , inference loss is 9.93, weight deacy loss is 6.71, training accuracy is 0.375000, time 56.341 samples/sec
+epoch 10, total_step 163020, total loss is 13.09 , inference loss is 6.38, weight deacy loss is 6.71, training accuracy is 0.562500, time 56.493 samples/sec
+epoch 10, total_step 163040, total loss is 10.55 , inference loss is 3.84, weight deacy loss is 6.71, training accuracy is 0.531250, time 57.055 samples/sec
+epoch 10, total_step 163060, total loss is 14.39 , inference loss is 7.68, weight deacy loss is 6.71, training accuracy is 0.562500, time 56.473 samples/sec
+epoch 10, total_step 163080, total loss is 14.83 , inference loss is 8.12, weight deacy loss is 6.71, training accuracy is 0.531250, time 55.747 samples/sec
+epoch 10, total_step 163100, total loss is 15.37 , inference loss is 8.66, weight deacy loss is 6.71, training accuracy is 0.437500, time 56.294 samples/sec
+epoch 10, total_step 163120, total loss is 12.70 , inference loss is 5.99, weight deacy loss is 6.71, training accuracy is 0.468750, time 56.766 samples/sec
+epoch 10, total_step 163140, total loss is 13.86 , inference loss is 7.14, weight deacy loss is 6.71, training accuracy is 0.281250, time 57.869 samples/sec
+epoch 10, total_step 163160, total loss is 16.20 , inference loss is 9.49, weight deacy loss is 6.71, training accuracy is 0.218750, time 55.483 samples/sec
+epoch 10, total_step 163180, total loss is 14.55 , inference loss is 7.84, weight deacy loss is 6.71, training accuracy is 0.468750, time 55.954 samples/sec
+epoch 10, total_step 163200, total loss is 11.51 , inference loss is 4.80, weight deacy loss is 6.71, training accuracy is 0.531250, time 56.260 samples/sec
+epoch 10, total_step 163220, total loss is 11.95 , inference loss is 5.24, weight deacy loss is 6.71, training accuracy is 0.406250, time 56.571 samples/sec
+epoch 10, total_step 163240, total loss is 12.70 , inference loss is 5.99, weight deacy loss is 6.71, training accuracy is 0.500000, time 56.571 samples/sec
+epoch 10, total_step 163260, total loss is 14.50 , inference loss is 7.79, weight deacy loss is 6.71, training accuracy is 0.437500, time 56.980 samples/sec
+epoch 10, total_step 163280, total loss is 13.77 , inference loss is 7.06, weight deacy loss is 6.71, training accuracy is 0.500000, time 56.784 samples/sec
+epoch 10, total_step 163300, total loss is 12.17 , inference loss is 5.47, weight deacy loss is 6.71, training accuracy is 0.468750, time 55.811 samples/sec
+epoch 10, total_step 163320, total loss is 14.16 , inference loss is 7.45, weight deacy loss is 6.71, training accuracy is 0.343750, time 56.879 samples/sec
+epoch 10, total_step 163340, total loss is 9.51 , inference loss is 2.81, weight deacy loss is 6.71, training accuracy is 0.656250, time 56.605 samples/sec
+epoch 10, total_step 163360, total loss is 16.61 , inference loss is 9.90, weight deacy loss is 6.71, training accuracy is 0.468750, time 56.441 samples/sec
+epoch 10, total_step 163380, total loss is 13.51 , inference loss is 6.80, weight deacy loss is 6.71, training accuracy is 0.500000, time 57.527 samples/sec
+epoch 10, total_step 163400, total loss is 14.05 , inference loss is 7.35, weight deacy loss is 6.71, training accuracy is 0.437500, time 56.381 samples/sec
+epoch 10, total_step 163420, total loss is 13.37 , inference loss is 6.66, weight deacy loss is 6.71, training accuracy is 0.500000, time 56.611 samples/sec
+epoch 10, total_step 163440, total loss is 13.65 , inference loss is 6.95, weight deacy loss is 6.71, training accuracy is 0.656250, time 56.885 samples/sec
+epoch 10, total_step 163460, total loss is 12.42 , inference loss is 5.71, weight deacy loss is 6.71, training accuracy is 0.531250, time 55.442 samples/sec
+epoch 10, total_step 163480, total loss is 18.50 , inference loss is 11.79, weight deacy loss is 6.71, training accuracy is 0.468750, time 56.706 samples/sec
+epoch 10, total_step 163500, total loss is 13.95 , inference loss is 7.24, weight deacy loss is 6.71, training accuracy is 0.406250, time 56.139 samples/sec
+epoch 10, total_step 163520, total loss is 17.86 , inference loss is 11.16, weight deacy loss is 6.71, training accuracy is 0.406250, time 56.223 samples/sec
+epoch 10, total_step 163540, total loss is 16.80 , inference loss is 10.10, weight deacy loss is 6.70, training accuracy is 0.375000, time 56.625 samples/sec
+epoch 10, total_step 163560, total loss is 12.78 , inference loss is 6.08, weight deacy loss is 6.70, training accuracy is 0.531250, time 56.003 samples/sec
+epoch 10, total_step 163580, total loss is 14.28 , inference loss is 7.57, weight deacy loss is 6.70, training accuracy is 0.437500, time 56.622 samples/sec
+epoch 10, total_step 163600, total loss is 13.11 , inference loss is 6.41, weight deacy loss is 6.70, training accuracy is 0.437500, time 56.182 samples/sec
+epoch 10, total_step 163620, total loss is 12.34 , inference loss is 5.63, weight deacy loss is 6.70, training accuracy is 0.500000, time 56.617 samples/sec
+epoch 10, total_step 163640, total loss is 11.85 , inference loss is 5.15, weight deacy loss is 6.70, training accuracy is 0.375000, time 55.892 samples/sec
+epoch 10, total_step 163660, total loss is 15.49 , inference loss is 8.79, weight deacy loss is 6.70, training accuracy is 0.500000, time 56.614 samples/sec
+epoch 10, total_step 163680, total loss is 12.30 , inference loss is 5.59, weight deacy loss is 6.70, training accuracy is 0.625000, time 56.571 samples/sec
+epoch 10, total_step 163700, total loss is 16.07 , inference loss is 9.37, weight deacy loss is 6.70, training accuracy is 0.406250, time 55.602 samples/sec
+epoch 10, total_step 163720, total loss is 10.31 , inference loss is 3.61, weight deacy loss is 6.70, training accuracy is 0.625000, time 56.454 samples/sec
+epoch 10, total_step 163740, total loss is 14.11 , inference loss is 7.41, weight deacy loss is 6.70, training accuracy is 0.343750, time 56.547 samples/sec
+epoch 10, total_step 163760, total loss is 12.30 , inference loss is 5.60, weight deacy loss is 6.70, training accuracy is 0.468750, time 56.893 samples/sec
+epoch 10, total_step 163780, total loss is 9.97 , inference loss is 3.27, weight deacy loss is 6.70, training accuracy is 0.593750, time 56.601 samples/sec
+epoch 10, total_step 163800, total loss is 14.23 , inference loss is 7.53, weight deacy loss is 6.70, training accuracy is 0.437500, time 55.987 samples/sec
+epoch 10, total_step 163820, total loss is 10.29 , inference loss is 3.59, weight deacy loss is 6.70, training accuracy is 0.656250, time 56.510 samples/sec
+epoch 10, total_step 163840, total loss is 15.06 , inference loss is 8.36, weight deacy loss is 6.70, training accuracy is 0.500000, time 56.874 samples/sec
+epoch 10, total_step 163860, total loss is 13.49 , inference loss is 6.79, weight deacy loss is 6.70, training accuracy is 0.562500, time 56.386 samples/sec
+epoch 10, total_step 163880, total loss is 15.31 , inference loss is 8.61, weight deacy loss is 6.70, training accuracy is 0.312500, time 56.851 samples/sec
+epoch 10, total_step 163900, total loss is 19.89 , inference loss is 13.19, weight deacy loss is 6.70, training accuracy is 0.218750, time 56.727 samples/sec
+epoch 10, total_step 163920, total loss is 14.21 , inference loss is 7.52, weight deacy loss is 6.70, training accuracy is 0.468750, time 55.923 samples/sec
+epoch 10, total_step 163940, total loss is 16.65 , inference loss is 9.95, weight deacy loss is 6.70, training accuracy is 0.406250, time 56.631 samples/sec
+epoch 10, total_step 163960, total loss is 14.92 , inference loss is 8.22, weight deacy loss is 6.70, training accuracy is 0.437500, time 55.880 samples/sec
+epoch 10, total_step 163980, total loss is 15.62 , inference loss is 8.93, weight deacy loss is 6.70, training accuracy is 0.343750, time 56.701 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.778248
+best_threshold_index 114 0.9862962962962963
+best_threshold_index 114 0.9855555555555555
+best_threshold_index 117 0.985
+best_threshold_index 110 0.9859259259259259
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 114 0.9848148148148148
+best_threshold_index 114 0.9853703703703703
+best_threshold_index 114 0.9848148148148148
+best_threshold_index 117 0.985
+best_threshold_index 117 0.9846296296296296
+[lfw][164000]XNorm: 20.479883
+[lfw][164000]Accuracy-Flip: 0.98350+-0.00519
+testing verification..
+(14000, 512)
+infer time 85.23601199999983
+best_threshold_index 119 0.9868253968253968
+best_threshold_index 120 0.9858730158730159
+best_threshold_index 120 0.9858730158730159
+best_threshold_index 120 0.9861904761904762
+best_threshold_index 120 0.9877777777777778
+best_threshold_index 120 0.9858730158730159
+best_threshold_index 120 0.9866666666666667
+best_threshold_index 120 0.9861904761904762
+best_threshold_index 120 0.9855555555555555
+best_threshold_index 120 0.9861904761904762
+[cfp_ff][164000]XNorm: 20.506383
+[cfp_ff][164000]Accuracy-Flip: 0.98586+-0.00580
+testing verification..
+(14000, 512)
+infer time 84.67803999999998
+best_threshold_index 138 0.9242857142857143
+best_threshold_index 139 0.923968253968254
+best_threshold_index 139 0.9217460317460318
+best_threshold_index 138 0.9250793650793651
+best_threshold_index 138 0.9266666666666666
+best_threshold_index 139 0.9212698412698412
+best_threshold_index 139 0.9215873015873016
+best_threshold_index 139 0.9219047619047619
+best_threshold_index 139 0.9236507936507936
+best_threshold_index 139 0.9234920634920635
+[cfp_fp][164000]XNorm: 20.917807
+[cfp_fp][164000]Accuracy-Flip: 0.92214+-0.01543
+testing verification..
+(12000, 512)
+infer time 73.27606200000008
+best_threshold_index 136 0.9020370370370371
+best_threshold_index 136 0.9007407407407407
+best_threshold_index 136 0.9033333333333333
+best_threshold_index 136 0.905
+best_threshold_index 136 0.9083333333333333
+best_threshold_index 136 0.9022222222222223
+best_threshold_index 136 0.9042592592592592
+best_threshold_index 136 0.9072222222222223
+best_threshold_index 136 0.9103703703703704
+best_threshold_index 136 0.9031481481481481
+[agedb_30][164000]XNorm: 21.768470
+[agedb_30][164000]Accuracy-Flip: 0.90467+-0.02627
+test accuracy is: 0.9834999999999999
+epoch 10, total_step 164000, total loss is 17.53 , inference loss is 10.83, weight deacy loss is 6.70, training accuracy is 0.343750, time 53.246 samples/sec
+epoch 10, total_step 164020, total loss is 17.45 , inference loss is 10.75, weight deacy loss is 6.70, training accuracy is 0.343750, time 56.178 samples/sec
+epoch 10, total_step 164040, total loss is 15.30 , inference loss is 8.61, weight deacy loss is 6.70, training accuracy is 0.437500, time 56.114 samples/sec
+epoch 10, total_step 164060, total loss is 15.80 , inference loss is 9.11, weight deacy loss is 6.70, training accuracy is 0.375000, time 56.534 samples/sec
+epoch 10, total_step 164080, total loss is 13.08 , inference loss is 6.39, weight deacy loss is 6.70, training accuracy is 0.500000, time 56.602 samples/sec
+epoch 10, total_step 164100, total loss is 14.69 , inference loss is 7.99, weight deacy loss is 6.70, training accuracy is 0.343750, time 56.212 samples/sec
+epoch 10, total_step 164120, total loss is 13.34 , inference loss is 6.65, weight deacy loss is 6.70, training accuracy is 0.468750, time 56.225 samples/sec
+epoch 10, total_step 164140, total loss is 15.62 , inference loss is 8.92, weight deacy loss is 6.70, training accuracy is 0.375000, time 57.188 samples/sec
+epoch 10, total_step 164160, total loss is 16.47 , inference loss is 9.78, weight deacy loss is 6.70, training accuracy is 0.343750, time 56.631 samples/sec
+epoch 10, total_step 164180, total loss is 11.18 , inference loss is 4.49, weight deacy loss is 6.70, training accuracy is 0.562500, time 56.823 samples/sec
+epoch 10, total_step 164200, total loss is 17.10 , inference loss is 10.40, weight deacy loss is 6.70, training accuracy is 0.343750, time 56.382 samples/sec
+epoch 10, total_step 164220, total loss is 11.73 , inference loss is 5.03, weight deacy loss is 6.70, training accuracy is 0.437500, time 57.214 samples/sec
+epoch 10, total_step 164240, total loss is 15.96 , inference loss is 9.27, weight deacy loss is 6.70, training accuracy is 0.281250, time 56.697 samples/sec
+epoch 10, total_step 164260, total loss is 16.11 , inference loss is 9.41, weight deacy loss is 6.70, training accuracy is 0.343750, time 56.616 samples/sec
+epoch 10, total_step 164280, total loss is 12.87 , inference loss is 6.18, weight deacy loss is 6.70, training accuracy is 0.531250, time 56.792 samples/sec
+epoch 10, total_step 164300, total loss is 13.04 , inference loss is 6.35, weight deacy loss is 6.70, training accuracy is 0.406250, time 57.136 samples/sec
+epoch 10, total_step 164320, total loss is 13.14 , inference loss is 6.44, weight deacy loss is 6.70, training accuracy is 0.406250, time 56.464 samples/sec
+epoch 10, total_step 164340, total loss is 12.05 , inference loss is 5.36, weight deacy loss is 6.70, training accuracy is 0.468750, time 56.127 samples/sec
+epoch 10, total_step 164360, total loss is 12.33 , inference loss is 5.64, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.636 samples/sec
+epoch 10, total_step 164380, total loss is 10.96 , inference loss is 4.26, weight deacy loss is 6.69, training accuracy is 0.468750, time 56.290 samples/sec
+epoch 10, total_step 164400, total loss is 13.59 , inference loss is 6.90, weight deacy loss is 6.69, training accuracy is 0.406250, time 55.447 samples/sec
+epoch 10, total_step 164420, total loss is 13.60 , inference loss is 6.91, weight deacy loss is 6.69, training accuracy is 0.500000, time 56.014 samples/sec
+epoch 10, total_step 164440, total loss is 16.23 , inference loss is 9.53, weight deacy loss is 6.69, training accuracy is 0.500000, time 56.057 samples/sec
+epoch 10, total_step 164460, total loss is 17.36 , inference loss is 10.67, weight deacy loss is 6.69, training accuracy is 0.312500, time 56.949 samples/sec
+epoch 10, total_step 164480, total loss is 13.28 , inference loss is 6.59, weight deacy loss is 6.69, training accuracy is 0.500000, time 57.311 samples/sec
+epoch 10, total_step 164500, total loss is 15.12 , inference loss is 8.42, weight deacy loss is 6.69, training accuracy is 0.218750, time 55.989 samples/sec
+epoch 10, total_step 164520, total loss is 14.83 , inference loss is 8.13, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.817 samples/sec
+epoch 10, total_step 164540, total loss is 16.26 , inference loss is 9.57, weight deacy loss is 6.69, training accuracy is 0.312500, time 56.278 samples/sec
+epoch 10, total_step 164560, total loss is 17.56 , inference loss is 10.87, weight deacy loss is 6.69, training accuracy is 0.437500, time 56.917 samples/sec
+epoch 10, total_step 164580, total loss is 14.77 , inference loss is 8.08, weight deacy loss is 6.69, training accuracy is 0.437500, time 56.505 samples/sec
+epoch 10, total_step 164600, total loss is 14.67 , inference loss is 7.98, weight deacy loss is 6.69, training accuracy is 0.468750, time 56.003 samples/sec
+epoch 10, total_step 164620, total loss is 9.88 , inference loss is 3.18, weight deacy loss is 6.69, training accuracy is 0.562500, time 56.806 samples/sec
+epoch 10, total_step 164640, total loss is 15.99 , inference loss is 9.30, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.090 samples/sec
+epoch 10, total_step 164660, total loss is 13.02 , inference loss is 6.33, weight deacy loss is 6.69, training accuracy is 0.500000, time 56.486 samples/sec
+epoch 10, total_step 164680, total loss is 14.62 , inference loss is 7.93, weight deacy loss is 6.69, training accuracy is 0.468750, time 56.606 samples/sec
+epoch 10, total_step 164700, total loss is 13.79 , inference loss is 7.10, weight deacy loss is 6.69, training accuracy is 0.375000, time 56.587 samples/sec
+epoch 10, total_step 164720, total loss is 18.58 , inference loss is 11.89, weight deacy loss is 6.69, training accuracy is 0.218750, time 56.918 samples/sec
+epoch 10, total_step 164740, total loss is 14.46 , inference loss is 7.77, weight deacy loss is 6.69, training accuracy is 0.468750, time 56.882 samples/sec
+epoch 10, total_step 164760, total loss is 13.42 , inference loss is 6.73, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.979 samples/sec
+epoch 10, total_step 164780, total loss is 14.36 , inference loss is 7.67, weight deacy loss is 6.69, training accuracy is 0.375000, time 56.739 samples/sec
+epoch 10, total_step 164800, total loss is 14.53 , inference loss is 7.84, weight deacy loss is 6.69, training accuracy is 0.312500, time 55.844 samples/sec
+epoch 10, total_step 164820, total loss is 16.07 , inference loss is 9.38, weight deacy loss is 6.69, training accuracy is 0.437500, time 56.052 samples/sec
+epoch 10, total_step 164840, total loss is 13.21 , inference loss is 6.52, weight deacy loss is 6.69, training accuracy is 0.375000, time 56.794 samples/sec
+epoch 10, total_step 164860, total loss is 19.85 , inference loss is 13.16, weight deacy loss is 6.69, training accuracy is 0.187500, time 55.921 samples/sec
+epoch 10, total_step 164880, total loss is 18.36 , inference loss is 11.67, weight deacy loss is 6.69, training accuracy is 0.312500, time 56.314 samples/sec
+epoch 10, total_step 164900, total loss is 13.00 , inference loss is 6.31, weight deacy loss is 6.69, training accuracy is 0.312500, time 56.516 samples/sec
+epoch 10, total_step 164920, total loss is 12.99 , inference loss is 6.30, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.136 samples/sec
+epoch 10, total_step 164940, total loss is 15.82 , inference loss is 9.13, weight deacy loss is 6.69, training accuracy is 0.312500, time 56.552 samples/sec
+epoch 10, total_step 164960, total loss is 14.39 , inference loss is 7.70, weight deacy loss is 6.69, training accuracy is 0.375000, time 56.512 samples/sec
+epoch 10, total_step 164980, total loss is 14.48 , inference loss is 7.79, weight deacy loss is 6.69, training accuracy is 0.375000, time 56.085 samples/sec
+epoch 10, total_step 165000, total loss is 15.46 , inference loss is 8.77, weight deacy loss is 6.69, training accuracy is 0.281250, time 55.722 samples/sec
+epoch 10, total_step 165020, total loss is 14.58 , inference loss is 7.89, weight deacy loss is 6.69, training accuracy is 0.343750, time 56.760 samples/sec
+epoch 10, total_step 165040, total loss is 12.97 , inference loss is 6.29, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.607 samples/sec
+epoch 10, total_step 165060, total loss is 16.28 , inference loss is 9.60, weight deacy loss is 6.69, training accuracy is 0.343750, time 56.355 samples/sec
+epoch 10, total_step 165080, total loss is 17.49 , inference loss is 10.80, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.137 samples/sec
+epoch 10, total_step 165100, total loss is 17.38 , inference loss is 10.69, weight deacy loss is 6.69, training accuracy is 0.375000, time 56.957 samples/sec
+epoch 10, total_step 165120, total loss is 14.72 , inference loss is 8.03, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.604 samples/sec
+epoch 10, total_step 165140, total loss is 13.81 , inference loss is 7.13, weight deacy loss is 6.69, training accuracy is 0.406250, time 56.449 samples/sec
+epoch 10, total_step 165160, total loss is 14.59 , inference loss is 7.90, weight deacy loss is 6.69, training accuracy is 0.343750, time 56.046 samples/sec
+epoch 10, total_step 165180, total loss is 14.66 , inference loss is 7.98, weight deacy loss is 6.68, training accuracy is 0.375000, time 56.245 samples/sec
+epoch 10, total_step 165200, total loss is 10.94 , inference loss is 4.25, weight deacy loss is 6.68, training accuracy is 0.468750, time 56.215 samples/sec
+epoch 10, total_step 165220, total loss is 17.10 , inference loss is 10.41, weight deacy loss is 6.68, training accuracy is 0.187500, time 56.722 samples/sec
+epoch 10, total_step 165240, total loss is 21.44 , inference loss is 14.75, weight deacy loss is 6.68, training accuracy is 0.406250, time 56.925 samples/sec
+epoch 10, total_step 165260, total loss is 16.53 , inference loss is 9.85, weight deacy loss is 6.68, training accuracy is 0.375000, time 56.011 samples/sec
+epoch 10, total_step 165280, total loss is 18.28 , inference loss is 11.59, weight deacy loss is 6.68, training accuracy is 0.250000, time 56.380 samples/sec
+epoch 10, total_step 165300, total loss is 15.24 , inference loss is 8.56, weight deacy loss is 6.68, training accuracy is 0.312500, time 55.851 samples/sec
+epoch 10, total_step 165320, total loss is 19.53 , inference loss is 12.85, weight deacy loss is 6.68, training accuracy is 0.281250, time 57.607 samples/sec
+epoch 10, total_step 165340, total loss is 17.35 , inference loss is 10.67, weight deacy loss is 6.68, training accuracy is 0.312500, time 55.777 samples/sec
+epoch 10, total_step 165360, total loss is 15.13 , inference loss is 8.45, weight deacy loss is 6.68, training accuracy is 0.375000, time 57.266 samples/sec
+epoch 10, total_step 165380, total loss is 15.46 , inference loss is 8.78, weight deacy loss is 6.68, training accuracy is 0.375000, time 57.133 samples/sec
+epoch 10, total_step 165400, total loss is 14.10 , inference loss is 7.42, weight deacy loss is 6.68, training accuracy is 0.250000, time 56.920 samples/sec
+epoch 10, total_step 165420, total loss is 14.68 , inference loss is 7.99, weight deacy loss is 6.68, training accuracy is 0.406250, time 56.932 samples/sec
+epoch 10, total_step 165440, total loss is 13.64 , inference loss is 6.96, weight deacy loss is 6.68, training accuracy is 0.343750, time 56.989 samples/sec
+epoch 10, total_step 165460, total loss is 20.25 , inference loss is 13.57, weight deacy loss is 6.68, training accuracy is 0.343750, time 56.882 samples/sec
+epoch 10, total_step 165480, total loss is 12.71 , inference loss is 6.02, weight deacy loss is 6.68, training accuracy is 0.468750, time 57.432 samples/sec
+epoch 10, total_step 165500, total loss is 14.55 , inference loss is 7.87, weight deacy loss is 6.68, training accuracy is 0.375000, time 56.170 samples/sec
+epoch 10, total_step 165520, total loss is 16.38 , inference loss is 9.70, weight deacy loss is 6.68, training accuracy is 0.375000, time 56.625 samples/sec
+epoch 10, total_step 165540, total loss is 16.04 , inference loss is 9.36, weight deacy loss is 6.68, training accuracy is 0.281250, time 56.566 samples/sec
+epoch 10, total_step 165560, total loss is 13.28 , inference loss is 6.59, weight deacy loss is 6.68, training accuracy is 0.437500, time 54.912 samples/sec
+epoch 10, total_step 165580, total loss is 16.37 , inference loss is 9.69, weight deacy loss is 6.68, training accuracy is 0.375000, time 56.558 samples/sec
+epoch 10, total_step 165600, total loss is 16.54 , inference loss is 9.86, weight deacy loss is 6.68, training accuracy is 0.375000, time 55.793 samples/sec
+epoch 10, total_step 165620, total loss is 15.53 , inference loss is 8.85, weight deacy loss is 6.68, training accuracy is 0.406250, time 56.859 samples/sec
+epoch 10, total_step 165640, total loss is 13.12 , inference loss is 6.44, weight deacy loss is 6.68, training accuracy is 0.343750, time 55.676 samples/sec
+epoch 10, total_step 165660, total loss is 14.57 , inference loss is 7.89, weight deacy loss is 6.68, training accuracy is 0.406250, time 56.584 samples/sec
+epoch 10, total_step 165680, total loss is 13.73 , inference loss is 7.05, weight deacy loss is 6.68, training accuracy is 0.250000, time 56.389 samples/sec
+epoch 10, total_step 165700, total loss is 16.20 , inference loss is 9.52, weight deacy loss is 6.68, training accuracy is 0.218750, time 56.848 samples/sec
+epoch 10, total_step 165720, total loss is 18.28 , inference loss is 11.60, weight deacy loss is 6.68, training accuracy is 0.312500, time 56.237 samples/sec
+epoch 10, total_step 165740, total loss is 12.69 , inference loss is 6.01, weight deacy loss is 6.68, training accuracy is 0.343750, time 56.400 samples/sec
+epoch 10, total_step 165760, total loss is 13.13 , inference loss is 6.45, weight deacy loss is 6.68, training accuracy is 0.531250, time 56.729 samples/sec
+epoch 10, total_step 165780, total loss is 14.63 , inference loss is 7.95, weight deacy loss is 6.68, training accuracy is 0.468750, time 56.216 samples/sec
+epoch 10, total_step 165800, total loss is 16.18 , inference loss is 9.50, weight deacy loss is 6.68, training accuracy is 0.312500, time 56.187 samples/sec
+epoch 10, total_step 165820, total loss is 16.24 , inference loss is 9.56, weight deacy loss is 6.68, training accuracy is 0.437500, time 56.750 samples/sec
+epoch 10, total_step 165840, total loss is 11.09 , inference loss is 4.41, weight deacy loss is 6.68, training accuracy is 0.500000, time 55.976 samples/sec
+epoch 10, total_step 165860, total loss is 15.79 , inference loss is 9.11, weight deacy loss is 6.68, training accuracy is 0.375000, time 56.836 samples/sec
+epoch 10, total_step 165880, total loss is 12.04 , inference loss is 5.37, weight deacy loss is 6.68, training accuracy is 0.531250, time 56.690 samples/sec
+epoch 10, total_step 165900, total loss is 11.81 , inference loss is 5.14, weight deacy loss is 6.68, training accuracy is 0.562500, time 56.220 samples/sec
+epoch 10, total_step 165920, total loss is 14.15 , inference loss is 7.47, weight deacy loss is 6.68, training accuracy is 0.406250, time 56.380 samples/sec
+epoch 10, total_step 165940, total loss is 13.04 , inference loss is 6.36, weight deacy loss is 6.67, training accuracy is 0.468750, time 56.824 samples/sec
+epoch 10, total_step 165960, total loss is 13.03 , inference loss is 6.35, weight deacy loss is 6.67, training accuracy is 0.406250, time 56.442 samples/sec
+epoch 10, total_step 165980, total loss is 14.20 , inference loss is 7.53, weight deacy loss is 6.67, training accuracy is 0.437500, time 56.209 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.77588100000004
+best_threshold_index 107 0.9831481481481481
+best_threshold_index 111 0.9822222222222222
+best_threshold_index 107 0.9833333333333333
+best_threshold_index 107 0.9844444444444445
+best_threshold_index 106 0.9822222222222222
+best_threshold_index 115 0.982037037037037
+best_threshold_index 111 0.9822222222222222
+best_threshold_index 107 0.9829629629629629
+best_threshold_index 115 0.9814814814814815
+best_threshold_index 107 0.9814814814814815
+[lfw][166000]XNorm: 21.826103
+[lfw][166000]Accuracy-Flip: 0.98017+-0.00780
+testing verification..
+(14000, 512)
+infer time 84.34065599999984
+best_threshold_index 118 0.9852380952380952
+best_threshold_index 119 0.984920634920635
+best_threshold_index 119 0.984920634920635
+best_threshold_index 119 0.9844444444444445
+best_threshold_index 118 0.9857142857142858
+best_threshold_index 119 0.9844444444444445
+best_threshold_index 119 0.984920634920635
+best_threshold_index 119 0.9844444444444445
+best_threshold_index 119 0.9842857142857143
+best_threshold_index 119 0.9852380952380952
+[cfp_ff][166000]XNorm: 21.068730
+[cfp_ff][166000]Accuracy-Flip: 0.98457+-0.00432
+testing verification..
+(14000, 512)
+infer time 85.17832499999999
+best_threshold_index 135 0.9230158730158731
+best_threshold_index 135 0.9231746031746032
+best_threshold_index 135 0.9217460317460318
+best_threshold_index 132 0.9257142857142857
+best_threshold_index 132 0.9268253968253968
+best_threshold_index 135 0.9228571428571428
+best_threshold_index 132 0.920952380952381
+best_threshold_index 132 0.9233333333333333
+best_threshold_index 135 0.9234920634920635
+best_threshold_index 138 0.9233333333333333
+[cfp_fp][166000]XNorm: 21.509486
+[cfp_fp][166000]Accuracy-Flip: 0.92029+-0.01518
+testing verification..
+(12000, 512)
+infer time 73.37771000000009
+best_threshold_index 145 0.91
+best_threshold_index 145 0.9070370370370371
+best_threshold_index 145 0.9081481481481481
+best_threshold_index 145 0.9096296296296297
+best_threshold_index 145 0.9133333333333333
+best_threshold_index 145 0.909074074074074
+best_threshold_index 145 0.9085185185185185
+best_threshold_index 145 0.9124074074074074
+best_threshold_index 145 0.9114814814814814
+best_threshold_index 145 0.9103703703703704
+[agedb_30][166000]XNorm: 22.228581
+[agedb_30][166000]Accuracy-Flip: 0.91000+-0.01675
+test accuracy is: 0.9801666666666667
+epoch 10, total_step 166000, total loss is 20.93 , inference loss is 14.26, weight deacy loss is 6.67, training accuracy is 0.281250, time 54.476 samples/sec
+epoch 10, total_step 166020, total loss is 13.98 , inference loss is 7.30, weight deacy loss is 6.67, training accuracy is 0.312500, time 57.339 samples/sec
+epoch 10, total_step 166040, total loss is 16.57 , inference loss is 9.89, weight deacy loss is 6.67, training accuracy is 0.312500, time 57.468 samples/sec
+epoch 10, total_step 166060, total loss is 16.53 , inference loss is 9.86, weight deacy loss is 6.67, training accuracy is 0.437500, time 57.008 samples/sec
+epoch 10, total_step 166080, total loss is 10.74 , inference loss is 4.07, weight deacy loss is 6.67, training accuracy is 0.468750, time 57.151 samples/sec
+epoch 10, total_step 166100, total loss is 16.44 , inference loss is 9.77, weight deacy loss is 6.67, training accuracy is 0.281250, time 56.749 samples/sec
+epoch 10, total_step 166120, total loss is 18.22 , inference loss is 11.55, weight deacy loss is 6.67, training accuracy is 0.281250, time 56.499 samples/sec
+epoch 10, total_step 166140, total loss is 15.85 , inference loss is 9.18, weight deacy loss is 6.67, training accuracy is 0.406250, time 56.151 samples/sec
+epoch 10, total_step 166160, total loss is 13.52 , inference loss is 6.85, weight deacy loss is 6.67, training accuracy is 0.406250, time 56.628 samples/sec
+epoch 10, total_step 166180, total loss is 12.40 , inference loss is 5.73, weight deacy loss is 6.67, training accuracy is 0.375000, time 56.492 samples/sec
+epoch 10, total_step 166200, total loss is 16.59 , inference loss is 9.92, weight deacy loss is 6.67, training accuracy is 0.375000, time 56.008 samples/sec
+epoch 10, total_step 166220, total loss is 11.96 , inference loss is 5.29, weight deacy loss is 6.67, training accuracy is 0.593750, time 56.988 samples/sec
+epoch 10, total_step 166240, total loss is 13.10 , inference loss is 6.43, weight deacy loss is 6.67, training accuracy is 0.343750, time 56.459 samples/sec
+epoch 10, total_step 166260, total loss is 12.33 , inference loss is 5.66, weight deacy loss is 6.67, training accuracy is 0.468750, time 57.635 samples/sec
+epoch 10, total_step 166280, total loss is 18.52 , inference loss is 11.85, weight deacy loss is 6.67, training accuracy is 0.343750, time 56.445 samples/sec
+epoch 10, total_step 166300, total loss is 13.29 , inference loss is 6.62, weight deacy loss is 6.67, training accuracy is 0.343750, time 55.987 samples/sec
+epoch 10, total_step 166320, total loss is 17.44 , inference loss is 10.77, weight deacy loss is 6.67, training accuracy is 0.281250, time 56.525 samples/sec
+epoch 10, total_step 166340, total loss is 14.87 , inference loss is 8.20, weight deacy loss is 6.67, training accuracy is 0.187500, time 56.657 samples/sec
+epoch 10, total_step 166360, total loss is 16.42 , inference loss is 9.75, weight deacy loss is 6.67, training accuracy is 0.343750, time 56.589 samples/sec
+epoch 10, total_step 166380, total loss is 17.89 , inference loss is 11.22, weight deacy loss is 6.67, training accuracy is 0.281250, time 55.757 samples/sec
+epoch 10, total_step 166400, total loss is 17.77 , inference loss is 11.10, weight deacy loss is 6.67, training accuracy is 0.437500, time 56.453 samples/sec
+epoch 10, total_step 166420, total loss is 15.20 , inference loss is 8.53, weight deacy loss is 6.67, training accuracy is 0.312500, time 56.519 samples/sec
+epoch 10, total_step 166440, total loss is 22.03 , inference loss is 15.37, weight deacy loss is 6.67, training accuracy is 0.250000, time 57.057 samples/sec
+epoch 10, total_step 166460, total loss is 17.88 , inference loss is 11.22, weight deacy loss is 6.67, training accuracy is 0.375000, time 56.362 samples/sec
+epoch 10, total_step 166480, total loss is 19.54 , inference loss is 12.87, weight deacy loss is 6.67, training accuracy is 0.312500, time 56.311 samples/sec
+epoch 10, total_step 166500, total loss is 17.77 , inference loss is 11.11, weight deacy loss is 6.67, training accuracy is 0.343750, time 55.315 samples/sec
+epoch 10, total_step 166520, total loss is 16.82 , inference loss is 10.16, weight deacy loss is 6.67, training accuracy is 0.406250, time 56.647 samples/sec
+epoch 10, total_step 166540, total loss is 16.76 , inference loss is 10.09, weight deacy loss is 6.67, training accuracy is 0.343750, time 55.523 samples/sec
+epoch 10, total_step 166560, total loss is 20.13 , inference loss is 13.47, weight deacy loss is 6.67, training accuracy is 0.312500, time 56.252 samples/sec
+epoch 10, total_step 166580, total loss is 14.58 , inference loss is 7.91, weight deacy loss is 6.67, training accuracy is 0.375000, time 57.095 samples/sec
+epoch 10, total_step 166600, total loss is 19.95 , inference loss is 13.28, weight deacy loss is 6.67, training accuracy is 0.187500, time 55.862 samples/sec
+epoch 10, total_step 166620, total loss is 18.23 , inference loss is 11.56, weight deacy loss is 6.67, training accuracy is 0.281250, time 56.127 samples/sec
+epoch 10, total_step 166640, total loss is 14.19 , inference loss is 7.52, weight deacy loss is 6.67, training accuracy is 0.375000, time 56.423 samples/sec
+epoch 10, total_step 166660, total loss is 12.80 , inference loss is 6.14, weight deacy loss is 6.67, training accuracy is 0.406250, time 57.021 samples/sec
+epoch 10, total_step 166680, total loss is 13.96 , inference loss is 7.30, weight deacy loss is 6.67, training accuracy is 0.437500, time 56.551 samples/sec
+epoch 10, total_step 166700, total loss is 14.63 , inference loss is 7.97, weight deacy loss is 6.66, training accuracy is 0.250000, time 56.827 samples/sec
+epoch 10, total_step 166720, total loss is 14.98 , inference loss is 8.32, weight deacy loss is 6.66, training accuracy is 0.312500, time 56.710 samples/sec
+epoch 10, total_step 166740, total loss is 13.59 , inference loss is 6.92, weight deacy loss is 6.66, training accuracy is 0.375000, time 56.045 samples/sec
+epoch 10, total_step 166760, total loss is 17.16 , inference loss is 10.50, weight deacy loss is 6.66, training accuracy is 0.375000, time 56.397 samples/sec
+epoch 10, total_step 166780, total loss is 14.28 , inference loss is 7.62, weight deacy loss is 6.66, training accuracy is 0.281250, time 56.060 samples/sec
+epoch 10, total_step 166800, total loss is 15.08 , inference loss is 8.42, weight deacy loss is 6.66, training accuracy is 0.437500, time 55.795 samples/sec
+epoch 10, total_step 166820, total loss is 13.06 , inference loss is 6.39, weight deacy loss is 6.66, training accuracy is 0.500000, time 56.759 samples/sec
+epoch 10, total_step 166840, total loss is 15.62 , inference loss is 8.96, weight deacy loss is 6.66, training accuracy is 0.343750, time 56.389 samples/sec
+epoch 10, total_step 166860, total loss is 18.33 , inference loss is 11.67, weight deacy loss is 6.66, training accuracy is 0.343750, time 57.045 samples/sec
+epoch 10, total_step 166880, total loss is 16.47 , inference loss is 9.81, weight deacy loss is 6.66, training accuracy is 0.281250, time 56.638 samples/sec
+epoch 10, total_step 166900, total loss is 16.89 , inference loss is 10.22, weight deacy loss is 6.66, training accuracy is 0.375000, time 55.975 samples/sec
+epoch 10, total_step 166920, total loss is 16.86 , inference loss is 10.20, weight deacy loss is 6.66, training accuracy is 0.375000, time 56.610 samples/sec
+epoch 10, total_step 166940, total loss is 12.78 , inference loss is 6.12, weight deacy loss is 6.66, training accuracy is 0.531250, time 56.629 samples/sec
+epoch 10, total_step 166960, total loss is 17.83 , inference loss is 11.17, weight deacy loss is 6.66, training accuracy is 0.312500, time 55.660 samples/sec
+epoch 10, total_step 166980, total loss is 17.96 , inference loss is 11.30, weight deacy loss is 6.66, training accuracy is 0.312500, time 56.685 samples/sec
+epoch 10, total_step 167000, total loss is 13.88 , inference loss is 7.22, weight deacy loss is 6.66, training accuracy is 0.500000, time 56.073 samples/sec
+epoch 10, total_step 167020, total loss is 16.59 , inference loss is 9.93, weight deacy loss is 6.66, training accuracy is 0.375000, time 57.305 samples/sec
+epoch 10, total_step 167040, total loss is 17.25 , inference loss is 10.60, weight deacy loss is 6.66, training accuracy is 0.343750, time 56.229 samples/sec
+epoch 10, total_step 167060, total loss is 19.90 , inference loss is 13.24, weight deacy loss is 6.66, training accuracy is 0.250000, time 55.972 samples/sec
+epoch 10, total_step 167080, total loss is 15.77 , inference loss is 9.11, weight deacy loss is 6.66, training accuracy is 0.156250, time 56.706 samples/sec
+epoch 10, total_step 167100, total loss is 14.73 , inference loss is 8.07, weight deacy loss is 6.66, training accuracy is 0.343750, time 56.525 samples/sec
+epoch 10, total_step 167120, total loss is 15.86 , inference loss is 9.20, weight deacy loss is 6.66, training accuracy is 0.375000, time 55.531 samples/sec
+epoch 10, total_step 167140, total loss is 15.14 , inference loss is 8.48, weight deacy loss is 6.66, training accuracy is 0.375000, time 56.411 samples/sec
+epoch 10, total_step 167160, total loss is 16.39 , inference loss is 9.73, weight deacy loss is 6.66, training accuracy is 0.312500, time 57.081 samples/sec
+epoch 10, total_step 167180, total loss is 12.05 , inference loss is 5.39, weight deacy loss is 6.66, training accuracy is 0.593750, time 57.315 samples/sec
+epoch 10, total_step 167200, total loss is 15.34 , inference loss is 8.68, weight deacy loss is 6.66, training accuracy is 0.312500, time 57.369 samples/sec
+epoch 10, total_step 167220, total loss is 16.23 , inference loss is 9.57, weight deacy loss is 6.66, training accuracy is 0.437500, time 56.594 samples/sec
+epoch 10, total_step 167240, total loss is 10.24 , inference loss is 3.58, weight deacy loss is 6.66, training accuracy is 0.562500, time 56.858 samples/sec
+epoch 10, total_step 167260, total loss is 14.05 , inference loss is 7.39, weight deacy loss is 6.66, training accuracy is 0.281250, time 56.476 samples/sec
+epoch 10, total_step 167280, total loss is 17.77 , inference loss is 11.11, weight deacy loss is 6.66, training accuracy is 0.312500, time 56.338 samples/sec
+epoch 10, total_step 167300, total loss is 14.32 , inference loss is 7.67, weight deacy loss is 6.66, training accuracy is 0.437500, time 56.483 samples/sec
+epoch 10, total_step 167320, total loss is 16.23 , inference loss is 9.58, weight deacy loss is 6.66, training accuracy is 0.343750, time 56.538 samples/sec
+epoch 10, total_step 167340, total loss is 17.47 , inference loss is 10.82, weight deacy loss is 6.66, training accuracy is 0.343750, time 57.069 samples/sec
+epoch 10, total_step 167360, total loss is 22.39 , inference loss is 15.74, weight deacy loss is 6.66, training accuracy is 0.250000, time 56.074 samples/sec
+epoch 10, total_step 167380, total loss is 12.52 , inference loss is 5.87, weight deacy loss is 6.65, training accuracy is 0.468750, time 56.317 samples/sec
+epoch 10, total_step 167400, total loss is 15.02 , inference loss is 8.37, weight deacy loss is 6.65, training accuracy is 0.375000, time 56.756 samples/sec
+epoch 10, total_step 167420, total loss is 19.77 , inference loss is 13.12, weight deacy loss is 6.65, training accuracy is 0.250000, time 56.367 samples/sec
+epoch 10, total_step 167440, total loss is 13.36 , inference loss is 6.71, weight deacy loss is 6.65, training accuracy is 0.437500, time 56.585 samples/sec
+epoch 10, total_step 167460, total loss is 12.60 , inference loss is 5.95, weight deacy loss is 6.65, training accuracy is 0.437500, time 57.100 samples/sec
+epoch 10, total_step 167480, total loss is 15.35 , inference loss is 8.69, weight deacy loss is 6.65, training accuracy is 0.312500, time 56.425 samples/sec
+epoch 10, total_step 167500, total loss is 19.22 , inference loss is 12.56, weight deacy loss is 6.65, training accuracy is 0.281250, time 56.084 samples/sec
+epoch 10, total_step 167520, total loss is 13.13 , inference loss is 6.47, weight deacy loss is 6.65, training accuracy is 0.468750, time 56.794 samples/sec
+epoch 10, total_step 167540, total loss is 19.13 , inference loss is 12.48, weight deacy loss is 6.65, training accuracy is 0.312500, time 56.579 samples/sec
+epoch 10, total_step 167560, total loss is 17.20 , inference loss is 10.55, weight deacy loss is 6.65, training accuracy is 0.250000, time 56.424 samples/sec
+epoch 10, total_step 167580, total loss is 21.46 , inference loss is 14.81, weight deacy loss is 6.65, training accuracy is 0.281250, time 56.237 samples/sec
+epoch 10, total_step 167600, total loss is 12.16 , inference loss is 5.51, weight deacy loss is 6.65, training accuracy is 0.406250, time 56.647 samples/sec
+epoch 10, total_step 167620, total loss is 14.81 , inference loss is 8.16, weight deacy loss is 6.65, training accuracy is 0.375000, time 56.433 samples/sec
+epoch 10, total_step 167640, total loss is 14.86 , inference loss is 8.21, weight deacy loss is 6.65, training accuracy is 0.468750, time 56.008 samples/sec
+epoch 10, total_step 167660, total loss is 16.22 , inference loss is 9.57, weight deacy loss is 6.65, training accuracy is 0.343750, time 56.198 samples/sec
+epoch 10, total_step 167680, total loss is 16.34 , inference loss is 9.69, weight deacy loss is 6.65, training accuracy is 0.406250, time 55.746 samples/sec
+epoch 10, total_step 167700, total loss is 16.78 , inference loss is 10.13, weight deacy loss is 6.65, training accuracy is 0.250000, time 56.457 samples/sec
+epoch 10, total_step 167720, total loss is 18.81 , inference loss is 12.16, weight deacy loss is 6.65, training accuracy is 0.312500, time 56.094 samples/sec
+epoch 10, total_step 167740, total loss is 15.31 , inference loss is 8.66, weight deacy loss is 6.65, training accuracy is 0.343750, time 55.581 samples/sec
+epoch 10, total_step 167760, total loss is 17.86 , inference loss is 11.21, weight deacy loss is 6.65, training accuracy is 0.250000, time 57.452 samples/sec
+epoch 10, total_step 167780, total loss is 12.01 , inference loss is 5.36, weight deacy loss is 6.65, training accuracy is 0.468750, time 55.964 samples/sec
+epoch 10, total_step 167800, total loss is 20.58 , inference loss is 13.93, weight deacy loss is 6.65, training accuracy is 0.312500, time 56.186 samples/sec
+epoch 10, total_step 167820, total loss is 13.18 , inference loss is 6.53, weight deacy loss is 6.65, training accuracy is 0.406250, time 56.535 samples/sec
+epoch 10, total_step 167840, total loss is 13.40 , inference loss is 6.75, weight deacy loss is 6.65, training accuracy is 0.312500, time 55.073 samples/sec
+epoch 10, total_step 167860, total loss is 15.62 , inference loss is 8.97, weight deacy loss is 6.65, training accuracy is 0.406250, time 56.481 samples/sec
+epoch 10, total_step 167880, total loss is 14.96 , inference loss is 8.32, weight deacy loss is 6.65, training accuracy is 0.437500, time 56.674 samples/sec
+epoch 10, total_step 167900, total loss is 19.29 , inference loss is 12.64, weight deacy loss is 6.65, training accuracy is 0.156250, time 55.667 samples/sec
+epoch 10, total_step 167920, total loss is 17.15 , inference loss is 10.50, weight deacy loss is 6.65, training accuracy is 0.281250, time 55.461 samples/sec
+epoch 10, total_step 167940, total loss is 17.67 , inference loss is 11.02, weight deacy loss is 6.65, training accuracy is 0.406250, time 56.910 samples/sec
+epoch 10, total_step 167960, total loss is 14.07 , inference loss is 7.43, weight deacy loss is 6.65, training accuracy is 0.312500, time 56.877 samples/sec
+epoch 10, total_step 167980, total loss is 16.62 , inference loss is 9.97, weight deacy loss is 6.65, training accuracy is 0.250000, time 56.104 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.15531300000009
+best_threshold_index 111 0.9874074074074074
+best_threshold_index 115 0.9859259259259259
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 106 0.9872222222222222
+best_threshold_index 111 0.9864814814814815
+best_threshold_index 115 0.9857407407407407
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 111 0.9864814814814815
+best_threshold_index 111 0.985
+best_threshold_index 111 0.9859259259259259
+[lfw][168000]XNorm: 22.352192
+[lfw][168000]Accuracy-Flip: 0.98500+-0.00742
+testing verification..
+(14000, 512)
+infer time 84.75487300000006
+best_threshold_index 123 0.9879365079365079
+best_threshold_index 123 0.9874603174603175
+best_threshold_index 123 0.9876190476190476
+best_threshold_index 123 0.9871428571428571
+best_threshold_index 123 0.9887301587301587
+best_threshold_index 123 0.9874603174603175
+best_threshold_index 123 0.9882539682539683
+best_threshold_index 123 0.9874603174603175
+best_threshold_index 123 0.9874603174603175
+best_threshold_index 123 0.9876190476190476
+[cfp_ff][168000]XNorm: 21.989889
+[cfp_ff][168000]Accuracy-Flip: 0.98771+-0.00400
+testing verification..
+(14000, 512)
+infer time 84.99083
+best_threshold_index 137 0.9266666666666666
+best_threshold_index 142 0.9268253968253968
+best_threshold_index 142 0.9265079365079365
+best_threshold_index 137 0.9288888888888889
+best_threshold_index 138 0.9293650793650794
+best_threshold_index 137 0.9265079365079365
+best_threshold_index 142 0.9241269841269841
+best_threshold_index 143 0.9258730158730158
+best_threshold_index 142 0.9274603174603174
+best_threshold_index 141 0.927936507936508
+[cfp_fp][168000]XNorm: 21.792232
+[cfp_fp][168000]Accuracy-Flip: 0.92429+-0.01319
+testing verification..
+(12000, 512)
+infer time 73.04480200000012
+best_threshold_index 141 0.9061111111111111
+best_threshold_index 141 0.9042592592592592
+best_threshold_index 141 0.9057407407407407
+best_threshold_index 141 0.9083333333333333
+best_threshold_index 141 0.9120370370370371
+best_threshold_index 141 0.9057407407407407
+best_threshold_index 141 0.9088888888888889
+best_threshold_index 141 0.9127777777777778
+best_threshold_index 141 0.9101851851851852
+best_threshold_index 141 0.9092592592592592
+[agedb_30][168000]XNorm: 22.866451
+[agedb_30][168000]Accuracy-Flip: 0.90833+-0.02432
+test accuracy is: 0.985
+epoch 10, total_step 168000, total loss is 12.87 , inference loss is 6.22, weight deacy loss is 6.65, training accuracy is 0.500000, time 53.725 samples/sec
+epoch 10, total_step 168020, total loss is 19.01 , inference loss is 12.37, weight deacy loss is 6.65, training accuracy is 0.250000, time 55.979 samples/sec
+epoch 10, total_step 168040, total loss is 14.63 , inference loss is 7.99, weight deacy loss is 6.65, training accuracy is 0.437500, time 56.710 samples/sec
+epoch 10, total_step 168060, total loss is 17.44 , inference loss is 10.79, weight deacy loss is 6.65, training accuracy is 0.375000, time 55.735 samples/sec
+epoch 10, total_step 168080, total loss is 16.87 , inference loss is 10.23, weight deacy loss is 6.65, training accuracy is 0.156250, time 56.920 samples/sec
+epoch 10, total_step 168100, total loss is 15.62 , inference loss is 8.97, weight deacy loss is 6.65, training accuracy is 0.156250, time 56.485 samples/sec
+epoch 10, total_step 168120, total loss is 15.20 , inference loss is 8.55, weight deacy loss is 6.64, training accuracy is 0.218750, time 56.740 samples/sec
+epoch 10, total_step 168140, total loss is 14.31 , inference loss is 7.67, weight deacy loss is 6.64, training accuracy is 0.281250, time 56.601 samples/sec
+epoch 10, total_step 168160, total loss is 17.13 , inference loss is 10.49, weight deacy loss is 6.64, training accuracy is 0.343750, time 55.363 samples/sec
+epoch 10, total_step 168180, total loss is 14.31 , inference loss is 7.67, weight deacy loss is 6.64, training accuracy is 0.406250, time 55.165 samples/sec
+epoch 10, total_step 168200, total loss is 16.42 , inference loss is 9.77, weight deacy loss is 6.64, training accuracy is 0.187500, time 56.424 samples/sec
+epoch 10, total_step 168220, total loss is 17.94 , inference loss is 11.30, weight deacy loss is 6.64, training accuracy is 0.250000, time 56.031 samples/sec
+epoch 10, total_step 168240, total loss is 16.30 , inference loss is 9.65, weight deacy loss is 6.64, training accuracy is 0.281250, time 58.083 samples/sec
+epoch 10, total_step 168260, total loss is 17.50 , inference loss is 10.86, weight deacy loss is 6.64, training accuracy is 0.250000, time 56.292 samples/sec
+epoch 10, total_step 168280, total loss is 15.16 , inference loss is 8.52, weight deacy loss is 6.64, training accuracy is 0.218750, time 55.877 samples/sec
+epoch 10, total_step 168300, total loss is 21.52 , inference loss is 14.88, weight deacy loss is 6.64, training accuracy is 0.218750, time 56.728 samples/sec
+epoch 10, total_step 168320, total loss is 16.51 , inference loss is 9.86, weight deacy loss is 6.64, training accuracy is 0.187500, time 56.445 samples/sec
+epoch 10, total_step 168340, total loss is 16.74 , inference loss is 10.10, weight deacy loss is 6.64, training accuracy is 0.218750, time 56.609 samples/sec
+epoch 10, total_step 168360, total loss is 13.88 , inference loss is 7.24, weight deacy loss is 6.64, training accuracy is 0.375000, time 57.338 samples/sec
+epoch 10, total_step 168380, total loss is 18.34 , inference loss is 11.70, weight deacy loss is 6.64, training accuracy is 0.312500, time 56.440 samples/sec
+epoch 10, total_step 168400, total loss is 12.69 , inference loss is 6.05, weight deacy loss is 6.64, training accuracy is 0.406250, time 56.278 samples/sec
+epoch 10, total_step 168420, total loss is 16.17 , inference loss is 9.53, weight deacy loss is 6.64, training accuracy is 0.500000, time 55.992 samples/sec
+epoch 10, total_step 168440, total loss is 15.52 , inference loss is 8.88, weight deacy loss is 6.64, training accuracy is 0.375000, time 56.859 samples/sec
+epoch 10, total_step 168460, total loss is 13.12 , inference loss is 6.48, weight deacy loss is 6.64, training accuracy is 0.375000, time 57.051 samples/sec
+epoch 10, total_step 168480, total loss is 16.48 , inference loss is 9.84, weight deacy loss is 6.64, training accuracy is 0.281250, time 55.649 samples/sec
+epoch 10, total_step 168500, total loss is 14.58 , inference loss is 7.94, weight deacy loss is 6.64, training accuracy is 0.312500, time 56.035 samples/sec
+epoch 10, total_step 168520, total loss is 13.76 , inference loss is 7.12, weight deacy loss is 6.64, training accuracy is 0.250000, time 56.856 samples/sec
+epoch 10, total_step 168540, total loss is 19.95 , inference loss is 13.31, weight deacy loss is 6.64, training accuracy is 0.343750, time 56.060 samples/sec
+epoch 10, total_step 168560, total loss is 15.38 , inference loss is 8.75, weight deacy loss is 6.64, training accuracy is 0.406250, time 56.070 samples/sec
+epoch 10, total_step 168580, total loss is 16.74 , inference loss is 10.11, weight deacy loss is 6.64, training accuracy is 0.375000, time 57.619 samples/sec
+epoch 10, total_step 168600, total loss is 17.98 , inference loss is 11.34, weight deacy loss is 6.64, training accuracy is 0.343750, time 56.375 samples/sec
+epoch 10, total_step 168620, total loss is 16.02 , inference loss is 9.38, weight deacy loss is 6.64, training accuracy is 0.343750, time 56.692 samples/sec
+epoch 10, total_step 168640, total loss is 15.71 , inference loss is 9.07, weight deacy loss is 6.64, training accuracy is 0.343750, time 56.469 samples/sec
+
+
+
+
+End of epoch 10
+epoch 11, total_step 168660, total loss is 16.68 , inference loss is 10.05, weight deacy loss is 6.64, training accuracy is 0.187500, time 55.758 samples/sec
+epoch 11, total_step 168680, total loss is 14.02 , inference loss is 7.39, weight deacy loss is 6.64, training accuracy is 0.468750, time 57.370 samples/sec
+epoch 11, total_step 168700, total loss is 14.17 , inference loss is 7.53, weight deacy loss is 6.64, training accuracy is 0.437500, time 56.034 samples/sec
+epoch 11, total_step 168720, total loss is 13.75 , inference loss is 7.11, weight deacy loss is 6.64, training accuracy is 0.625000, time 55.634 samples/sec
+epoch 11, total_step 168740, total loss is 13.78 , inference loss is 7.14, weight deacy loss is 6.64, training accuracy is 0.593750, time 55.834 samples/sec
+epoch 11, total_step 168760, total loss is 9.37 , inference loss is 2.73, weight deacy loss is 6.64, training accuracy is 0.718750, time 55.883 samples/sec
+epoch 11, total_step 168780, total loss is 8.39 , inference loss is 1.75, weight deacy loss is 6.64, training accuracy is 0.718750, time 56.319 samples/sec
+epoch 11, total_step 168800, total loss is 12.04 , inference loss is 5.40, weight deacy loss is 6.64, training accuracy is 0.687500, time 56.329 samples/sec
+epoch 11, total_step 168820, total loss is 11.39 , inference loss is 4.75, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.932 samples/sec
+epoch 11, total_step 168840, total loss is 11.67 , inference loss is 5.04, weight deacy loss is 6.63, training accuracy is 0.656250, time 55.498 samples/sec
+epoch 11, total_step 168860, total loss is 12.17 , inference loss is 5.54, weight deacy loss is 6.63, training accuracy is 0.687500, time 55.907 samples/sec
+epoch 11, total_step 168880, total loss is 12.43 , inference loss is 5.80, weight deacy loss is 6.63, training accuracy is 0.593750, time 55.024 samples/sec
+epoch 11, total_step 168900, total loss is 10.47 , inference loss is 3.84, weight deacy loss is 6.63, training accuracy is 0.625000, time 55.656 samples/sec
+epoch 11, total_step 168920, total loss is 9.82 , inference loss is 3.19, weight deacy loss is 6.63, training accuracy is 0.687500, time 56.505 samples/sec
+epoch 11, total_step 168940, total loss is 10.53 , inference loss is 3.90, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.203 samples/sec
+epoch 11, total_step 168960, total loss is 11.40 , inference loss is 4.77, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.578 samples/sec
+epoch 11, total_step 168980, total loss is 14.09 , inference loss is 7.46, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.363 samples/sec
+epoch 11, total_step 169000, total loss is 14.33 , inference loss is 7.69, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.584 samples/sec
+epoch 11, total_step 169020, total loss is 11.75 , inference loss is 5.12, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.293 samples/sec
+epoch 11, total_step 169040, total loss is 11.56 , inference loss is 4.93, weight deacy loss is 6.63, training accuracy is 0.656250, time 55.841 samples/sec
+epoch 11, total_step 169060, total loss is 9.93 , inference loss is 3.30, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.470 samples/sec
+epoch 11, total_step 169080, total loss is 12.06 , inference loss is 5.43, weight deacy loss is 6.63, training accuracy is 0.656250, time 57.435 samples/sec
+epoch 11, total_step 169100, total loss is 10.73 , inference loss is 4.10, weight deacy loss is 6.63, training accuracy is 0.656250, time 57.378 samples/sec
+epoch 11, total_step 169120, total loss is 10.01 , inference loss is 3.38, weight deacy loss is 6.63, training accuracy is 0.750000, time 56.465 samples/sec
+epoch 11, total_step 169140, total loss is 12.16 , inference loss is 5.53, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.978 samples/sec
+epoch 11, total_step 169160, total loss is 15.16 , inference loss is 8.53, weight deacy loss is 6.63, training accuracy is 0.531250, time 56.824 samples/sec
+epoch 11, total_step 169180, total loss is 15.08 , inference loss is 8.45, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.849 samples/sec
+epoch 11, total_step 169200, total loss is 9.84 , inference loss is 3.21, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.571 samples/sec
+epoch 11, total_step 169220, total loss is 10.77 , inference loss is 4.14, weight deacy loss is 6.63, training accuracy is 0.812500, time 56.916 samples/sec
+epoch 11, total_step 169240, total loss is 18.05 , inference loss is 11.42, weight deacy loss is 6.63, training accuracy is 0.468750, time 56.863 samples/sec
+epoch 11, total_step 169260, total loss is 10.62 , inference loss is 3.99, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.575 samples/sec
+epoch 11, total_step 169280, total loss is 9.95 , inference loss is 3.32, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.565 samples/sec
+epoch 11, total_step 169300, total loss is 15.42 , inference loss is 8.79, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.109 samples/sec
+epoch 11, total_step 169320, total loss is 15.55 , inference loss is 8.92, weight deacy loss is 6.63, training accuracy is 0.531250, time 56.425 samples/sec
+epoch 11, total_step 169340, total loss is 10.41 , inference loss is 3.77, weight deacy loss is 6.63, training accuracy is 0.718750, time 56.339 samples/sec
+epoch 11, total_step 169360, total loss is 10.61 , inference loss is 3.98, weight deacy loss is 6.63, training accuracy is 0.750000, time 56.662 samples/sec
+epoch 11, total_step 169380, total loss is 14.36 , inference loss is 7.73, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.612 samples/sec
+epoch 11, total_step 169400, total loss is 13.34 , inference loss is 6.71, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.128 samples/sec
+epoch 11, total_step 169420, total loss is 11.23 , inference loss is 4.59, weight deacy loss is 6.63, training accuracy is 0.625000, time 57.162 samples/sec
+epoch 11, total_step 169440, total loss is 16.26 , inference loss is 9.63, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.803 samples/sec
+epoch 11, total_step 169460, total loss is 12.33 , inference loss is 5.70, weight deacy loss is 6.63, training accuracy is 0.593750, time 57.042 samples/sec
+epoch 11, total_step 169480, total loss is 14.87 , inference loss is 8.24, weight deacy loss is 6.63, training accuracy is 0.468750, time 56.993 samples/sec
+epoch 11, total_step 169500, total loss is 12.30 , inference loss is 5.66, weight deacy loss is 6.63, training accuracy is 0.656250, time 56.645 samples/sec
+epoch 11, total_step 169520, total loss is 14.02 , inference loss is 7.39, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.825 samples/sec
+epoch 11, total_step 169540, total loss is 14.25 , inference loss is 7.62, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.951 samples/sec
+epoch 11, total_step 169560, total loss is 10.41 , inference loss is 3.78, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.719 samples/sec
+epoch 11, total_step 169580, total loss is 11.50 , inference loss is 4.86, weight deacy loss is 6.63, training accuracy is 0.593750, time 55.957 samples/sec
+epoch 11, total_step 169600, total loss is 10.68 , inference loss is 4.05, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.612 samples/sec
+epoch 11, total_step 169620, total loss is 13.68 , inference loss is 7.05, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.495 samples/sec
+epoch 11, total_step 169640, total loss is 19.01 , inference loss is 12.38, weight deacy loss is 6.63, training accuracy is 0.406250, time 56.365 samples/sec
+epoch 11, total_step 169660, total loss is 15.36 , inference loss is 8.73, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.387 samples/sec
+epoch 11, total_step 169680, total loss is 17.31 , inference loss is 10.68, weight deacy loss is 6.63, training accuracy is 0.468750, time 56.925 samples/sec
+epoch 11, total_step 169700, total loss is 11.40 , inference loss is 4.76, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.814 samples/sec
+epoch 11, total_step 169720, total loss is 11.43 , inference loss is 4.80, weight deacy loss is 6.63, training accuracy is 0.562500, time 55.990 samples/sec
+epoch 11, total_step 169740, total loss is 13.05 , inference loss is 6.42, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.873 samples/sec
+epoch 11, total_step 169760, total loss is 11.83 , inference loss is 5.20, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.250 samples/sec
+epoch 11, total_step 169780, total loss is 15.65 , inference loss is 9.02, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.903 samples/sec
+epoch 11, total_step 169800, total loss is 20.38 , inference loss is 13.75, weight deacy loss is 6.63, training accuracy is 0.312500, time 55.681 samples/sec
+epoch 11, total_step 169820, total loss is 15.35 , inference loss is 8.71, weight deacy loss is 6.63, training accuracy is 0.562500, time 55.312 samples/sec
+epoch 11, total_step 169840, total loss is 14.77 , inference loss is 8.14, weight deacy loss is 6.63, training accuracy is 0.406250, time 56.558 samples/sec
+epoch 11, total_step 169860, total loss is 12.37 , inference loss is 5.74, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.577 samples/sec
+epoch 11, total_step 169880, total loss is 12.04 , inference loss is 5.40, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.812 samples/sec
+epoch 11, total_step 169900, total loss is 10.66 , inference loss is 4.03, weight deacy loss is 6.63, training accuracy is 0.531250, time 56.291 samples/sec
+epoch 11, total_step 169920, total loss is 14.20 , inference loss is 7.57, weight deacy loss is 6.63, training accuracy is 0.468750, time 56.622 samples/sec
+epoch 11, total_step 169940, total loss is 13.19 , inference loss is 6.56, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.693 samples/sec
+epoch 11, total_step 169960, total loss is 12.22 , inference loss is 5.58, weight deacy loss is 6.63, training accuracy is 0.531250, time 56.338 samples/sec
+epoch 11, total_step 169980, total loss is 14.76 , inference loss is 8.12, weight deacy loss is 6.63, training accuracy is 0.343750, time 56.942 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.84193700000004
+best_threshold_index 120 0.9837037037037037
+best_threshold_index 120 0.9838888888888889
+best_threshold_index 120 0.9846296296296296
+best_threshold_index 120 0.9838888888888889
+best_threshold_index 120 0.9840740740740741
+best_threshold_index 120 0.9835185185185186
+best_threshold_index 120 0.9851851851851852
+best_threshold_index 120 0.9848148148148148
+best_threshold_index 120 0.9831481481481481
+best_threshold_index 120 0.9831481481481481
+[lfw][170000]XNorm: 19.137640
+[lfw][170000]Accuracy-Flip: 0.98400+-0.00588
+testing verification..
+(14000, 512)
+infer time 86.45471300000003
+best_threshold_index 121 0.9812698412698413
+best_threshold_index 123 0.9817460317460317
+best_threshold_index 121 0.9814285714285714
+best_threshold_index 121 0.9812698412698413
+best_threshold_index 121 0.9806349206349206
+best_threshold_index 121 0.9796825396825397
+best_threshold_index 121 0.9809523809523809
+best_threshold_index 121 0.9804761904761905
+best_threshold_index 119 0.9806349206349206
+best_threshold_index 121 0.9825396825396825
+[cfp_ff][170000]XNorm: 18.628470
+[cfp_ff][170000]Accuracy-Flip: 0.98000+-0.00691
+testing verification..
+(14000, 512)
+infer time 86.37122499999994
+best_threshold_index 141 0.9157142857142857
+best_threshold_index 141 0.9160317460317461
+best_threshold_index 141 0.9147619047619048
+best_threshold_index 141 0.9180952380952381
+best_threshold_index 141 0.9176190476190477
+best_threshold_index 141 0.9122222222222223
+best_threshold_index 141 0.9141269841269841
+best_threshold_index 141 0.9128571428571428
+best_threshold_index 141 0.9157142857142857
+best_threshold_index 141 0.9171428571428571
+[cfp_fp][170000]XNorm: 19.844217
+[cfp_fp][170000]Accuracy-Flip: 0.91543+-0.01669
+testing verification..
+(12000, 512)
+infer time 73.98580899999997
+best_threshold_index 143 0.9074074074074074
+best_threshold_index 143 0.9048148148148148
+best_threshold_index 143 0.9057407407407407
+best_threshold_index 143 0.9070370370370371
+best_threshold_index 143 0.914074074074074
+best_threshold_index 143 0.9081481481481481
+best_threshold_index 143 0.9075925925925926
+best_threshold_index 143 0.9083333333333333
+best_threshold_index 144 0.9092592592592592
+best_threshold_index 143 0.9081481481481481
+[agedb_30][170000]XNorm: 20.287898
+[agedb_30][170000]Accuracy-Flip: 0.90733+-0.02126
+test accuracy is: 0.9840000000000002
+epoch 11, total_step 170000, total loss is 14.30 , inference loss is 7.67, weight deacy loss is 6.63, training accuracy is 0.625000, time 53.303 samples/sec
+epoch 11, total_step 170020, total loss is 13.58 , inference loss is 6.95, weight deacy loss is 6.63, training accuracy is 0.468750, time 56.557 samples/sec
+epoch 11, total_step 170040, total loss is 11.42 , inference loss is 4.79, weight deacy loss is 6.63, training accuracy is 0.593750, time 57.017 samples/sec
+epoch 11, total_step 170060, total loss is 10.22 , inference loss is 3.58, weight deacy loss is 6.63, training accuracy is 0.718750, time 56.817 samples/sec
+epoch 11, total_step 170080, total loss is 16.08 , inference loss is 9.45, weight deacy loss is 6.63, training accuracy is 0.437500, time 56.077 samples/sec
+epoch 11, total_step 170100, total loss is 11.67 , inference loss is 5.03, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.508 samples/sec
+epoch 11, total_step 170120, total loss is 13.46 , inference loss is 6.82, weight deacy loss is 6.63, training accuracy is 0.468750, time 57.009 samples/sec
+epoch 11, total_step 170140, total loss is 16.25 , inference loss is 9.61, weight deacy loss is 6.63, training accuracy is 0.375000, time 56.689 samples/sec
+epoch 11, total_step 170160, total loss is 12.54 , inference loss is 5.90, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.531 samples/sec
+epoch 11, total_step 170180, total loss is 13.64 , inference loss is 7.01, weight deacy loss is 6.63, training accuracy is 0.531250, time 57.132 samples/sec
+epoch 11, total_step 170200, total loss is 9.93 , inference loss is 3.30, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.551 samples/sec
+epoch 11, total_step 170220, total loss is 17.60 , inference loss is 10.97, weight deacy loss is 6.64, training accuracy is 0.343750, time 56.708 samples/sec
+epoch 11, total_step 170240, total loss is 13.13 , inference loss is 6.50, weight deacy loss is 6.64, training accuracy is 0.562500, time 56.844 samples/sec
+epoch 11, total_step 170260, total loss is 15.38 , inference loss is 8.74, weight deacy loss is 6.64, training accuracy is 0.562500, time 56.135 samples/sec
+epoch 11, total_step 170280, total loss is 14.58 , inference loss is 7.94, weight deacy loss is 6.64, training accuracy is 0.468750, time 55.981 samples/sec
+epoch 11, total_step 170300, total loss is 13.94 , inference loss is 7.31, weight deacy loss is 6.64, training accuracy is 0.500000, time 57.518 samples/sec
+epoch 11, total_step 170320, total loss is 13.74 , inference loss is 7.10, weight deacy loss is 6.64, training accuracy is 0.468750, time 57.118 samples/sec
+epoch 11, total_step 170340, total loss is 13.95 , inference loss is 7.31, weight deacy loss is 6.64, training accuracy is 0.562500, time 56.140 samples/sec
+epoch 11, total_step 170360, total loss is 14.49 , inference loss is 7.86, weight deacy loss is 6.64, training accuracy is 0.437500, time 56.712 samples/sec
+epoch 11, total_step 170380, total loss is 14.81 , inference loss is 8.17, weight deacy loss is 6.64, training accuracy is 0.500000, time 56.698 samples/sec
+epoch 11, total_step 170400, total loss is 13.75 , inference loss is 7.12, weight deacy loss is 6.64, training accuracy is 0.500000, time 56.262 samples/sec
+epoch 11, total_step 170420, total loss is 11.32 , inference loss is 4.68, weight deacy loss is 6.64, training accuracy is 0.625000, time 56.680 samples/sec
+epoch 11, total_step 170440, total loss is 13.97 , inference loss is 7.33, weight deacy loss is 6.64, training accuracy is 0.531250, time 55.180 samples/sec
+epoch 11, total_step 170460, total loss is 12.58 , inference loss is 5.94, weight deacy loss is 6.64, training accuracy is 0.531250, time 55.613 samples/sec
+epoch 11, total_step 170480, total loss is 12.64 , inference loss is 6.00, weight deacy loss is 6.64, training accuracy is 0.312500, time 57.646 samples/sec
+epoch 11, total_step 170500, total loss is 14.55 , inference loss is 7.91, weight deacy loss is 6.64, training accuracy is 0.437500, time 56.875 samples/sec
+epoch 11, total_step 170520, total loss is 10.33 , inference loss is 3.69, weight deacy loss is 6.64, training accuracy is 0.781250, time 57.128 samples/sec
+epoch 11, total_step 170540, total loss is 13.66 , inference loss is 7.03, weight deacy loss is 6.64, training accuracy is 0.500000, time 56.273 samples/sec
+epoch 11, total_step 170560, total loss is 15.47 , inference loss is 8.83, weight deacy loss is 6.63, training accuracy is 0.656250, time 56.417 samples/sec
+epoch 11, total_step 170580, total loss is 11.11 , inference loss is 4.48, weight deacy loss is 6.63, training accuracy is 0.437500, time 56.502 samples/sec
+epoch 11, total_step 170600, total loss is 10.81 , inference loss is 4.18, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.001 samples/sec
+epoch 11, total_step 170620, total loss is 14.92 , inference loss is 8.29, weight deacy loss is 6.63, training accuracy is 0.437500, time 56.517 samples/sec
+epoch 11, total_step 170640, total loss is 15.87 , inference loss is 9.23, weight deacy loss is 6.63, training accuracy is 0.468750, time 54.964 samples/sec
+epoch 11, total_step 170660, total loss is 12.68 , inference loss is 6.05, weight deacy loss is 6.63, training accuracy is 0.531250, time 56.548 samples/sec
+epoch 11, total_step 170680, total loss is 12.70 , inference loss is 6.07, weight deacy loss is 6.63, training accuracy is 0.656250, time 56.480 samples/sec
+epoch 11, total_step 170700, total loss is 11.00 , inference loss is 4.37, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.788 samples/sec
+epoch 11, total_step 170720, total loss is 14.35 , inference loss is 7.72, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.000 samples/sec
+epoch 11, total_step 170740, total loss is 14.79 , inference loss is 8.16, weight deacy loss is 6.63, training accuracy is 0.343750, time 57.680 samples/sec
+epoch 11, total_step 170760, total loss is 10.81 , inference loss is 4.18, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.370 samples/sec
+epoch 11, total_step 170780, total loss is 11.36 , inference loss is 4.73, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.266 samples/sec
+epoch 11, total_step 170800, total loss is 10.09 , inference loss is 3.45, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.305 samples/sec
+epoch 11, total_step 170820, total loss is 12.38 , inference loss is 5.74, weight deacy loss is 6.63, training accuracy is 0.531250, time 55.895 samples/sec
+epoch 11, total_step 170840, total loss is 14.07 , inference loss is 7.44, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.419 samples/sec
+epoch 11, total_step 170860, total loss is 13.26 , inference loss is 6.63, weight deacy loss is 6.63, training accuracy is 0.562500, time 55.555 samples/sec
+epoch 11, total_step 170880, total loss is 12.37 , inference loss is 5.73, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.657 samples/sec
+epoch 11, total_step 170900, total loss is 9.60 , inference loss is 2.96, weight deacy loss is 6.63, training accuracy is 0.687500, time 56.321 samples/sec
+epoch 11, total_step 170920, total loss is 13.18 , inference loss is 6.55, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.107 samples/sec
+epoch 11, total_step 170940, total loss is 11.96 , inference loss is 5.33, weight deacy loss is 6.63, training accuracy is 0.656250, time 55.934 samples/sec
+epoch 11, total_step 170960, total loss is 12.36 , inference loss is 5.73, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.458 samples/sec
+epoch 11, total_step 170980, total loss is 10.31 , inference loss is 3.68, weight deacy loss is 6.63, training accuracy is 0.562500, time 55.448 samples/sec
+epoch 11, total_step 171000, total loss is 8.74 , inference loss is 2.11, weight deacy loss is 6.63, training accuracy is 0.656250, time 56.271 samples/sec
+epoch 11, total_step 171020, total loss is 11.29 , inference loss is 4.66, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.414 samples/sec
+epoch 11, total_step 171040, total loss is 9.70 , inference loss is 3.07, weight deacy loss is 6.63, training accuracy is 0.593750, time 56.149 samples/sec
+epoch 11, total_step 171060, total loss is 11.85 , inference loss is 5.22, weight deacy loss is 6.63, training accuracy is 0.625000, time 57.026 samples/sec
+epoch 11, total_step 171080, total loss is 11.54 , inference loss is 4.91, weight deacy loss is 6.63, training accuracy is 0.531250, time 56.941 samples/sec
+epoch 11, total_step 171100, total loss is 11.49 , inference loss is 4.86, weight deacy loss is 6.63, training accuracy is 0.750000, time 57.323 samples/sec
+epoch 11, total_step 171120, total loss is 15.17 , inference loss is 8.54, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.486 samples/sec
+epoch 11, total_step 171140, total loss is 11.25 , inference loss is 4.62, weight deacy loss is 6.63, training accuracy is 0.625000, time 55.642 samples/sec
+epoch 11, total_step 171160, total loss is 14.55 , inference loss is 7.92, weight deacy loss is 6.63, training accuracy is 0.437500, time 55.769 samples/sec
+epoch 11, total_step 171180, total loss is 9.25 , inference loss is 2.62, weight deacy loss is 6.63, training accuracy is 0.718750, time 56.019 samples/sec
+epoch 11, total_step 171200, total loss is 13.59 , inference loss is 6.96, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.231 samples/sec
+epoch 11, total_step 171220, total loss is 12.74 , inference loss is 6.11, weight deacy loss is 6.63, training accuracy is 0.468750, time 57.057 samples/sec
+epoch 11, total_step 171240, total loss is 15.12 , inference loss is 8.50, weight deacy loss is 6.63, training accuracy is 0.468750, time 56.186 samples/sec
+epoch 11, total_step 171260, total loss is 11.49 , inference loss is 4.86, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.699 samples/sec
+epoch 11, total_step 171280, total loss is 9.60 , inference loss is 2.97, weight deacy loss is 6.63, training accuracy is 0.687500, time 56.350 samples/sec
+epoch 11, total_step 171300, total loss is 10.90 , inference loss is 4.28, weight deacy loss is 6.63, training accuracy is 0.625000, time 56.118 samples/sec
+epoch 11, total_step 171320, total loss is 13.45 , inference loss is 6.82, weight deacy loss is 6.63, training accuracy is 0.500000, time 56.458 samples/sec
+epoch 11, total_step 171340, total loss is 11.18 , inference loss is 4.56, weight deacy loss is 6.63, training accuracy is 0.656250, time 56.613 samples/sec
+epoch 11, total_step 171360, total loss is 12.53 , inference loss is 5.90, weight deacy loss is 6.63, training accuracy is 0.531250, time 56.188 samples/sec
+epoch 11, total_step 171380, total loss is 12.77 , inference loss is 6.15, weight deacy loss is 6.63, training accuracy is 0.656250, time 57.273 samples/sec
+epoch 11, total_step 171400, total loss is 13.95 , inference loss is 7.32, weight deacy loss is 6.63, training accuracy is 0.562500, time 56.763 samples/sec
+epoch 11, total_step 171420, total loss is 10.41 , inference loss is 3.78, weight deacy loss is 6.63, training accuracy is 0.656250, time 55.631 samples/sec
+epoch 11, total_step 171440, total loss is 12.72 , inference loss is 6.10, weight deacy loss is 6.63, training accuracy is 0.656250, time 56.540 samples/sec
+epoch 11, total_step 171460, total loss is 9.59 , inference loss is 2.97, weight deacy loss is 6.63, training accuracy is 0.750000, time 56.054 samples/sec
+epoch 11, total_step 171480, total loss is 12.20 , inference loss is 5.57, weight deacy loss is 6.63, training accuracy is 0.593750, time 55.758 samples/sec
+epoch 11, total_step 171500, total loss is 12.45 , inference loss is 5.83, weight deacy loss is 6.62, training accuracy is 0.562500, time 56.573 samples/sec
+epoch 11, total_step 171520, total loss is 10.26 , inference loss is 3.63, weight deacy loss is 6.62, training accuracy is 0.687500, time 56.692 samples/sec
+epoch 11, total_step 171540, total loss is 18.15 , inference loss is 11.53, weight deacy loss is 6.62, training accuracy is 0.375000, time 57.113 samples/sec
+epoch 11, total_step 171560, total loss is 11.24 , inference loss is 4.62, weight deacy loss is 6.62, training accuracy is 0.625000, time 56.790 samples/sec
+epoch 11, total_step 171580, total loss is 10.95 , inference loss is 4.32, weight deacy loss is 6.62, training accuracy is 0.656250, time 57.176 samples/sec
+epoch 11, total_step 171600, total loss is 12.63 , inference loss is 6.00, weight deacy loss is 6.62, training accuracy is 0.593750, time 56.766 samples/sec
+epoch 11, total_step 171620, total loss is 12.80 , inference loss is 6.18, weight deacy loss is 6.62, training accuracy is 0.375000, time 56.674 samples/sec
+epoch 11, total_step 171640, total loss is 12.10 , inference loss is 5.48, weight deacy loss is 6.62, training accuracy is 0.406250, time 57.250 samples/sec
+epoch 11, total_step 171660, total loss is 13.21 , inference loss is 6.59, weight deacy loss is 6.62, training accuracy is 0.656250, time 57.044 samples/sec
+epoch 11, total_step 171680, total loss is 11.82 , inference loss is 5.20, weight deacy loss is 6.62, training accuracy is 0.687500, time 56.635 samples/sec
+epoch 11, total_step 171700, total loss is 13.71 , inference loss is 7.09, weight deacy loss is 6.62, training accuracy is 0.562500, time 55.257 samples/sec
+epoch 11, total_step 171720, total loss is 14.74 , inference loss is 8.12, weight deacy loss is 6.62, training accuracy is 0.500000, time 55.196 samples/sec
+epoch 11, total_step 171740, total loss is 10.07 , inference loss is 3.44, weight deacy loss is 6.62, training accuracy is 0.656250, time 57.120 samples/sec
+epoch 11, total_step 171760, total loss is 12.13 , inference loss is 5.51, weight deacy loss is 6.62, training accuracy is 0.687500, time 56.530 samples/sec
+epoch 11, total_step 171780, total loss is 16.73 , inference loss is 10.10, weight deacy loss is 6.62, training accuracy is 0.500000, time 56.474 samples/sec
+epoch 11, total_step 171800, total loss is 10.42 , inference loss is 3.80, weight deacy loss is 6.62, training accuracy is 0.500000, time 56.148 samples/sec
+epoch 11, total_step 171820, total loss is 18.87 , inference loss is 12.25, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.475 samples/sec
+epoch 11, total_step 171840, total loss is 15.07 , inference loss is 8.45, weight deacy loss is 6.62, training accuracy is 0.468750, time 55.753 samples/sec
+epoch 11, total_step 171860, total loss is 13.44 , inference loss is 6.81, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.554 samples/sec
+epoch 11, total_step 171880, total loss is 11.94 , inference loss is 5.32, weight deacy loss is 6.62, training accuracy is 0.593750, time 56.249 samples/sec
+epoch 11, total_step 171900, total loss is 10.49 , inference loss is 3.87, weight deacy loss is 6.62, training accuracy is 0.656250, time 57.289 samples/sec
+epoch 11, total_step 171920, total loss is 15.16 , inference loss is 8.53, weight deacy loss is 6.62, training accuracy is 0.375000, time 56.173 samples/sec
+epoch 11, total_step 171940, total loss is 14.28 , inference loss is 7.65, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.802 samples/sec
+epoch 11, total_step 171960, total loss is 14.86 , inference loss is 8.24, weight deacy loss is 6.62, training accuracy is 0.406250, time 56.401 samples/sec
+epoch 11, total_step 171980, total loss is 14.03 , inference loss is 7.40, weight deacy loss is 6.62, training accuracy is 0.531250, time 57.243 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.58953999999996
+best_threshold_index 121 0.9877777777777778
+best_threshold_index 121 0.9868518518518519
+best_threshold_index 120 0.9874074074074074
+best_threshold_index 121 0.9874074074074074
+best_threshold_index 121 0.9874074074074074
+best_threshold_index 121 0.987037037037037
+best_threshold_index 121 0.9877777777777778
+best_threshold_index 121 0.9872222222222222
+best_threshold_index 121 0.9862962962962963
+best_threshold_index 121 0.9864814814814815
+[lfw][172000]XNorm: 20.795145
+[lfw][172000]Accuracy-Flip: 0.98683+-0.00456
+testing verification..
+(14000, 512)
+infer time 86.15889899999998
+best_threshold_index 119 0.9876190476190476
+best_threshold_index 119 0.986984126984127
+best_threshold_index 119 0.9871428571428571
+best_threshold_index 119 0.9866666666666667
+best_threshold_index 119 0.9880952380952381
+best_threshold_index 124 0.9871428571428571
+best_threshold_index 119 0.9882539682539683
+best_threshold_index 119 0.9874603174603175
+best_threshold_index 122 0.986984126984127
+best_threshold_index 119 0.9884126984126984
+[cfp_ff][172000]XNorm: 20.465811
+[cfp_ff][172000]Accuracy-Flip: 0.98629+-0.00483
+testing verification..
+(14000, 512)
+infer time 85.55145099999984
+best_threshold_index 144 0.9222222222222223
+best_threshold_index 143 0.922063492063492
+best_threshold_index 143 0.9217460317460318
+best_threshold_index 143 0.9223809523809524
+best_threshold_index 142 0.9242857142857143
+best_threshold_index 143 0.9203174603174603
+best_threshold_index 143 0.9203174603174603
+best_threshold_index 142 0.9204761904761904
+best_threshold_index 143 0.9212698412698412
+best_threshold_index 143 0.9222222222222223
+[cfp_fp][172000]XNorm: 21.208623
+[cfp_fp][172000]Accuracy-Flip: 0.92086+-0.01083
+testing verification..
+(12000, 512)
+infer time 73.86035000000001
+best_threshold_index 147 0.9014814814814814
+best_threshold_index 147 0.899074074074074
+best_threshold_index 147 0.8996296296296297
+best_threshold_index 147 0.9014814814814814
+best_threshold_index 147 0.9038888888888889
+best_threshold_index 147 0.9018518518518519
+best_threshold_index 147 0.9016666666666666
+best_threshold_index 147 0.9057407407407407
+best_threshold_index 147 0.9074074074074074
+best_threshold_index 147 0.9027777777777778
+[agedb_30][172000]XNorm: 21.736171
+[agedb_30][172000]Accuracy-Flip: 0.90250+-0.02200
+test accuracy is: 0.9868333333333335
+epoch 11, total_step 172000, total loss is 13.31 , inference loss is 6.69, weight deacy loss is 6.62, training accuracy is 0.343750, time 53.485 samples/sec
+epoch 11, total_step 172020, total loss is 12.65 , inference loss is 6.03, weight deacy loss is 6.62, training accuracy is 0.593750, time 56.616 samples/sec
+epoch 11, total_step 172040, total loss is 11.99 , inference loss is 5.37, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.834 samples/sec
+epoch 11, total_step 172060, total loss is 12.71 , inference loss is 6.08, weight deacy loss is 6.62, training accuracy is 0.562500, time 56.211 samples/sec
+epoch 11, total_step 172080, total loss is 16.35 , inference loss is 9.73, weight deacy loss is 6.62, training accuracy is 0.218750, time 56.770 samples/sec
+epoch 11, total_step 172100, total loss is 11.19 , inference loss is 4.56, weight deacy loss is 6.62, training accuracy is 0.718750, time 56.338 samples/sec
+epoch 11, total_step 172120, total loss is 11.83 , inference loss is 5.21, weight deacy loss is 6.62, training accuracy is 0.593750, time 55.969 samples/sec
+epoch 11, total_step 172140, total loss is 13.02 , inference loss is 6.40, weight deacy loss is 6.62, training accuracy is 0.437500, time 56.263 samples/sec
+epoch 11, total_step 172160, total loss is 14.79 , inference loss is 8.17, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.942 samples/sec
+epoch 11, total_step 172180, total loss is 15.62 , inference loss is 9.00, weight deacy loss is 6.62, training accuracy is 0.562500, time 56.691 samples/sec
+epoch 11, total_step 172200, total loss is 15.75 , inference loss is 9.13, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.330 samples/sec
+epoch 11, total_step 172220, total loss is 14.46 , inference loss is 7.83, weight deacy loss is 6.62, training accuracy is 0.562500, time 56.638 samples/sec
+epoch 11, total_step 172240, total loss is 17.09 , inference loss is 10.47, weight deacy loss is 6.62, training accuracy is 0.500000, time 56.423 samples/sec
+epoch 11, total_step 172260, total loss is 12.70 , inference loss is 6.08, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.757 samples/sec
+epoch 11, total_step 172280, total loss is 13.72 , inference loss is 7.10, weight deacy loss is 6.62, training accuracy is 0.406250, time 56.424 samples/sec
+epoch 11, total_step 172300, total loss is 11.15 , inference loss is 4.53, weight deacy loss is 6.62, training accuracy is 0.500000, time 56.914 samples/sec
+epoch 11, total_step 172320, total loss is 14.82 , inference loss is 8.20, weight deacy loss is 6.62, training accuracy is 0.406250, time 56.753 samples/sec
+epoch 11, total_step 172340, total loss is 10.33 , inference loss is 3.71, weight deacy loss is 6.62, training accuracy is 0.500000, time 57.453 samples/sec
+epoch 11, total_step 172360, total loss is 14.44 , inference loss is 7.82, weight deacy loss is 6.62, training accuracy is 0.593750, time 55.425 samples/sec
+epoch 11, total_step 172380, total loss is 16.12 , inference loss is 9.50, weight deacy loss is 6.62, training accuracy is 0.531250, time 57.402 samples/sec
+epoch 11, total_step 172400, total loss is 12.97 , inference loss is 6.35, weight deacy loss is 6.62, training accuracy is 0.593750, time 56.684 samples/sec
+epoch 11, total_step 172420, total loss is 17.75 , inference loss is 11.13, weight deacy loss is 6.62, training accuracy is 0.312500, time 56.490 samples/sec
+epoch 11, total_step 172440, total loss is 12.85 , inference loss is 6.24, weight deacy loss is 6.62, training accuracy is 0.406250, time 56.327 samples/sec
+epoch 11, total_step 172460, total loss is 15.74 , inference loss is 9.12, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.771 samples/sec
+epoch 11, total_step 172480, total loss is 17.84 , inference loss is 11.22, weight deacy loss is 6.62, training accuracy is 0.468750, time 56.102 samples/sec
+epoch 11, total_step 172500, total loss is 13.36 , inference loss is 6.74, weight deacy loss is 6.62, training accuracy is 0.531250, time 56.115 samples/sec
+epoch 11, total_step 172520, total loss is 14.85 , inference loss is 8.23, weight deacy loss is 6.62, training accuracy is 0.437500, time 56.550 samples/sec
+epoch 11, total_step 172540, total loss is 14.80 , inference loss is 8.18, weight deacy loss is 6.62, training accuracy is 0.406250, time 56.741 samples/sec
+epoch 11, total_step 172560, total loss is 14.16 , inference loss is 7.54, weight deacy loss is 6.62, training accuracy is 0.531250, time 55.950 samples/sec
+epoch 11, total_step 172580, total loss is 15.21 , inference loss is 8.59, weight deacy loss is 6.62, training accuracy is 0.343750, time 56.895 samples/sec
+epoch 11, total_step 172600, total loss is 15.63 , inference loss is 9.01, weight deacy loss is 6.62, training accuracy is 0.531250, time 56.377 samples/sec
+epoch 11, total_step 172620, total loss is 15.39 , inference loss is 8.77, weight deacy loss is 6.62, training accuracy is 0.437500, time 56.627 samples/sec
+epoch 11, total_step 172640, total loss is 10.77 , inference loss is 4.15, weight deacy loss is 6.62, training accuracy is 0.687500, time 56.189 samples/sec
+epoch 11, total_step 172660, total loss is 15.33 , inference loss is 8.71, weight deacy loss is 6.62, training accuracy is 0.343750, time 56.200 samples/sec
+epoch 11, total_step 172680, total loss is 15.28 , inference loss is 8.66, weight deacy loss is 6.62, training accuracy is 0.406250, time 57.122 samples/sec
+epoch 11, total_step 172700, total loss is 15.32 , inference loss is 8.70, weight deacy loss is 6.62, training accuracy is 0.375000, time 55.683 samples/sec
+epoch 11, total_step 172720, total loss is 13.29 , inference loss is 6.68, weight deacy loss is 6.62, training accuracy is 0.500000, time 56.897 samples/sec
+epoch 11, total_step 172740, total loss is 16.86 , inference loss is 10.24, weight deacy loss is 6.62, training accuracy is 0.437500, time 56.103 samples/sec
+epoch 11, total_step 172760, total loss is 10.96 , inference loss is 4.34, weight deacy loss is 6.62, training accuracy is 0.500000, time 57.188 samples/sec
+epoch 11, total_step 172780, total loss is 12.85 , inference loss is 6.23, weight deacy loss is 6.62, training accuracy is 0.625000, time 56.544 samples/sec
+epoch 11, total_step 172800, total loss is 14.71 , inference loss is 8.09, weight deacy loss is 6.62, training accuracy is 0.531250, time 56.561 samples/sec
+epoch 11, total_step 172820, total loss is 11.65 , inference loss is 5.03, weight deacy loss is 6.62, training accuracy is 0.625000, time 56.488 samples/sec
+epoch 11, total_step 172840, total loss is 14.06 , inference loss is 7.44, weight deacy loss is 6.62, training accuracy is 0.500000, time 56.249 samples/sec
+epoch 11, total_step 172860, total loss is 13.16 , inference loss is 6.55, weight deacy loss is 6.62, training accuracy is 0.468750, time 57.343 samples/sec
+epoch 11, total_step 172880, total loss is 12.89 , inference loss is 6.27, weight deacy loss is 6.62, training accuracy is 0.437500, time 56.775 samples/sec
+epoch 11, total_step 172900, total loss is 17.66 , inference loss is 11.05, weight deacy loss is 6.62, training accuracy is 0.281250, time 56.379 samples/sec
+epoch 11, total_step 172920, total loss is 12.27 , inference loss is 5.66, weight deacy loss is 6.62, training accuracy is 0.375000, time 56.926 samples/sec
+epoch 11, total_step 172940, total loss is 17.40 , inference loss is 10.78, weight deacy loss is 6.62, training accuracy is 0.281250, time 56.471 samples/sec
+epoch 11, total_step 172960, total loss is 14.40 , inference loss is 7.79, weight deacy loss is 6.62, training accuracy is 0.406250, time 56.971 samples/sec
+epoch 11, total_step 172980, total loss is 13.97 , inference loss is 7.35, weight deacy loss is 6.61, training accuracy is 0.437500, time 56.609 samples/sec
+epoch 11, total_step 173000, total loss is 15.81 , inference loss is 9.19, weight deacy loss is 6.61, training accuracy is 0.562500, time 56.312 samples/sec
+epoch 11, total_step 173020, total loss is 15.48 , inference loss is 8.87, weight deacy loss is 6.61, training accuracy is 0.437500, time 56.431 samples/sec
+epoch 11, total_step 173040, total loss is 15.20 , inference loss is 8.59, weight deacy loss is 6.61, training accuracy is 0.406250, time 57.165 samples/sec
+epoch 11, total_step 173060, total loss is 16.59 , inference loss is 9.98, weight deacy loss is 6.61, training accuracy is 0.375000, time 56.493 samples/sec
+epoch 11, total_step 173080, total loss is 16.78 , inference loss is 10.16, weight deacy loss is 6.61, training accuracy is 0.468750, time 56.989 samples/sec
+epoch 11, total_step 173100, total loss is 17.37 , inference loss is 10.76, weight deacy loss is 6.61, training accuracy is 0.406250, time 56.641 samples/sec
+epoch 11, total_step 173120, total loss is 14.17 , inference loss is 7.55, weight deacy loss is 6.61, training accuracy is 0.562500, time 56.371 samples/sec
+epoch 11, total_step 173140, total loss is 15.09 , inference loss is 8.48, weight deacy loss is 6.61, training accuracy is 0.437500, time 57.682 samples/sec
+epoch 11, total_step 173160, total loss is 15.24 , inference loss is 8.62, weight deacy loss is 6.61, training accuracy is 0.406250, time 56.930 samples/sec
+epoch 11, total_step 173180, total loss is 12.39 , inference loss is 5.77, weight deacy loss is 6.61, training accuracy is 0.531250, time 56.449 samples/sec
+epoch 11, total_step 173200, total loss is 9.42 , inference loss is 2.81, weight deacy loss is 6.61, training accuracy is 0.750000, time 56.252 samples/sec
+epoch 11, total_step 173220, total loss is 13.17 , inference loss is 6.56, weight deacy loss is 6.61, training accuracy is 0.500000, time 56.280 samples/sec
+epoch 11, total_step 173240, total loss is 16.88 , inference loss is 10.26, weight deacy loss is 6.61, training accuracy is 0.281250, time 57.463 samples/sec
+epoch 11, total_step 173260, total loss is 14.57 , inference loss is 7.96, weight deacy loss is 6.61, training accuracy is 0.468750, time 56.553 samples/sec
+epoch 11, total_step 173280, total loss is 14.77 , inference loss is 8.16, weight deacy loss is 6.61, training accuracy is 0.406250, time 56.177 samples/sec
+epoch 11, total_step 173300, total loss is 14.94 , inference loss is 8.33, weight deacy loss is 6.61, training accuracy is 0.500000, time 55.793 samples/sec
+epoch 11, total_step 173320, total loss is 15.87 , inference loss is 9.26, weight deacy loss is 6.61, training accuracy is 0.562500, time 57.284 samples/sec
+epoch 11, total_step 173340, total loss is 12.61 , inference loss is 6.00, weight deacy loss is 6.61, training accuracy is 0.468750, time 57.049 samples/sec
+epoch 11, total_step 173360, total loss is 19.29 , inference loss is 12.68, weight deacy loss is 6.61, training accuracy is 0.375000, time 56.747 samples/sec
+epoch 11, total_step 173380, total loss is 12.07 , inference loss is 5.46, weight deacy loss is 6.61, training accuracy is 0.500000, time 55.411 samples/sec
+epoch 11, total_step 173400, total loss is 14.87 , inference loss is 8.26, weight deacy loss is 6.61, training accuracy is 0.531250, time 56.438 samples/sec
+epoch 11, total_step 173420, total loss is 15.16 , inference loss is 8.55, weight deacy loss is 6.61, training accuracy is 0.531250, time 56.474 samples/sec
+epoch 11, total_step 173440, total loss is 12.64 , inference loss is 6.03, weight deacy loss is 6.61, training accuracy is 0.531250, time 56.851 samples/sec
+epoch 11, total_step 173460, total loss is 15.85 , inference loss is 9.24, weight deacy loss is 6.61, training accuracy is 0.562500, time 55.964 samples/sec
+epoch 11, total_step 173480, total loss is 14.43 , inference loss is 7.82, weight deacy loss is 6.61, training accuracy is 0.437500, time 55.790 samples/sec
+epoch 11, total_step 173500, total loss is 14.61 , inference loss is 8.00, weight deacy loss is 6.61, training accuracy is 0.406250, time 55.392 samples/sec
+epoch 11, total_step 173520, total loss is 15.07 , inference loss is 8.46, weight deacy loss is 6.61, training accuracy is 0.343750, time 56.200 samples/sec
+epoch 11, total_step 173540, total loss is 14.38 , inference loss is 7.77, weight deacy loss is 6.61, training accuracy is 0.406250, time 57.271 samples/sec
+epoch 11, total_step 173560, total loss is 13.73 , inference loss is 7.12, weight deacy loss is 6.61, training accuracy is 0.437500, time 57.084 samples/sec
+epoch 11, total_step 173580, total loss is 14.98 , inference loss is 8.37, weight deacy loss is 6.61, training accuracy is 0.656250, time 57.089 samples/sec
+epoch 11, total_step 173600, total loss is 17.34 , inference loss is 10.74, weight deacy loss is 6.61, training accuracy is 0.343750, time 55.601 samples/sec
+epoch 11, total_step 173620, total loss is 13.19 , inference loss is 6.58, weight deacy loss is 6.61, training accuracy is 0.375000, time 57.078 samples/sec
+epoch 11, total_step 173640, total loss is 12.38 , inference loss is 5.77, weight deacy loss is 6.61, training accuracy is 0.500000, time 56.208 samples/sec
+epoch 11, total_step 173660, total loss is 13.18 , inference loss is 6.57, weight deacy loss is 6.61, training accuracy is 0.500000, time 55.926 samples/sec
+epoch 11, total_step 173680, total loss is 11.61 , inference loss is 5.00, weight deacy loss is 6.61, training accuracy is 0.500000, time 56.003 samples/sec
+epoch 11, total_step 173700, total loss is 11.63 , inference loss is 5.02, weight deacy loss is 6.61, training accuracy is 0.593750, time 55.736 samples/sec
+epoch 11, total_step 173720, total loss is 13.46 , inference loss is 6.85, weight deacy loss is 6.61, training accuracy is 0.468750, time 57.238 samples/sec
+epoch 11, total_step 173740, total loss is 13.53 , inference loss is 6.92, weight deacy loss is 6.61, training accuracy is 0.500000, time 57.024 samples/sec
+epoch 11, total_step 173760, total loss is 14.78 , inference loss is 8.17, weight deacy loss is 6.61, training accuracy is 0.531250, time 57.060 samples/sec
+epoch 11, total_step 173780, total loss is 11.72 , inference loss is 5.11, weight deacy loss is 6.61, training accuracy is 0.562500, time 56.605 samples/sec
+epoch 11, total_step 173800, total loss is 13.85 , inference loss is 7.24, weight deacy loss is 6.61, training accuracy is 0.468750, time 55.861 samples/sec
+epoch 11, total_step 173820, total loss is 13.52 , inference loss is 6.91, weight deacy loss is 6.61, training accuracy is 0.468750, time 56.330 samples/sec
+epoch 11, total_step 173840, total loss is 14.28 , inference loss is 7.67, weight deacy loss is 6.61, training accuracy is 0.562500, time 56.735 samples/sec
+epoch 11, total_step 173860, total loss is 11.56 , inference loss is 4.95, weight deacy loss is 6.61, training accuracy is 0.437500, time 56.551 samples/sec
+epoch 11, total_step 173880, total loss is 12.40 , inference loss is 5.79, weight deacy loss is 6.61, training accuracy is 0.562500, time 56.657 samples/sec
+epoch 11, total_step 173900, total loss is 14.44 , inference loss is 7.84, weight deacy loss is 6.61, training accuracy is 0.406250, time 56.623 samples/sec
+epoch 11, total_step 173920, total loss is 11.25 , inference loss is 4.64, weight deacy loss is 6.61, training accuracy is 0.468750, time 56.519 samples/sec
+epoch 11, total_step 173940, total loss is 11.77 , inference loss is 5.16, weight deacy loss is 6.61, training accuracy is 0.468750, time 56.976 samples/sec
+epoch 11, total_step 173960, total loss is 14.02 , inference loss is 7.41, weight deacy loss is 6.61, training accuracy is 0.437500, time 56.436 samples/sec
+epoch 11, total_step 173980, total loss is 12.90 , inference loss is 6.30, weight deacy loss is 6.61, training accuracy is 0.500000, time 57.135 samples/sec
+testing verification..
+(12000, 512)
+infer time 74.04180899999993
+best_threshold_index 112 0.9859259259259259
+best_threshold_index 115 0.9848148148148148
+best_threshold_index 113 0.9851851851851852
+best_threshold_index 113 0.9857407407407407
+best_threshold_index 112 0.985
+best_threshold_index 113 0.9840740740740741
+best_threshold_index 113 0.9851851851851852
+best_threshold_index 113 0.9862962962962963
+best_threshold_index 113 0.9838888888888889
+best_threshold_index 113 0.9842592592592593
+[lfw][174000]XNorm: 21.642534
+[lfw][174000]Accuracy-Flip: 0.98417+-0.00704
+testing verification..
+(14000, 512)
+infer time 85.95645800000003
+best_threshold_index 127 0.9865079365079366
+best_threshold_index 127 0.986031746031746
+best_threshold_index 127 0.9858730158730159
+best_threshold_index 126 0.9855555555555555
+best_threshold_index 127 0.9861904761904762
+best_threshold_index 127 0.9850793650793651
+best_threshold_index 124 0.986984126984127
+best_threshold_index 127 0.9853968253968254
+best_threshold_index 127 0.9863492063492063
+best_threshold_index 127 0.9874603174603175
+[cfp_ff][174000]XNorm: 21.438427
+[cfp_ff][174000]Accuracy-Flip: 0.98557+-0.00671
+testing verification..
+(14000, 512)
+infer time 85.61919800000008
+best_threshold_index 142 0.9226984126984127
+best_threshold_index 142 0.9223809523809524
+best_threshold_index 142 0.9215873015873016
+best_threshold_index 142 0.9257142857142857
+best_threshold_index 142 0.9252380952380952
+best_threshold_index 142 0.9203174603174603
+best_threshold_index 142 0.9215873015873016
+best_threshold_index 142 0.9219047619047619
+best_threshold_index 142 0.9236507936507936
+best_threshold_index 142 0.9234920634920635
+[cfp_fp][174000]XNorm: 21.898762
+[cfp_fp][174000]Accuracy-Flip: 0.92286+-0.01443
+testing verification..
+(12000, 512)
+infer time 73.85910199999994
+best_threshold_index 139 0.9051851851851852
+best_threshold_index 142 0.9031481481481481
+best_threshold_index 142 0.9046296296296297
+best_threshold_index 143 0.9046296296296297
+best_threshold_index 139 0.9118518518518518
+best_threshold_index 142 0.9025925925925926
+best_threshold_index 139 0.904074074074074
+best_threshold_index 139 0.9096296296296297
+best_threshold_index 139 0.9081481481481481
+best_threshold_index 139 0.905
+[agedb_30][174000]XNorm: 22.479513
+[agedb_30][174000]Accuracy-Flip: 0.90183+-0.02447
+test accuracy is: 0.9841666666666666
+epoch 11, total_step 174000, total loss is 12.83 , inference loss is 6.22, weight deacy loss is 6.61, training accuracy is 0.687500, time 54.337 samples/sec
+epoch 11, total_step 174020, total loss is 12.55 , inference loss is 5.94, weight deacy loss is 6.61, training accuracy is 0.437500, time 56.652 samples/sec
+epoch 11, total_step 174040, total loss is 13.13 , inference loss is 6.52, weight deacy loss is 6.61, training accuracy is 0.531250, time 56.905 samples/sec
+epoch 11, total_step 174060, total loss is 10.73 , inference loss is 4.13, weight deacy loss is 6.61, training accuracy is 0.468750, time 56.130 samples/sec
+epoch 11, total_step 174080, total loss is 12.60 , inference loss is 6.00, weight deacy loss is 6.61, training accuracy is 0.593750, time 56.055 samples/sec
+epoch 11, total_step 174100, total loss is 12.29 , inference loss is 5.69, weight deacy loss is 6.61, training accuracy is 0.531250, time 56.626 samples/sec
+epoch 11, total_step 174120, total loss is 11.29 , inference loss is 4.68, weight deacy loss is 6.61, training accuracy is 0.437500, time 56.994 samples/sec
+epoch 11, total_step 174140, total loss is 16.68 , inference loss is 10.07, weight deacy loss is 6.61, training accuracy is 0.406250, time 56.367 samples/sec
+epoch 11, total_step 174160, total loss is 15.16 , inference loss is 8.55, weight deacy loss is 6.61, training accuracy is 0.375000, time 56.929 samples/sec
+epoch 11, total_step 174180, total loss is 15.26 , inference loss is 8.66, weight deacy loss is 6.60, training accuracy is 0.437500, time 55.684 samples/sec
+epoch 11, total_step 174200, total loss is 14.36 , inference loss is 7.75, weight deacy loss is 6.60, training accuracy is 0.343750, time 56.881 samples/sec
+epoch 11, total_step 174220, total loss is 15.79 , inference loss is 9.18, weight deacy loss is 6.60, training accuracy is 0.375000, time 56.888 samples/sec
+epoch 11, total_step 174240, total loss is 15.75 , inference loss is 9.15, weight deacy loss is 6.60, training accuracy is 0.468750, time 56.369 samples/sec
+epoch 11, total_step 174260, total loss is 16.65 , inference loss is 10.05, weight deacy loss is 6.60, training accuracy is 0.437500, time 56.427 samples/sec
+epoch 11, total_step 174280, total loss is 12.03 , inference loss is 5.43, weight deacy loss is 6.60, training accuracy is 0.468750, time 56.314 samples/sec
+epoch 11, total_step 174300, total loss is 12.04 , inference loss is 5.44, weight deacy loss is 6.60, training accuracy is 0.562500, time 56.632 samples/sec
+epoch 11, total_step 174320, total loss is 13.69 , inference loss is 7.08, weight deacy loss is 6.60, training accuracy is 0.437500, time 56.300 samples/sec
+epoch 11, total_step 174340, total loss is 11.15 , inference loss is 4.54, weight deacy loss is 6.60, training accuracy is 0.625000, time 54.609 samples/sec
+epoch 11, total_step 174360, total loss is 13.67 , inference loss is 7.07, weight deacy loss is 6.60, training accuracy is 0.593750, time 56.602 samples/sec
+epoch 11, total_step 174380, total loss is 14.50 , inference loss is 7.89, weight deacy loss is 6.60, training accuracy is 0.406250, time 56.616 samples/sec
+epoch 11, total_step 174400, total loss is 12.39 , inference loss is 5.79, weight deacy loss is 6.60, training accuracy is 0.531250, time 57.118 samples/sec
+epoch 11, total_step 174420, total loss is 12.57 , inference loss is 5.96, weight deacy loss is 6.60, training accuracy is 0.531250, time 57.158 samples/sec
+epoch 11, total_step 174440, total loss is 10.18 , inference loss is 3.58, weight deacy loss is 6.60, training accuracy is 0.593750, time 56.238 samples/sec
+epoch 11, total_step 174460, total loss is 11.75 , inference loss is 5.15, weight deacy loss is 6.60, training accuracy is 0.562500, time 56.649 samples/sec
+epoch 11, total_step 174480, total loss is 13.36 , inference loss is 6.76, weight deacy loss is 6.60, training accuracy is 0.625000, time 55.784 samples/sec
+epoch 11, total_step 174500, total loss is 12.84 , inference loss is 6.24, weight deacy loss is 6.60, training accuracy is 0.406250, time 56.192 samples/sec
+epoch 11, total_step 174520, total loss is 15.89 , inference loss is 9.29, weight deacy loss is 6.60, training accuracy is 0.437500, time 56.166 samples/sec
+epoch 11, total_step 174540, total loss is 13.28 , inference loss is 6.68, weight deacy loss is 6.60, training accuracy is 0.437500, time 56.525 samples/sec
+epoch 11, total_step 174560, total loss is 14.90 , inference loss is 8.30, weight deacy loss is 6.60, training accuracy is 0.375000, time 57.043 samples/sec
+epoch 11, total_step 174580, total loss is 15.45 , inference loss is 8.85, weight deacy loss is 6.60, training accuracy is 0.312500, time 54.533 samples/sec
+epoch 11, total_step 174600, total loss is 14.42 , inference loss is 7.82, weight deacy loss is 6.60, training accuracy is 0.468750, time 55.501 samples/sec
+epoch 11, total_step 174620, total loss is 16.58 , inference loss is 9.98, weight deacy loss is 6.60, training accuracy is 0.281250, time 56.767 samples/sec
+epoch 11, total_step 174640, total loss is 13.71 , inference loss is 7.11, weight deacy loss is 6.60, training accuracy is 0.437500, time 56.627 samples/sec
+epoch 11, total_step 174660, total loss is 16.11 , inference loss is 9.51, weight deacy loss is 6.60, training accuracy is 0.312500, time 56.182 samples/sec
+epoch 11, total_step 174680, total loss is 15.29 , inference loss is 8.69, weight deacy loss is 6.60, training accuracy is 0.437500, time 56.979 samples/sec
+epoch 11, total_step 174700, total loss is 12.13 , inference loss is 5.53, weight deacy loss is 6.60, training accuracy is 0.562500, time 56.775 samples/sec
+epoch 11, total_step 174720, total loss is 17.04 , inference loss is 10.44, weight deacy loss is 6.60, training accuracy is 0.375000, time 56.180 samples/sec
+epoch 11, total_step 174740, total loss is 15.89 , inference loss is 9.29, weight deacy loss is 6.60, training accuracy is 0.375000, time 56.084 samples/sec
+epoch 11, total_step 174760, total loss is 13.35 , inference loss is 6.75, weight deacy loss is 6.60, training accuracy is 0.375000, time 56.669 samples/sec
+epoch 11, total_step 174780, total loss is 12.46 , inference loss is 5.86, weight deacy loss is 6.60, training accuracy is 0.625000, time 56.747 samples/sec
+epoch 11, total_step 174800, total loss is 13.27 , inference loss is 6.67, weight deacy loss is 6.60, training accuracy is 0.312500, time 56.679 samples/sec
+epoch 11, total_step 174820, total loss is 16.05 , inference loss is 9.45, weight deacy loss is 6.60, training accuracy is 0.343750, time 56.465 samples/sec
+epoch 11, total_step 174840, total loss is 12.21 , inference loss is 5.61, weight deacy loss is 6.60, training accuracy is 0.687500, time 56.379 samples/sec
+epoch 11, total_step 174860, total loss is 15.97 , inference loss is 9.38, weight deacy loss is 6.60, training accuracy is 0.531250, time 55.493 samples/sec
+epoch 11, total_step 174880, total loss is 12.37 , inference loss is 5.77, weight deacy loss is 6.60, training accuracy is 0.593750, time 56.048 samples/sec
+epoch 11, total_step 174900, total loss is 15.10 , inference loss is 8.50, weight deacy loss is 6.60, training accuracy is 0.406250, time 55.814 samples/sec
+epoch 11, total_step 174920, total loss is 12.00 , inference loss is 5.40, weight deacy loss is 6.60, training accuracy is 0.468750, time 56.915 samples/sec
+epoch 11, total_step 174940, total loss is 15.71 , inference loss is 9.11, weight deacy loss is 6.60, training accuracy is 0.531250, time 56.509 samples/sec
+epoch 11, total_step 174960, total loss is 15.22 , inference loss is 8.62, weight deacy loss is 6.60, training accuracy is 0.437500, time 55.190 samples/sec
+epoch 11, total_step 174980, total loss is 14.62 , inference loss is 8.03, weight deacy loss is 6.60, training accuracy is 0.562500, time 56.115 samples/sec
+epoch 11, total_step 175000, total loss is 13.92 , inference loss is 7.32, weight deacy loss is 6.60, training accuracy is 0.500000, time 56.938 samples/sec
+epoch 11, total_step 175020, total loss is 11.66 , inference loss is 5.06, weight deacy loss is 6.60, training accuracy is 0.468750, time 56.386 samples/sec
+epoch 11, total_step 175040, total loss is 13.44 , inference loss is 6.84, weight deacy loss is 6.60, training accuracy is 0.531250, time 52.816 samples/sec
+epoch 11, total_step 175060, total loss is 16.94 , inference loss is 10.34, weight deacy loss is 6.60, training accuracy is 0.500000, time 54.876 samples/sec
+epoch 11, total_step 175080, total loss is 16.18 , inference loss is 9.58, weight deacy loss is 6.60, training accuracy is 0.375000, time 55.843 samples/sec
+epoch 11, total_step 175100, total loss is 17.44 , inference loss is 10.85, weight deacy loss is 6.60, training accuracy is 0.281250, time 54.786 samples/sec
+epoch 11, total_step 175120, total loss is 14.30 , inference loss is 7.71, weight deacy loss is 6.60, training accuracy is 0.500000, time 57.135 samples/sec
+epoch 11, total_step 175140, total loss is 14.70 , inference loss is 8.10, weight deacy loss is 6.60, training accuracy is 0.468750, time 56.680 samples/sec
+epoch 11, total_step 175160, total loss is 17.25 , inference loss is 10.65, weight deacy loss is 6.60, training accuracy is 0.375000, time 57.019 samples/sec
+epoch 11, total_step 175180, total loss is 14.14 , inference loss is 7.54, weight deacy loss is 6.60, training accuracy is 0.406250, time 57.327 samples/sec
+epoch 11, total_step 175200, total loss is 11.07 , inference loss is 4.47, weight deacy loss is 6.60, training accuracy is 0.531250, time 56.162 samples/sec
+epoch 11, total_step 175220, total loss is 15.48 , inference loss is 8.88, weight deacy loss is 6.60, training accuracy is 0.406250, time 57.340 samples/sec
+epoch 11, total_step 175240, total loss is 20.58 , inference loss is 13.98, weight deacy loss is 6.60, training accuracy is 0.343750, time 56.728 samples/sec
+epoch 11, total_step 175260, total loss is 16.89 , inference loss is 10.29, weight deacy loss is 6.60, training accuracy is 0.312500, time 55.648 samples/sec
+epoch 11, total_step 175280, total loss is 12.07 , inference loss is 5.48, weight deacy loss is 6.59, training accuracy is 0.500000, time 55.476 samples/sec
+epoch 11, total_step 175300, total loss is 17.32 , inference loss is 10.72, weight deacy loss is 6.59, training accuracy is 0.500000, time 57.187 samples/sec
+epoch 11, total_step 175320, total loss is 12.27 , inference loss is 5.68, weight deacy loss is 6.59, training accuracy is 0.437500, time 54.332 samples/sec
+epoch 11, total_step 175340, total loss is 13.90 , inference loss is 7.30, weight deacy loss is 6.59, training accuracy is 0.500000, time 56.878 samples/sec
+epoch 11, total_step 175360, total loss is 13.38 , inference loss is 6.79, weight deacy loss is 6.59, training accuracy is 0.531250, time 56.252 samples/sec
+epoch 11, total_step 175380, total loss is 11.22 , inference loss is 4.62, weight deacy loss is 6.59, training accuracy is 0.531250, time 56.800 samples/sec
+epoch 11, total_step 175400, total loss is 17.98 , inference loss is 11.38, weight deacy loss is 6.59, training accuracy is 0.406250, time 56.953 samples/sec
+epoch 11, total_step 175420, total loss is 11.56 , inference loss is 4.96, weight deacy loss is 6.59, training accuracy is 0.500000, time 56.347 samples/sec
+epoch 11, total_step 175440, total loss is 13.67 , inference loss is 7.08, weight deacy loss is 6.59, training accuracy is 0.437500, time 56.432 samples/sec
+epoch 11, total_step 175460, total loss is 14.20 , inference loss is 7.61, weight deacy loss is 6.59, training accuracy is 0.437500, time 56.985 samples/sec
+epoch 11, total_step 175480, total loss is 10.20 , inference loss is 3.61, weight deacy loss is 6.59, training accuracy is 0.562500, time 56.565 samples/sec
+epoch 11, total_step 175500, total loss is 16.88 , inference loss is 10.29, weight deacy loss is 6.59, training accuracy is 0.343750, time 56.305 samples/sec
+epoch 11, total_step 175520, total loss is 17.81 , inference loss is 11.22, weight deacy loss is 6.59, training accuracy is 0.250000, time 56.412 samples/sec
+epoch 11, total_step 175540, total loss is 15.31 , inference loss is 8.72, weight deacy loss is 6.59, training accuracy is 0.531250, time 56.516 samples/sec
+epoch 11, total_step 175560, total loss is 14.34 , inference loss is 7.75, weight deacy loss is 6.59, training accuracy is 0.406250, time 56.929 samples/sec
+epoch 11, total_step 175580, total loss is 9.34 , inference loss is 2.75, weight deacy loss is 6.59, training accuracy is 0.718750, time 56.950 samples/sec
+epoch 11, total_step 175600, total loss is 11.33 , inference loss is 4.74, weight deacy loss is 6.59, training accuracy is 0.500000, time 56.810 samples/sec
+epoch 11, total_step 175620, total loss is 14.52 , inference loss is 7.93, weight deacy loss is 6.59, training accuracy is 0.343750, time 56.251 samples/sec
+epoch 11, total_step 175640, total loss is 14.35 , inference loss is 7.76, weight deacy loss is 6.59, training accuracy is 0.531250, time 56.450 samples/sec
+epoch 11, total_step 175660, total loss is 15.48 , inference loss is 8.89, weight deacy loss is 6.59, training accuracy is 0.468750, time 56.419 samples/sec
+epoch 11, total_step 175680, total loss is 11.36 , inference loss is 4.77, weight deacy loss is 6.59, training accuracy is 0.531250, time 55.488 samples/sec
+epoch 11, total_step 175700, total loss is 12.22 , inference loss is 5.63, weight deacy loss is 6.59, training accuracy is 0.437500, time 56.781 samples/sec
+epoch 11, total_step 175720, total loss is 13.54 , inference loss is 6.94, weight deacy loss is 6.59, training accuracy is 0.625000, time 56.626 samples/sec
+epoch 11, total_step 175740, total loss is 15.64 , inference loss is 9.05, weight deacy loss is 6.59, training accuracy is 0.562500, time 56.278 samples/sec
+epoch 11, total_step 175760, total loss is 17.63 , inference loss is 11.04, weight deacy loss is 6.59, training accuracy is 0.437500, time 56.380 samples/sec
+epoch 11, total_step 175780, total loss is 17.75 , inference loss is 11.15, weight deacy loss is 6.59, training accuracy is 0.500000, time 56.293 samples/sec
+epoch 11, total_step 175800, total loss is 13.14 , inference loss is 6.55, weight deacy loss is 6.59, training accuracy is 0.343750, time 56.561 samples/sec
+epoch 11, total_step 175820, total loss is 13.36 , inference loss is 6.77, weight deacy loss is 6.59, training accuracy is 0.406250, time 56.452 samples/sec
+epoch 11, total_step 175840, total loss is 13.82 , inference loss is 7.23, weight deacy loss is 6.59, training accuracy is 0.531250, time 56.299 samples/sec
+epoch 11, total_step 175860, total loss is 14.50 , inference loss is 7.91, weight deacy loss is 6.59, training accuracy is 0.468750, time 55.839 samples/sec
+epoch 11, total_step 175880, total loss is 15.15 , inference loss is 8.56, weight deacy loss is 6.59, training accuracy is 0.375000, time 56.769 samples/sec
+epoch 11, total_step 175900, total loss is 10.61 , inference loss is 4.02, weight deacy loss is 6.59, training accuracy is 0.593750, time 56.231 samples/sec
+epoch 11, total_step 175920, total loss is 13.13 , inference loss is 6.54, weight deacy loss is 6.59, training accuracy is 0.437500, time 56.181 samples/sec
+epoch 11, total_step 175940, total loss is 16.02 , inference loss is 9.43, weight deacy loss is 6.59, training accuracy is 0.468750, time 56.544 samples/sec
+epoch 11, total_step 175960, total loss is 13.31 , inference loss is 6.72, weight deacy loss is 6.59, training accuracy is 0.375000, time 57.423 samples/sec
+epoch 11, total_step 175980, total loss is 11.37 , inference loss is 4.78, weight deacy loss is 6.59, training accuracy is 0.562500, time 56.030 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.72013400000002
+best_threshold_index 118 0.9885185185185185
+best_threshold_index 123 0.9879629629629629
+best_threshold_index 123 0.9885185185185185
+best_threshold_index 123 0.9883333333333333
+best_threshold_index 118 0.9885185185185185
+best_threshold_index 123 0.9875925925925926
+best_threshold_index 123 0.9888888888888889
+best_threshold_index 118 0.9877777777777778
+best_threshold_index 123 0.9872222222222222
+best_threshold_index 123 0.987037037037037
+[lfw][176000]XNorm: 19.820502
+[lfw][176000]Accuracy-Flip: 0.98717+-0.00553
+testing verification..
+(14000, 512)
+infer time 85.94355799999998
+best_threshold_index 115 0.9884126984126984
+best_threshold_index 120 0.9874603174603175
+best_threshold_index 120 0.9879365079365079
+best_threshold_index 120 0.9873015873015873
+best_threshold_index 125 0.9874603174603175
+best_threshold_index 120 0.9873015873015873
+best_threshold_index 120 0.9887301587301587
+best_threshold_index 120 0.9874603174603175
+best_threshold_index 121 0.9876190476190476
+best_threshold_index 125 0.9880952380952381
+[cfp_ff][176000]XNorm: 19.215493
+[cfp_ff][176000]Accuracy-Flip: 0.98657+-0.00470
+testing verification..
+(14000, 512)
+infer time 85.76572699999988
+best_threshold_index 135 0.9304761904761905
+best_threshold_index 135 0.9285714285714286
+best_threshold_index 135 0.927936507936508
+best_threshold_index 135 0.9292063492063493
+best_threshold_index 134 0.9303174603174603
+best_threshold_index 135 0.9265079365079365
+best_threshold_index 135 0.9257142857142857
+best_threshold_index 135 0.9276190476190476
+best_threshold_index 134 0.9288888888888889
+best_threshold_index 135 0.9292063492063493
+[cfp_fp][176000]XNorm: 19.489530
+[cfp_fp][176000]Accuracy-Flip: 0.92771+-0.01359
+testing verification..
+(12000, 512)
+infer time 72.74881499999998
+best_threshold_index 134 0.91
+best_threshold_index 134 0.9083333333333333
+best_threshold_index 134 0.9101851851851852
+best_threshold_index 134 0.9133333333333333
+best_threshold_index 134 0.9155555555555556
+best_threshold_index 134 0.909074074074074
+best_threshold_index 134 0.9111111111111111
+best_threshold_index 136 0.9162962962962963
+best_threshold_index 134 0.9118518518518518
+best_threshold_index 134 0.9127777777777778
+[agedb_30][176000]XNorm: 21.136872
+[agedb_30][176000]Accuracy-Flip: 0.91067+-0.02454
+test accuracy is: 0.9871666666666666
+epoch 11, total_step 176000, total loss is 16.31 , inference loss is 9.72, weight deacy loss is 6.59, training accuracy is 0.343750, time 53.523 samples/sec
+epoch 11, total_step 176020, total loss is 12.94 , inference loss is 6.35, weight deacy loss is 6.59, training accuracy is 0.500000, time 57.456 samples/sec
+epoch 11, total_step 176040, total loss is 13.08 , inference loss is 6.49, weight deacy loss is 6.59, training accuracy is 0.562500, time 56.377 samples/sec
+epoch 11, total_step 176060, total loss is 13.38 , inference loss is 6.79, weight deacy loss is 6.59, training accuracy is 0.406250, time 56.627 samples/sec
+epoch 11, total_step 176080, total loss is 15.20 , inference loss is 8.61, weight deacy loss is 6.59, training accuracy is 0.500000, time 56.289 samples/sec
+epoch 11, total_step 176100, total loss is 12.32 , inference loss is 5.73, weight deacy loss is 6.59, training accuracy is 0.593750, time 56.522 samples/sec
+epoch 11, total_step 176120, total loss is 15.46 , inference loss is 8.87, weight deacy loss is 6.59, training accuracy is 0.531250, time 56.178 samples/sec
+epoch 11, total_step 176140, total loss is 17.15 , inference loss is 10.56, weight deacy loss is 6.59, training accuracy is 0.437500, time 56.613 samples/sec
+epoch 11, total_step 176160, total loss is 11.26 , inference loss is 4.67, weight deacy loss is 6.59, training accuracy is 0.562500, time 56.824 samples/sec
+epoch 11, total_step 176180, total loss is 18.53 , inference loss is 11.94, weight deacy loss is 6.59, training accuracy is 0.375000, time 56.705 samples/sec
+epoch 11, total_step 176200, total loss is 16.22 , inference loss is 9.63, weight deacy loss is 6.59, training accuracy is 0.437500, time 55.896 samples/sec
+epoch 11, total_step 176220, total loss is 17.86 , inference loss is 11.28, weight deacy loss is 6.59, training accuracy is 0.406250, time 56.987 samples/sec
+epoch 11, total_step 176240, total loss is 18.15 , inference loss is 11.56, weight deacy loss is 6.59, training accuracy is 0.593750, time 55.576 samples/sec
+epoch 11, total_step 176260, total loss is 16.43 , inference loss is 9.85, weight deacy loss is 6.59, training accuracy is 0.500000, time 57.374 samples/sec
+epoch 11, total_step 176280, total loss is 13.56 , inference loss is 6.97, weight deacy loss is 6.59, training accuracy is 0.468750, time 55.912 samples/sec
+epoch 11, total_step 176300, total loss is 14.29 , inference loss is 7.70, weight deacy loss is 6.59, training accuracy is 0.468750, time 56.147 samples/sec
+epoch 11, total_step 176320, total loss is 15.00 , inference loss is 8.42, weight deacy loss is 6.59, training accuracy is 0.437500, time 56.088 samples/sec
+epoch 11, total_step 176340, total loss is 14.34 , inference loss is 7.75, weight deacy loss is 6.59, training accuracy is 0.406250, time 56.185 samples/sec
+epoch 11, total_step 176360, total loss is 11.71 , inference loss is 5.12, weight deacy loss is 6.59, training accuracy is 0.500000, time 56.352 samples/sec
+epoch 11, total_step 176380, total loss is 12.58 , inference loss is 6.00, weight deacy loss is 6.59, training accuracy is 0.500000, time 55.198 samples/sec
+epoch 11, total_step 176400, total loss is 16.27 , inference loss is 9.69, weight deacy loss is 6.59, training accuracy is 0.437500, time 56.215 samples/sec
+epoch 11, total_step 176420, total loss is 11.16 , inference loss is 4.57, weight deacy loss is 6.58, training accuracy is 0.531250, time 57.119 samples/sec
+epoch 11, total_step 176440, total loss is 14.81 , inference loss is 8.22, weight deacy loss is 6.58, training accuracy is 0.500000, time 55.924 samples/sec
+epoch 11, total_step 176460, total loss is 14.96 , inference loss is 8.38, weight deacy loss is 6.58, training accuracy is 0.343750, time 56.380 samples/sec
+epoch 11, total_step 176480, total loss is 11.91 , inference loss is 5.32, weight deacy loss is 6.58, training accuracy is 0.500000, time 56.274 samples/sec
+epoch 11, total_step 176500, total loss is 14.20 , inference loss is 7.62, weight deacy loss is 6.58, training accuracy is 0.531250, time 56.376 samples/sec
+epoch 11, total_step 176520, total loss is 11.00 , inference loss is 4.42, weight deacy loss is 6.58, training accuracy is 0.562500, time 55.131 samples/sec
+epoch 11, total_step 176540, total loss is 11.09 , inference loss is 4.51, weight deacy loss is 6.58, training accuracy is 0.500000, time 56.519 samples/sec
+epoch 11, total_step 176560, total loss is 14.33 , inference loss is 7.75, weight deacy loss is 6.58, training accuracy is 0.437500, time 56.691 samples/sec
+epoch 11, total_step 176580, total loss is 11.52 , inference loss is 4.94, weight deacy loss is 6.58, training accuracy is 0.500000, time 56.326 samples/sec
+epoch 11, total_step 176600, total loss is 14.56 , inference loss is 7.98, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.713 samples/sec
+epoch 11, total_step 176620, total loss is 15.57 , inference loss is 8.99, weight deacy loss is 6.58, training accuracy is 0.406250, time 56.139 samples/sec
+epoch 11, total_step 176640, total loss is 14.44 , inference loss is 7.86, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.673 samples/sec
+epoch 11, total_step 176660, total loss is 11.06 , inference loss is 4.48, weight deacy loss is 6.58, training accuracy is 0.500000, time 57.212 samples/sec
+epoch 11, total_step 176680, total loss is 18.33 , inference loss is 11.75, weight deacy loss is 6.58, training accuracy is 0.468750, time 55.998 samples/sec
+epoch 11, total_step 176700, total loss is 14.22 , inference loss is 7.64, weight deacy loss is 6.58, training accuracy is 0.593750, time 56.466 samples/sec
+epoch 11, total_step 176720, total loss is 11.98 , inference loss is 5.40, weight deacy loss is 6.58, training accuracy is 0.562500, time 55.337 samples/sec
+epoch 11, total_step 176740, total loss is 14.72 , inference loss is 8.14, weight deacy loss is 6.58, training accuracy is 0.437500, time 56.752 samples/sec
+epoch 11, total_step 176760, total loss is 13.53 , inference loss is 6.95, weight deacy loss is 6.58, training accuracy is 0.468750, time 55.723 samples/sec
+epoch 11, total_step 176780, total loss is 13.23 , inference loss is 6.65, weight deacy loss is 6.58, training accuracy is 0.437500, time 56.578 samples/sec
+epoch 11, total_step 176800, total loss is 11.85 , inference loss is 5.27, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.775 samples/sec
+epoch 11, total_step 176820, total loss is 16.56 , inference loss is 9.98, weight deacy loss is 6.58, training accuracy is 0.343750, time 56.492 samples/sec
+epoch 11, total_step 176840, total loss is 13.07 , inference loss is 6.49, weight deacy loss is 6.58, training accuracy is 0.500000, time 57.195 samples/sec
+epoch 11, total_step 176860, total loss is 13.70 , inference loss is 7.12, weight deacy loss is 6.58, training accuracy is 0.531250, time 56.740 samples/sec
+epoch 11, total_step 176880, total loss is 16.64 , inference loss is 10.06, weight deacy loss is 6.58, training accuracy is 0.343750, time 56.504 samples/sec
+epoch 11, total_step 176900, total loss is 19.40 , inference loss is 12.82, weight deacy loss is 6.58, training accuracy is 0.312500, time 56.060 samples/sec
+epoch 11, total_step 176920, total loss is 14.64 , inference loss is 8.06, weight deacy loss is 6.58, training accuracy is 0.406250, time 56.706 samples/sec
+epoch 11, total_step 176940, total loss is 10.53 , inference loss is 3.95, weight deacy loss is 6.58, training accuracy is 0.531250, time 56.231 samples/sec
+epoch 11, total_step 176960, total loss is 16.29 , inference loss is 9.71, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.483 samples/sec
+epoch 11, total_step 176980, total loss is 17.03 , inference loss is 10.45, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.619 samples/sec
+epoch 11, total_step 177000, total loss is 20.62 , inference loss is 14.04, weight deacy loss is 6.58, training accuracy is 0.343750, time 56.556 samples/sec
+epoch 11, total_step 177020, total loss is 14.30 , inference loss is 7.72, weight deacy loss is 6.58, training accuracy is 0.437500, time 56.732 samples/sec
+epoch 11, total_step 177040, total loss is 15.54 , inference loss is 8.96, weight deacy loss is 6.58, training accuracy is 0.406250, time 56.873 samples/sec
+epoch 11, total_step 177060, total loss is 13.41 , inference loss is 6.83, weight deacy loss is 6.58, training accuracy is 0.437500, time 56.315 samples/sec
+epoch 11, total_step 177080, total loss is 15.39 , inference loss is 8.81, weight deacy loss is 6.58, training accuracy is 0.406250, time 57.462 samples/sec
+epoch 11, total_step 177100, total loss is 19.45 , inference loss is 12.87, weight deacy loss is 6.58, training accuracy is 0.406250, time 56.696 samples/sec
+epoch 11, total_step 177120, total loss is 11.99 , inference loss is 5.42, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.818 samples/sec
+epoch 11, total_step 177140, total loss is 14.11 , inference loss is 7.54, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.774 samples/sec
+epoch 11, total_step 177160, total loss is 18.45 , inference loss is 11.87, weight deacy loss is 6.58, training accuracy is 0.375000, time 56.622 samples/sec
+epoch 11, total_step 177180, total loss is 11.39 , inference loss is 4.81, weight deacy loss is 6.58, training accuracy is 0.531250, time 56.882 samples/sec
+epoch 11, total_step 177200, total loss is 16.33 , inference loss is 9.75, weight deacy loss is 6.58, training accuracy is 0.406250, time 56.869 samples/sec
+epoch 11, total_step 177220, total loss is 13.50 , inference loss is 6.92, weight deacy loss is 6.58, training accuracy is 0.375000, time 56.489 samples/sec
+epoch 11, total_step 177240, total loss is 16.00 , inference loss is 9.43, weight deacy loss is 6.58, training accuracy is 0.406250, time 56.493 samples/sec
+epoch 11, total_step 177260, total loss is 12.70 , inference loss is 6.12, weight deacy loss is 6.58, training accuracy is 0.562500, time 56.649 samples/sec
+epoch 11, total_step 177280, total loss is 15.59 , inference loss is 9.02, weight deacy loss is 6.58, training accuracy is 0.375000, time 56.925 samples/sec
+epoch 11, total_step 177300, total loss is 12.93 , inference loss is 6.36, weight deacy loss is 6.58, training accuracy is 0.593750, time 56.099 samples/sec
+epoch 11, total_step 177320, total loss is 12.73 , inference loss is 6.15, weight deacy loss is 6.58, training accuracy is 0.500000, time 56.349 samples/sec
+epoch 11, total_step 177340, total loss is 12.98 , inference loss is 6.41, weight deacy loss is 6.58, training accuracy is 0.343750, time 56.291 samples/sec
+epoch 11, total_step 177360, total loss is 15.96 , inference loss is 9.38, weight deacy loss is 6.58, training accuracy is 0.375000, time 56.752 samples/sec
+epoch 11, total_step 177380, total loss is 14.57 , inference loss is 7.99, weight deacy loss is 6.58, training accuracy is 0.281250, time 56.195 samples/sec
+epoch 11, total_step 177400, total loss is 14.12 , inference loss is 7.54, weight deacy loss is 6.58, training accuracy is 0.375000, time 57.274 samples/sec
+epoch 11, total_step 177420, total loss is 11.93 , inference loss is 5.35, weight deacy loss is 6.58, training accuracy is 0.437500, time 56.198 samples/sec
+epoch 11, total_step 177440, total loss is 15.49 , inference loss is 8.91, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.955 samples/sec
+epoch 11, total_step 177460, total loss is 18.67 , inference loss is 12.09, weight deacy loss is 6.58, training accuracy is 0.281250, time 56.596 samples/sec
+epoch 11, total_step 177480, total loss is 16.01 , inference loss is 9.43, weight deacy loss is 6.58, training accuracy is 0.343750, time 56.922 samples/sec
+epoch 11, total_step 177500, total loss is 15.25 , inference loss is 8.68, weight deacy loss is 6.58, training accuracy is 0.406250, time 56.667 samples/sec
+epoch 11, total_step 177520, total loss is 17.74 , inference loss is 11.16, weight deacy loss is 6.58, training accuracy is 0.312500, time 56.120 samples/sec
+epoch 11, total_step 177540, total loss is 14.06 , inference loss is 7.48, weight deacy loss is 6.58, training accuracy is 0.406250, time 57.297 samples/sec
+epoch 11, total_step 177560, total loss is 12.31 , inference loss is 5.73, weight deacy loss is 6.58, training accuracy is 0.625000, time 56.349 samples/sec
+epoch 11, total_step 177580, total loss is 15.70 , inference loss is 9.13, weight deacy loss is 6.58, training accuracy is 0.468750, time 56.759 samples/sec
+epoch 11, total_step 177600, total loss is 13.15 , inference loss is 6.58, weight deacy loss is 6.58, training accuracy is 0.500000, time 55.488 samples/sec
+epoch 11, total_step 177620, total loss is 14.98 , inference loss is 8.40, weight deacy loss is 6.58, training accuracy is 0.562500, time 56.661 samples/sec
+epoch 11, total_step 177640, total loss is 11.75 , inference loss is 5.17, weight deacy loss is 6.58, training accuracy is 0.375000, time 56.868 samples/sec
+epoch 11, total_step 177660, total loss is 16.02 , inference loss is 9.44, weight deacy loss is 6.58, training accuracy is 0.437500, time 57.391 samples/sec
+epoch 11, total_step 177680, total loss is 13.72 , inference loss is 7.15, weight deacy loss is 6.57, training accuracy is 0.406250, time 57.547 samples/sec
+epoch 11, total_step 177700, total loss is 10.63 , inference loss is 4.06, weight deacy loss is 6.57, training accuracy is 0.625000, time 55.993 samples/sec
+epoch 11, total_step 177720, total loss is 12.06 , inference loss is 5.48, weight deacy loss is 6.57, training accuracy is 0.562500, time 56.869 samples/sec
+epoch 11, total_step 177740, total loss is 13.58 , inference loss is 7.01, weight deacy loss is 6.57, training accuracy is 0.406250, time 55.874 samples/sec
+epoch 11, total_step 177760, total loss is 12.14 , inference loss is 5.57, weight deacy loss is 6.57, training accuracy is 0.562500, time 56.273 samples/sec
+epoch 11, total_step 177780, total loss is 12.79 , inference loss is 6.22, weight deacy loss is 6.57, training accuracy is 0.406250, time 56.622 samples/sec
+epoch 11, total_step 177800, total loss is 12.66 , inference loss is 6.08, weight deacy loss is 6.57, training accuracy is 0.375000, time 56.708 samples/sec
+epoch 11, total_step 177820, total loss is 13.83 , inference loss is 7.26, weight deacy loss is 6.57, training accuracy is 0.468750, time 56.566 samples/sec
+epoch 11, total_step 177840, total loss is 12.30 , inference loss is 5.73, weight deacy loss is 6.57, training accuracy is 0.437500, time 55.987 samples/sec
+epoch 11, total_step 177860, total loss is 15.95 , inference loss is 9.37, weight deacy loss is 6.57, training accuracy is 0.437500, time 56.987 samples/sec
+epoch 11, total_step 177880, total loss is 15.06 , inference loss is 8.49, weight deacy loss is 6.57, training accuracy is 0.437500, time 56.453 samples/sec
+epoch 11, total_step 177900, total loss is 12.92 , inference loss is 6.35, weight deacy loss is 6.57, training accuracy is 0.593750, time 56.889 samples/sec
+epoch 11, total_step 177920, total loss is 16.35 , inference loss is 9.77, weight deacy loss is 6.57, training accuracy is 0.531250, time 56.789 samples/sec
+epoch 11, total_step 177940, total loss is 16.54 , inference loss is 9.97, weight deacy loss is 6.57, training accuracy is 0.500000, time 55.715 samples/sec
+epoch 11, total_step 177960, total loss is 12.82 , inference loss is 6.24, weight deacy loss is 6.57, training accuracy is 0.468750, time 56.732 samples/sec
+epoch 11, total_step 177980, total loss is 13.77 , inference loss is 7.19, weight deacy loss is 6.57, training accuracy is 0.437500, time 56.366 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.75500400000001
+best_threshold_index 115 0.9890740740740741
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 115 0.9881481481481481
+best_threshold_index 115 0.9883333333333333
+best_threshold_index 121 0.9887037037037038
+best_threshold_index 115 0.9875925925925926
+best_threshold_index 121 0.9885185185185185
+best_threshold_index 115 0.9875925925925926
+best_threshold_index 115 0.987037037037037
+best_threshold_index 115 0.9875925925925926
+[lfw][178000]XNorm: 21.310687
+[lfw][178000]Accuracy-Flip: 0.98733+-0.00569
+testing verification..
+(14000, 512)
+infer time 85.99496899999993
+best_threshold_index 120 0.9890476190476191
+best_threshold_index 120 0.9877777777777778
+best_threshold_index 120 0.9876190476190476
+best_threshold_index 120 0.9877777777777778
+best_threshold_index 120 0.9887301587301587
+best_threshold_index 120 0.9877777777777778
+best_threshold_index 120 0.9882539682539683
+best_threshold_index 120 0.9880952380952381
+best_threshold_index 120 0.9882539682539683
+best_threshold_index 120 0.9880952380952381
+[cfp_ff][178000]XNorm: 20.864720
+[cfp_ff][178000]Accuracy-Flip: 0.98814+-0.00389
+testing verification..
+(14000, 512)
+infer time 85.73147600000004
+best_threshold_index 142 0.9273015873015873
+best_threshold_index 142 0.9266666666666666
+best_threshold_index 145 0.9258730158730158
+best_threshold_index 140 0.9280952380952381
+best_threshold_index 142 0.9307936507936508
+best_threshold_index 142 0.9250793650793651
+best_threshold_index 145 0.9234920634920635
+best_threshold_index 145 0.926031746031746
+best_threshold_index 141 0.9258730158730158
+best_threshold_index 141 0.9265079365079365
+[cfp_fp][178000]XNorm: 21.000609
+[cfp_fp][178000]Accuracy-Flip: 0.92400+-0.01652
+testing verification..
+(12000, 512)
+infer time 73.00091299999995
+best_threshold_index 140 0.8977777777777778
+best_threshold_index 140 0.8975925925925926
+best_threshold_index 140 0.8987037037037037
+best_threshold_index 140 0.9
+best_threshold_index 140 0.9027777777777778
+best_threshold_index 140 0.8981481481481481
+best_threshold_index 140 0.9022222222222223
+best_threshold_index 142 0.9068518518518518
+best_threshold_index 140 0.9025925925925926
+best_threshold_index 140 0.9009259259259259
+[agedb_30][178000]XNorm: 22.448825
+[agedb_30][178000]Accuracy-Flip: 0.89950+-0.02576
+test accuracy is: 0.9873333333333333
+epoch 11, total_step 178000, total loss is 13.59 , inference loss is 7.02, weight deacy loss is 6.57, training accuracy is 0.593750, time 53.529 samples/sec
+epoch 11, total_step 178020, total loss is 16.96 , inference loss is 10.39, weight deacy loss is 6.57, training accuracy is 0.468750, time 56.664 samples/sec
+epoch 11, total_step 178040, total loss is 13.18 , inference loss is 6.61, weight deacy loss is 6.57, training accuracy is 0.406250, time 57.257 samples/sec
+epoch 11, total_step 178060, total loss is 18.24 , inference loss is 11.66, weight deacy loss is 6.57, training accuracy is 0.375000, time 57.640 samples/sec
+epoch 11, total_step 178080, total loss is 12.22 , inference loss is 5.65, weight deacy loss is 6.57, training accuracy is 0.375000, time 55.776 samples/sec
+epoch 11, total_step 178100, total loss is 14.53 , inference loss is 7.96, weight deacy loss is 6.57, training accuracy is 0.406250, time 56.623 samples/sec
+epoch 11, total_step 178120, total loss is 13.07 , inference loss is 6.50, weight deacy loss is 6.57, training accuracy is 0.500000, time 55.714 samples/sec
+epoch 11, total_step 178140, total loss is 15.07 , inference loss is 8.50, weight deacy loss is 6.57, training accuracy is 0.468750, time 57.052 samples/sec
+epoch 11, total_step 178160, total loss is 12.74 , inference loss is 6.17, weight deacy loss is 6.57, training accuracy is 0.343750, time 57.087 samples/sec
+epoch 11, total_step 178180, total loss is 11.80 , inference loss is 5.22, weight deacy loss is 6.57, training accuracy is 0.500000, time 56.391 samples/sec
+epoch 11, total_step 178200, total loss is 14.48 , inference loss is 7.91, weight deacy loss is 6.57, training accuracy is 0.218750, time 56.047 samples/sec
+epoch 11, total_step 178220, total loss is 15.66 , inference loss is 9.09, weight deacy loss is 6.57, training accuracy is 0.375000, time 57.028 samples/sec
+epoch 11, total_step 178240, total loss is 13.21 , inference loss is 6.64, weight deacy loss is 6.57, training accuracy is 0.375000, time 55.984 samples/sec
+epoch 11, total_step 178260, total loss is 10.93 , inference loss is 4.36, weight deacy loss is 6.57, training accuracy is 0.562500, time 56.566 samples/sec
+epoch 11, total_step 178280, total loss is 16.18 , inference loss is 9.61, weight deacy loss is 6.57, training accuracy is 0.375000, time 56.651 samples/sec
+epoch 11, total_step 178300, total loss is 18.63 , inference loss is 12.06, weight deacy loss is 6.57, training accuracy is 0.343750, time 56.596 samples/sec
+epoch 11, total_step 178320, total loss is 11.70 , inference loss is 5.13, weight deacy loss is 6.57, training accuracy is 0.406250, time 55.034 samples/sec
+epoch 11, total_step 178340, total loss is 10.91 , inference loss is 4.34, weight deacy loss is 6.57, training accuracy is 0.531250, time 56.765 samples/sec
+epoch 11, total_step 178360, total loss is 17.47 , inference loss is 10.90, weight deacy loss is 6.57, training accuracy is 0.468750, time 56.668 samples/sec
+epoch 11, total_step 178380, total loss is 13.52 , inference loss is 6.95, weight deacy loss is 6.57, training accuracy is 0.437500, time 55.350 samples/sec
+epoch 11, total_step 178400, total loss is 18.05 , inference loss is 11.48, weight deacy loss is 6.57, training accuracy is 0.343750, time 56.289 samples/sec
+epoch 11, total_step 178420, total loss is 12.15 , inference loss is 5.58, weight deacy loss is 6.57, training accuracy is 0.437500, time 56.116 samples/sec
+epoch 11, total_step 178440, total loss is 13.09 , inference loss is 6.52, weight deacy loss is 6.57, training accuracy is 0.531250, time 57.358 samples/sec
+epoch 11, total_step 178460, total loss is 16.33 , inference loss is 9.76, weight deacy loss is 6.57, training accuracy is 0.468750, time 56.922 samples/sec
+epoch 11, total_step 178480, total loss is 16.67 , inference loss is 10.10, weight deacy loss is 6.57, training accuracy is 0.468750, time 56.055 samples/sec
+epoch 11, total_step 178500, total loss is 16.10 , inference loss is 9.53, weight deacy loss is 6.57, training accuracy is 0.375000, time 56.296 samples/sec
+epoch 11, total_step 178520, total loss is 13.10 , inference loss is 6.53, weight deacy loss is 6.57, training accuracy is 0.562500, time 56.999 samples/sec
+epoch 11, total_step 178540, total loss is 12.45 , inference loss is 5.88, weight deacy loss is 6.57, training accuracy is 0.437500, time 56.440 samples/sec
+epoch 11, total_step 178560, total loss is 10.65 , inference loss is 4.08, weight deacy loss is 6.57, training accuracy is 0.625000, time 55.324 samples/sec
+epoch 11, total_step 178580, total loss is 15.44 , inference loss is 8.87, weight deacy loss is 6.57, training accuracy is 0.375000, time 56.463 samples/sec
+epoch 11, total_step 178600, total loss is 11.71 , inference loss is 5.14, weight deacy loss is 6.57, training accuracy is 0.468750, time 56.555 samples/sec
+epoch 11, total_step 178620, total loss is 11.18 , inference loss is 4.61, weight deacy loss is 6.57, training accuracy is 0.531250, time 56.622 samples/sec
+epoch 11, total_step 178640, total loss is 18.03 , inference loss is 11.46, weight deacy loss is 6.57, training accuracy is 0.406250, time 55.712 samples/sec
+epoch 11, total_step 178660, total loss is 14.29 , inference loss is 7.72, weight deacy loss is 6.57, training accuracy is 0.406250, time 56.432 samples/sec
+epoch 11, total_step 178680, total loss is 12.25 , inference loss is 5.68, weight deacy loss is 6.57, training accuracy is 0.406250, time 57.090 samples/sec
+epoch 11, total_step 178700, total loss is 12.90 , inference loss is 6.34, weight deacy loss is 6.57, training accuracy is 0.562500, time 57.362 samples/sec
+epoch 11, total_step 178720, total loss is 13.80 , inference loss is 7.24, weight deacy loss is 6.57, training accuracy is 0.468750, time 55.880 samples/sec
+epoch 11, total_step 178740, total loss is 17.46 , inference loss is 10.90, weight deacy loss is 6.57, training accuracy is 0.437500, time 56.953 samples/sec
+epoch 11, total_step 178760, total loss is 13.14 , inference loss is 6.57, weight deacy loss is 6.57, training accuracy is 0.406250, time 55.574 samples/sec
+epoch 11, total_step 178780, total loss is 14.69 , inference loss is 8.12, weight deacy loss is 6.57, training accuracy is 0.468750, time 57.140 samples/sec
+epoch 11, total_step 178800, total loss is 13.50 , inference loss is 6.94, weight deacy loss is 6.57, training accuracy is 0.437500, time 54.905 samples/sec
+epoch 11, total_step 178820, total loss is 16.02 , inference loss is 9.45, weight deacy loss is 6.57, training accuracy is 0.375000, time 56.719 samples/sec
+epoch 11, total_step 178840, total loss is 12.66 , inference loss is 6.10, weight deacy loss is 6.57, training accuracy is 0.343750, time 56.682 samples/sec
+epoch 11, total_step 178860, total loss is 16.55 , inference loss is 9.99, weight deacy loss is 6.56, training accuracy is 0.375000, time 56.872 samples/sec
+epoch 11, total_step 178880, total loss is 17.53 , inference loss is 10.97, weight deacy loss is 6.56, training accuracy is 0.343750, time 56.465 samples/sec
+epoch 11, total_step 178900, total loss is 16.47 , inference loss is 9.90, weight deacy loss is 6.56, training accuracy is 0.531250, time 56.348 samples/sec
+epoch 11, total_step 178920, total loss is 14.75 , inference loss is 8.18, weight deacy loss is 6.56, training accuracy is 0.437500, time 56.291 samples/sec
+epoch 11, total_step 178940, total loss is 14.68 , inference loss is 8.11, weight deacy loss is 6.56, training accuracy is 0.406250, time 56.817 samples/sec
+epoch 11, total_step 178960, total loss is 15.37 , inference loss is 8.80, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.587 samples/sec
+epoch 11, total_step 178980, total loss is 15.70 , inference loss is 9.14, weight deacy loss is 6.56, training accuracy is 0.406250, time 56.422 samples/sec
+epoch 11, total_step 179000, total loss is 14.72 , inference loss is 8.16, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.832 samples/sec
+epoch 11, total_step 179020, total loss is 12.31 , inference loss is 5.75, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.639 samples/sec
+epoch 11, total_step 179040, total loss is 14.76 , inference loss is 8.20, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.253 samples/sec
+epoch 11, total_step 179060, total loss is 9.65 , inference loss is 3.09, weight deacy loss is 6.56, training accuracy is 0.625000, time 55.294 samples/sec
+epoch 11, total_step 179080, total loss is 13.37 , inference loss is 6.81, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.076 samples/sec
+epoch 11, total_step 179100, total loss is 14.52 , inference loss is 7.95, weight deacy loss is 6.56, training accuracy is 0.562500, time 56.944 samples/sec
+epoch 11, total_step 179120, total loss is 13.32 , inference loss is 6.75, weight deacy loss is 6.56, training accuracy is 0.500000, time 56.980 samples/sec
+epoch 11, total_step 179140, total loss is 17.32 , inference loss is 10.76, weight deacy loss is 6.56, training accuracy is 0.343750, time 56.877 samples/sec
+epoch 11, total_step 179160, total loss is 15.25 , inference loss is 8.69, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.307 samples/sec
+epoch 11, total_step 179180, total loss is 14.07 , inference loss is 7.50, weight deacy loss is 6.56, training accuracy is 0.437500, time 56.328 samples/sec
+epoch 11, total_step 179200, total loss is 14.18 , inference loss is 7.62, weight deacy loss is 6.56, training accuracy is 0.500000, time 56.061 samples/sec
+epoch 11, total_step 179220, total loss is 15.82 , inference loss is 9.26, weight deacy loss is 6.56, training accuracy is 0.437500, time 56.998 samples/sec
+epoch 11, total_step 179240, total loss is 14.93 , inference loss is 8.36, weight deacy loss is 6.56, training accuracy is 0.437500, time 57.422 samples/sec
+epoch 11, total_step 179260, total loss is 14.09 , inference loss is 7.53, weight deacy loss is 6.56, training accuracy is 0.562500, time 55.840 samples/sec
+epoch 11, total_step 179280, total loss is 16.72 , inference loss is 10.16, weight deacy loss is 6.56, training accuracy is 0.406250, time 56.285 samples/sec
+epoch 11, total_step 179300, total loss is 13.68 , inference loss is 7.12, weight deacy loss is 6.56, training accuracy is 0.406250, time 56.809 samples/sec
+epoch 11, total_step 179320, total loss is 14.90 , inference loss is 8.34, weight deacy loss is 6.56, training accuracy is 0.406250, time 57.255 samples/sec
+epoch 11, total_step 179340, total loss is 16.02 , inference loss is 9.46, weight deacy loss is 6.56, training accuracy is 0.437500, time 56.317 samples/sec
+epoch 11, total_step 179360, total loss is 11.40 , inference loss is 4.84, weight deacy loss is 6.56, training accuracy is 0.593750, time 57.344 samples/sec
+epoch 11, total_step 179380, total loss is 13.86 , inference loss is 7.30, weight deacy loss is 6.56, training accuracy is 0.593750, time 56.655 samples/sec
+epoch 11, total_step 179400, total loss is 13.78 , inference loss is 7.22, weight deacy loss is 6.56, training accuracy is 0.562500, time 56.382 samples/sec
+epoch 11, total_step 179420, total loss is 14.07 , inference loss is 7.51, weight deacy loss is 6.56, training accuracy is 0.406250, time 55.903 samples/sec
+epoch 11, total_step 179440, total loss is 11.51 , inference loss is 4.95, weight deacy loss is 6.56, training accuracy is 0.593750, time 56.414 samples/sec
+epoch 11, total_step 179460, total loss is 11.75 , inference loss is 5.19, weight deacy loss is 6.56, training accuracy is 0.593750, time 56.775 samples/sec
+epoch 11, total_step 179480, total loss is 15.63 , inference loss is 9.07, weight deacy loss is 6.56, training accuracy is 0.312500, time 56.548 samples/sec
+epoch 11, total_step 179500, total loss is 15.26 , inference loss is 8.70, weight deacy loss is 6.56, training accuracy is 0.593750, time 56.564 samples/sec
+epoch 11, total_step 179520, total loss is 16.97 , inference loss is 10.41, weight deacy loss is 6.56, training accuracy is 0.406250, time 57.483 samples/sec
+epoch 11, total_step 179540, total loss is 12.87 , inference loss is 6.31, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.704 samples/sec
+epoch 11, total_step 179560, total loss is 13.00 , inference loss is 6.44, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.042 samples/sec
+epoch 11, total_step 179580, total loss is 13.75 , inference loss is 7.19, weight deacy loss is 6.56, training accuracy is 0.500000, time 56.534 samples/sec
+epoch 11, total_step 179600, total loss is 18.14 , inference loss is 11.58, weight deacy loss is 6.56, training accuracy is 0.343750, time 56.810 samples/sec
+epoch 11, total_step 179620, total loss is 13.55 , inference loss is 6.99, weight deacy loss is 6.56, training accuracy is 0.406250, time 56.263 samples/sec
+epoch 11, total_step 179640, total loss is 14.92 , inference loss is 8.36, weight deacy loss is 6.56, training accuracy is 0.406250, time 57.370 samples/sec
+epoch 11, total_step 179660, total loss is 11.39 , inference loss is 4.83, weight deacy loss is 6.56, training accuracy is 0.562500, time 56.582 samples/sec
+epoch 11, total_step 179680, total loss is 15.11 , inference loss is 8.55, weight deacy loss is 6.56, training accuracy is 0.437500, time 57.206 samples/sec
+epoch 11, total_step 179700, total loss is 15.24 , inference loss is 8.68, weight deacy loss is 6.56, training accuracy is 0.500000, time 56.381 samples/sec
+epoch 11, total_step 179720, total loss is 14.79 , inference loss is 8.23, weight deacy loss is 6.56, training accuracy is 0.281250, time 56.379 samples/sec
+epoch 11, total_step 179740, total loss is 13.57 , inference loss is 7.01, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.723 samples/sec
+epoch 11, total_step 179760, total loss is 19.56 , inference loss is 13.00, weight deacy loss is 6.56, training accuracy is 0.437500, time 56.993 samples/sec
+epoch 11, total_step 179780, total loss is 13.40 , inference loss is 6.84, weight deacy loss is 6.56, training accuracy is 0.375000, time 56.430 samples/sec
+epoch 11, total_step 179800, total loss is 16.52 , inference loss is 9.97, weight deacy loss is 6.56, training accuracy is 0.375000, time 56.136 samples/sec
+epoch 11, total_step 179820, total loss is 14.50 , inference loss is 7.94, weight deacy loss is 6.56, training accuracy is 0.468750, time 56.582 samples/sec
+epoch 11, total_step 179840, total loss is 13.37 , inference loss is 6.81, weight deacy loss is 6.56, training accuracy is 0.531250, time 56.995 samples/sec
+epoch 11, total_step 179860, total loss is 17.16 , inference loss is 10.60, weight deacy loss is 6.56, training accuracy is 0.312500, time 55.702 samples/sec
+epoch 11, total_step 179880, total loss is 14.93 , inference loss is 8.37, weight deacy loss is 6.56, training accuracy is 0.343750, time 56.474 samples/sec
+epoch 11, total_step 179900, total loss is 15.70 , inference loss is 9.14, weight deacy loss is 6.56, training accuracy is 0.343750, time 56.258 samples/sec
+epoch 11, total_step 179920, total loss is 13.11 , inference loss is 6.56, weight deacy loss is 6.56, training accuracy is 0.593750, time 55.862 samples/sec
+epoch 11, total_step 179940, total loss is 11.75 , inference loss is 5.20, weight deacy loss is 6.56, training accuracy is 0.531250, time 56.402 samples/sec
+epoch 11, total_step 179960, total loss is 17.93 , inference loss is 11.37, weight deacy loss is 6.56, training accuracy is 0.343750, time 56.464 samples/sec
+epoch 11, total_step 179980, total loss is 14.69 , inference loss is 8.14, weight deacy loss is 6.56, training accuracy is 0.406250, time 56.819 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.55042199999998
+best_threshold_index 107 0.9848148148148148
+best_threshold_index 107 0.9844444444444445
+best_threshold_index 107 0.9840740740740741
+best_threshold_index 107 0.985
+best_threshold_index 107 0.9842592592592593
+best_threshold_index 107 0.9837037037037037
+best_threshold_index 107 0.9842592592592593
+best_threshold_index 107 0.9840740740740741
+best_threshold_index 107 0.9824074074074074
+best_threshold_index 107 0.9829629629629629
+[lfw][180000]XNorm: 22.005148
+[lfw][180000]Accuracy-Flip: 0.98400+-0.00680
+testing verification..
+(14000, 512)
+infer time 85.22056699999993
+best_threshold_index 119 0.9868253968253968
+best_threshold_index 115 0.9853968253968254
+best_threshold_index 115 0.984920634920635
+best_threshold_index 115 0.9847619047619047
+best_threshold_index 118 0.9863492063492063
+best_threshold_index 119 0.9850793650793651
+best_threshold_index 115 0.986031746031746
+best_threshold_index 115 0.9850793650793651
+best_threshold_index 115 0.984920634920635
+best_threshold_index 115 0.9855555555555555
+[cfp_ff][180000]XNorm: 21.965093
+[cfp_ff][180000]Accuracy-Flip: 0.98429+-0.00655
+testing verification..
+(14000, 512)
+infer time 85.34899499999989
+best_threshold_index 136 0.9222222222222223
+best_threshold_index 136 0.9252380952380952
+best_threshold_index 136 0.9214285714285714
+best_threshold_index 136 0.9234920634920635
+best_threshold_index 136 0.9261904761904762
+best_threshold_index 136 0.9214285714285714
+best_threshold_index 136 0.922063492063492
+best_threshold_index 136 0.9228571428571428
+best_threshold_index 136 0.9226984126984127
+best_threshold_index 136 0.9238095238095239
+[cfp_fp][180000]XNorm: 22.308247
+[cfp_fp][180000]Accuracy-Flip: 0.92314+-0.01347
+testing verification..
+(12000, 512)
+infer time 73.38494199999987
+best_threshold_index 136 0.905
+best_threshold_index 135 0.904074074074074
+best_threshold_index 135 0.905
+best_threshold_index 136 0.9062962962962963
+best_threshold_index 136 0.9112962962962963
+best_threshold_index 135 0.9057407407407407
+best_threshold_index 136 0.9074074074074074
+best_threshold_index 135 0.91
+best_threshold_index 136 0.9081481481481481
+best_threshold_index 136 0.9081481481481481
+[agedb_30][180000]XNorm: 22.737799
+[agedb_30][180000]Accuracy-Flip: 0.90533+-0.01962
+test accuracy is: 0.984
+epoch 11, total_step 180000, total loss is 16.44 , inference loss is 9.88, weight deacy loss is 6.56, training accuracy is 0.343750, time 52.938 samples/sec
+epoch 11, total_step 180020, total loss is 14.25 , inference loss is 7.69, weight deacy loss is 6.56, training accuracy is 0.250000, time 57.268 samples/sec
+epoch 11, total_step 180040, total loss is 14.12 , inference loss is 7.57, weight deacy loss is 6.56, training accuracy is 0.375000, time 56.544 samples/sec
+epoch 11, total_step 180060, total loss is 16.33 , inference loss is 9.77, weight deacy loss is 6.55, training accuracy is 0.468750, time 56.758 samples/sec
+epoch 11, total_step 180080, total loss is 11.46 , inference loss is 4.90, weight deacy loss is 6.55, training accuracy is 0.562500, time 56.813 samples/sec
+epoch 11, total_step 180100, total loss is 13.69 , inference loss is 7.13, weight deacy loss is 6.55, training accuracy is 0.375000, time 56.373 samples/sec
+epoch 11, total_step 180120, total loss is 13.20 , inference loss is 6.65, weight deacy loss is 6.55, training accuracy is 0.437500, time 56.665 samples/sec
+epoch 11, total_step 180140, total loss is 14.43 , inference loss is 7.88, weight deacy loss is 6.55, training accuracy is 0.593750, time 56.641 samples/sec
+epoch 11, total_step 180160, total loss is 13.20 , inference loss is 6.64, weight deacy loss is 6.55, training accuracy is 0.375000, time 56.389 samples/sec
+epoch 11, total_step 180180, total loss is 17.03 , inference loss is 10.48, weight deacy loss is 6.55, training accuracy is 0.406250, time 56.744 samples/sec
+epoch 11, total_step 180200, total loss is 14.20 , inference loss is 7.65, weight deacy loss is 6.55, training accuracy is 0.468750, time 57.069 samples/sec
+epoch 11, total_step 180220, total loss is 11.85 , inference loss is 5.29, weight deacy loss is 6.55, training accuracy is 0.562500, time 56.770 samples/sec
+epoch 11, total_step 180240, total loss is 13.57 , inference loss is 7.02, weight deacy loss is 6.55, training accuracy is 0.375000, time 56.727 samples/sec
+epoch 11, total_step 180260, total loss is 15.97 , inference loss is 9.41, weight deacy loss is 6.55, training accuracy is 0.343750, time 56.092 samples/sec
+epoch 11, total_step 180280, total loss is 15.40 , inference loss is 8.84, weight deacy loss is 6.55, training accuracy is 0.500000, time 56.549 samples/sec
+epoch 11, total_step 180300, total loss is 14.88 , inference loss is 8.32, weight deacy loss is 6.55, training accuracy is 0.406250, time 55.713 samples/sec
+epoch 11, total_step 180320, total loss is 10.86 , inference loss is 4.31, weight deacy loss is 6.55, training accuracy is 0.468750, time 57.085 samples/sec
+epoch 11, total_step 180340, total loss is 14.23 , inference loss is 7.68, weight deacy loss is 6.55, training accuracy is 0.281250, time 56.148 samples/sec
+epoch 11, total_step 180360, total loss is 15.07 , inference loss is 8.52, weight deacy loss is 6.55, training accuracy is 0.500000, time 56.880 samples/sec
+epoch 11, total_step 180380, total loss is 12.23 , inference loss is 5.68, weight deacy loss is 6.55, training accuracy is 0.468750, time 56.448 samples/sec
+epoch 11, total_step 180400, total loss is 15.35 , inference loss is 8.80, weight deacy loss is 6.55, training accuracy is 0.343750, time 56.539 samples/sec
+epoch 11, total_step 180420, total loss is 16.66 , inference loss is 10.10, weight deacy loss is 6.55, training accuracy is 0.250000, time 56.434 samples/sec
+epoch 11, total_step 180440, total loss is 13.25 , inference loss is 6.69, weight deacy loss is 6.55, training accuracy is 0.406250, time 56.730 samples/sec
+epoch 11, total_step 180460, total loss is 15.74 , inference loss is 9.19, weight deacy loss is 6.55, training accuracy is 0.406250, time 56.955 samples/sec
+epoch 11, total_step 180480, total loss is 14.60 , inference loss is 8.05, weight deacy loss is 6.55, training accuracy is 0.406250, time 56.700 samples/sec
+epoch 11, total_step 180500, total loss is 12.73 , inference loss is 6.18, weight deacy loss is 6.55, training accuracy is 0.468750, time 56.416 samples/sec
+epoch 11, total_step 180520, total loss is 17.23 , inference loss is 10.67, weight deacy loss is 6.55, training accuracy is 0.312500, time 56.479 samples/sec
+epoch 11, total_step 180540, total loss is 13.40 , inference loss is 6.85, weight deacy loss is 6.55, training accuracy is 0.437500, time 56.300 samples/sec
+epoch 11, total_step 180560, total loss is 14.60 , inference loss is 8.05, weight deacy loss is 6.55, training accuracy is 0.281250, time 56.860 samples/sec
+epoch 11, total_step 180580, total loss is 14.46 , inference loss is 7.91, weight deacy loss is 6.55, training accuracy is 0.562500, time 57.354 samples/sec
+epoch 11, total_step 180600, total loss is 13.41 , inference loss is 6.86, weight deacy loss is 6.55, training accuracy is 0.468750, time 56.497 samples/sec
+epoch 11, total_step 180620, total loss is 11.37 , inference loss is 4.82, weight deacy loss is 6.55, training accuracy is 0.531250, time 56.481 samples/sec
+epoch 11, total_step 180640, total loss is 17.97 , inference loss is 11.42, weight deacy loss is 6.55, training accuracy is 0.218750, time 56.663 samples/sec
+epoch 11, total_step 180660, total loss is 14.14 , inference loss is 7.59, weight deacy loss is 6.55, training accuracy is 0.250000, time 56.879 samples/sec
+epoch 11, total_step 180680, total loss is 13.84 , inference loss is 7.29, weight deacy loss is 6.55, training accuracy is 0.343750, time 56.521 samples/sec
+epoch 11, total_step 180700, total loss is 13.61 , inference loss is 7.06, weight deacy loss is 6.55, training accuracy is 0.468750, time 56.574 samples/sec
+epoch 11, total_step 180720, total loss is 15.54 , inference loss is 8.99, weight deacy loss is 6.55, training accuracy is 0.187500, time 56.826 samples/sec
+epoch 11, total_step 180740, total loss is 19.07 , inference loss is 12.52, weight deacy loss is 6.55, training accuracy is 0.187500, time 56.144 samples/sec
+epoch 11, total_step 180760, total loss is 16.62 , inference loss is 10.07, weight deacy loss is 6.55, training accuracy is 0.437500, time 55.892 samples/sec
+epoch 11, total_step 180780, total loss is 19.55 , inference loss is 13.00, weight deacy loss is 6.55, training accuracy is 0.343750, time 56.527 samples/sec
+epoch 11, total_step 180800, total loss is 13.73 , inference loss is 7.18, weight deacy loss is 6.55, training accuracy is 0.281250, time 56.142 samples/sec
+epoch 11, total_step 180820, total loss is 18.78 , inference loss is 12.23, weight deacy loss is 6.55, training accuracy is 0.312500, time 56.232 samples/sec
+epoch 11, total_step 180840, total loss is 14.40 , inference loss is 7.86, weight deacy loss is 6.55, training accuracy is 0.500000, time 57.354 samples/sec
+epoch 11, total_step 180860, total loss is 13.61 , inference loss is 7.06, weight deacy loss is 6.55, training accuracy is 0.468750, time 56.056 samples/sec
+epoch 11, total_step 180880, total loss is 18.28 , inference loss is 11.73, weight deacy loss is 6.55, training accuracy is 0.375000, time 57.249 samples/sec
+epoch 11, total_step 180900, total loss is 15.26 , inference loss is 8.71, weight deacy loss is 6.55, training accuracy is 0.437500, time 56.923 samples/sec
+epoch 11, total_step 180920, total loss is 16.35 , inference loss is 9.80, weight deacy loss is 6.55, training accuracy is 0.312500, time 56.773 samples/sec
+epoch 11, total_step 180940, total loss is 14.76 , inference loss is 8.21, weight deacy loss is 6.55, training accuracy is 0.406250, time 56.833 samples/sec
+epoch 11, total_step 180960, total loss is 15.04 , inference loss is 8.50, weight deacy loss is 6.55, training accuracy is 0.437500, time 56.156 samples/sec
+epoch 11, total_step 180980, total loss is 11.72 , inference loss is 5.17, weight deacy loss is 6.55, training accuracy is 0.281250, time 56.705 samples/sec
+epoch 11, total_step 181000, total loss is 18.14 , inference loss is 11.59, weight deacy loss is 6.55, training accuracy is 0.343750, time 56.149 samples/sec
+epoch 11, total_step 181020, total loss is 13.29 , inference loss is 6.75, weight deacy loss is 6.55, training accuracy is 0.437500, time 56.833 samples/sec
+epoch 11, total_step 181040, total loss is 14.55 , inference loss is 8.01, weight deacy loss is 6.55, training accuracy is 0.312500, time 56.056 samples/sec
+epoch 11, total_step 181060, total loss is 12.16 , inference loss is 5.61, weight deacy loss is 6.55, training accuracy is 0.343750, time 56.163 samples/sec
+epoch 11, total_step 181080, total loss is 18.23 , inference loss is 11.68, weight deacy loss is 6.55, training accuracy is 0.312500, time 55.330 samples/sec
+epoch 11, total_step 181100, total loss is 12.50 , inference loss is 5.95, weight deacy loss is 6.55, training accuracy is 0.531250, time 56.071 samples/sec
+epoch 11, total_step 181120, total loss is 13.13 , inference loss is 6.59, weight deacy loss is 6.55, training accuracy is 0.343750, time 56.272 samples/sec
+epoch 11, total_step 181140, total loss is 17.36 , inference loss is 10.81, weight deacy loss is 6.55, training accuracy is 0.343750, time 56.973 samples/sec
+epoch 11, total_step 181160, total loss is 16.98 , inference loss is 10.44, weight deacy loss is 6.55, training accuracy is 0.281250, time 56.685 samples/sec
+epoch 11, total_step 181180, total loss is 16.75 , inference loss is 10.20, weight deacy loss is 6.54, training accuracy is 0.250000, time 56.885 samples/sec
+epoch 11, total_step 181200, total loss is 20.65 , inference loss is 14.11, weight deacy loss is 6.54, training accuracy is 0.187500, time 56.818 samples/sec
+epoch 11, total_step 181220, total loss is 12.07 , inference loss is 5.53, weight deacy loss is 6.54, training accuracy is 0.500000, time 56.976 samples/sec
+epoch 11, total_step 181240, total loss is 16.73 , inference loss is 10.19, weight deacy loss is 6.54, training accuracy is 0.343750, time 56.027 samples/sec
+epoch 11, total_step 181260, total loss is 14.85 , inference loss is 8.31, weight deacy loss is 6.54, training accuracy is 0.406250, time 57.809 samples/sec
+epoch 11, total_step 181280, total loss is 15.93 , inference loss is 9.39, weight deacy loss is 6.54, training accuracy is 0.406250, time 56.554 samples/sec
+epoch 11, total_step 181300, total loss is 15.84 , inference loss is 9.30, weight deacy loss is 6.54, training accuracy is 0.468750, time 56.368 samples/sec
+epoch 11, total_step 181320, total loss is 13.90 , inference loss is 7.35, weight deacy loss is 6.54, training accuracy is 0.343750, time 56.035 samples/sec
+epoch 11, total_step 181340, total loss is 12.20 , inference loss is 5.65, weight deacy loss is 6.54, training accuracy is 0.593750, time 56.646 samples/sec
+epoch 11, total_step 181360, total loss is 17.56 , inference loss is 11.02, weight deacy loss is 6.54, training accuracy is 0.250000, time 56.263 samples/sec
+epoch 11, total_step 181380, total loss is 14.36 , inference loss is 7.82, weight deacy loss is 6.54, training accuracy is 0.312500, time 56.028 samples/sec
+epoch 11, total_step 181400, total loss is 17.46 , inference loss is 10.92, weight deacy loss is 6.54, training accuracy is 0.375000, time 56.740 samples/sec
+epoch 11, total_step 181420, total loss is 15.82 , inference loss is 9.28, weight deacy loss is 6.54, training accuracy is 0.375000, time 57.230 samples/sec
+epoch 11, total_step 181440, total loss is 15.87 , inference loss is 9.33, weight deacy loss is 6.54, training accuracy is 0.343750, time 56.413 samples/sec
+epoch 11, total_step 181460, total loss is 17.18 , inference loss is 10.64, weight deacy loss is 6.54, training accuracy is 0.156250, time 55.987 samples/sec
+epoch 11, total_step 181480, total loss is 12.21 , inference loss is 5.67, weight deacy loss is 6.54, training accuracy is 0.468750, time 57.220 samples/sec
+epoch 11, total_step 181500, total loss is 15.46 , inference loss is 8.92, weight deacy loss is 6.54, training accuracy is 0.312500, time 56.042 samples/sec
+epoch 11, total_step 181520, total loss is 16.23 , inference loss is 9.69, weight deacy loss is 6.54, training accuracy is 0.281250, time 56.264 samples/sec
+epoch 11, total_step 181540, total loss is 15.93 , inference loss is 9.39, weight deacy loss is 6.54, training accuracy is 0.406250, time 55.675 samples/sec
+epoch 11, total_step 181560, total loss is 13.14 , inference loss is 6.60, weight deacy loss is 6.54, training accuracy is 0.406250, time 57.489 samples/sec
+epoch 11, total_step 181580, total loss is 12.49 , inference loss is 5.95, weight deacy loss is 6.54, training accuracy is 0.406250, time 56.417 samples/sec
+epoch 11, total_step 181600, total loss is 15.42 , inference loss is 8.88, weight deacy loss is 6.54, training accuracy is 0.156250, time 56.540 samples/sec
+epoch 11, total_step 181620, total loss is 12.71 , inference loss is 6.17, weight deacy loss is 6.54, training accuracy is 0.375000, time 56.479 samples/sec
+epoch 11, total_step 181640, total loss is 14.48 , inference loss is 7.94, weight deacy loss is 6.54, training accuracy is 0.406250, time 56.550 samples/sec
+epoch 11, total_step 181660, total loss is 14.71 , inference loss is 8.17, weight deacy loss is 6.54, training accuracy is 0.468750, time 56.404 samples/sec
+epoch 11, total_step 181680, total loss is 15.74 , inference loss is 9.20, weight deacy loss is 6.54, training accuracy is 0.500000, time 55.204 samples/sec
+epoch 11, total_step 181700, total loss is 17.40 , inference loss is 10.86, weight deacy loss is 6.54, training accuracy is 0.250000, time 55.941 samples/sec
+epoch 11, total_step 181720, total loss is 11.08 , inference loss is 4.55, weight deacy loss is 6.54, training accuracy is 0.406250, time 56.760 samples/sec
+epoch 11, total_step 181740, total loss is 15.05 , inference loss is 8.51, weight deacy loss is 6.54, training accuracy is 0.250000, time 56.089 samples/sec
+epoch 11, total_step 181760, total loss is 18.94 , inference loss is 12.40, weight deacy loss is 6.54, training accuracy is 0.312500, time 56.444 samples/sec
+epoch 11, total_step 181780, total loss is 17.86 , inference loss is 11.32, weight deacy loss is 6.54, training accuracy is 0.250000, time 55.957 samples/sec
+epoch 11, total_step 181800, total loss is 13.64 , inference loss is 7.10, weight deacy loss is 6.54, training accuracy is 0.375000, time 56.361 samples/sec
+epoch 11, total_step 181820, total loss is 17.67 , inference loss is 11.13, weight deacy loss is 6.54, training accuracy is 0.406250, time 56.077 samples/sec
+epoch 11, total_step 181840, total loss is 11.73 , inference loss is 5.19, weight deacy loss is 6.54, training accuracy is 0.437500, time 56.280 samples/sec
+epoch 11, total_step 181860, total loss is 16.14 , inference loss is 9.60, weight deacy loss is 6.54, training accuracy is 0.375000, time 56.756 samples/sec
+epoch 11, total_step 181880, total loss is 17.78 , inference loss is 11.24, weight deacy loss is 6.54, training accuracy is 0.343750, time 56.899 samples/sec
+epoch 11, total_step 181900, total loss is 15.43 , inference loss is 8.89, weight deacy loss is 6.54, training accuracy is 0.531250, time 56.955 samples/sec
+epoch 11, total_step 181920, total loss is 13.27 , inference loss is 6.74, weight deacy loss is 6.54, training accuracy is 0.437500, time 55.907 samples/sec
+epoch 11, total_step 181940, total loss is 18.16 , inference loss is 11.63, weight deacy loss is 6.54, training accuracy is 0.343750, time 57.092 samples/sec
+epoch 11, total_step 181960, total loss is 16.09 , inference loss is 9.55, weight deacy loss is 6.54, training accuracy is 0.312500, time 56.422 samples/sec
+epoch 11, total_step 181980, total loss is 12.77 , inference loss is 6.23, weight deacy loss is 6.54, training accuracy is 0.437500, time 56.354 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.49335699999997
+best_threshold_index 110 0.9855555555555555
+best_threshold_index 111 0.9844444444444445
+best_threshold_index 109 0.9848148148148148
+best_threshold_index 109 0.9853703703703703
+best_threshold_index 109 0.985
+best_threshold_index 108 0.9837037037037037
+best_threshold_index 108 0.9842592592592593
+best_threshold_index 108 0.9842592592592593
+best_threshold_index 108 0.9827777777777778
+best_threshold_index 111 0.9838888888888889
+[lfw][182000]XNorm: 22.941111
+[lfw][182000]Accuracy-Flip: 0.98300+-0.00657
+testing verification..
+(14000, 512)
+infer time 85.74023200000002
+best_threshold_index 116 0.9866666666666667
+best_threshold_index 116 0.9853968253968254
+best_threshold_index 118 0.9858730158730159
+best_threshold_index 116 0.9857142857142858
+best_threshold_index 118 0.9874603174603175
+best_threshold_index 116 0.9855555555555555
+best_threshold_index 118 0.9861904761904762
+best_threshold_index 118 0.9852380952380952
+best_threshold_index 118 0.9855555555555555
+best_threshold_index 116 0.9865079365079366
+[cfp_ff][182000]XNorm: 22.654344
+[cfp_ff][182000]Accuracy-Flip: 0.98514+-0.00590
+testing verification..
+(14000, 512)
+infer time 85.35279299999995
+best_threshold_index 135 0.9253968253968254
+best_threshold_index 135 0.9258730158730158
+best_threshold_index 137 0.9244444444444444
+best_threshold_index 135 0.9280952380952381
+best_threshold_index 135 0.9285714285714286
+best_threshold_index 139 0.9246031746031746
+best_threshold_index 139 0.9230158730158731
+best_threshold_index 139 0.9246031746031746
+best_threshold_index 135 0.926031746031746
+best_threshold_index 135 0.9263492063492064
+[cfp_fp][182000]XNorm: 22.627341
+[cfp_fp][182000]Accuracy-Flip: 0.92329+-0.01340
+testing verification..
+(12000, 512)
+infer time 73.321154
+best_threshold_index 140 0.8998148148148148
+best_threshold_index 140 0.9003703703703704
+best_threshold_index 140 0.9022222222222223
+best_threshold_index 140 0.9033333333333333
+best_threshold_index 140 0.9083333333333333
+best_threshold_index 140 0.9027777777777778
+best_threshold_index 140 0.9042592592592592
+best_threshold_index 140 0.9074074074074074
+best_threshold_index 140 0.9053703703703704
+best_threshold_index 140 0.9044444444444445
+[agedb_30][182000]XNorm: 23.390916
+[agedb_30][182000]Accuracy-Flip: 0.90383+-0.02348
+test accuracy is: 0.983
+epoch 11, total_step 182000, total loss is 12.95 , inference loss is 6.41, weight deacy loss is 6.54, training accuracy is 0.312500, time 53.524 samples/sec
+epoch 11, total_step 182020, total loss is 13.93 , inference loss is 7.39, weight deacy loss is 6.54, training accuracy is 0.437500, time 56.150 samples/sec
+epoch 11, total_step 182040, total loss is 16.43 , inference loss is 9.90, weight deacy loss is 6.54, training accuracy is 0.343750, time 56.215 samples/sec
+epoch 11, total_step 182060, total loss is 15.00 , inference loss is 8.46, weight deacy loss is 6.54, training accuracy is 0.468750, time 57.029 samples/sec
+epoch 11, total_step 182080, total loss is 17.59 , inference loss is 11.05, weight deacy loss is 6.54, training accuracy is 0.343750, time 56.337 samples/sec
+epoch 11, total_step 182100, total loss is 17.59 , inference loss is 11.05, weight deacy loss is 6.53, training accuracy is 0.406250, time 56.943 samples/sec
+epoch 11, total_step 182120, total loss is 15.53 , inference loss is 9.00, weight deacy loss is 6.53, training accuracy is 0.406250, time 56.973 samples/sec
+epoch 11, total_step 182140, total loss is 19.47 , inference loss is 12.93, weight deacy loss is 6.53, training accuracy is 0.343750, time 56.694 samples/sec
+epoch 11, total_step 182160, total loss is 17.13 , inference loss is 10.59, weight deacy loss is 6.53, training accuracy is 0.250000, time 56.176 samples/sec
+epoch 11, total_step 182180, total loss is 14.98 , inference loss is 8.45, weight deacy loss is 6.53, training accuracy is 0.312500, time 56.107 samples/sec
+epoch 11, total_step 182200, total loss is 15.82 , inference loss is 9.29, weight deacy loss is 6.53, training accuracy is 0.437500, time 56.708 samples/sec
+epoch 11, total_step 182220, total loss is 13.65 , inference loss is 7.11, weight deacy loss is 6.53, training accuracy is 0.375000, time 55.923 samples/sec
+epoch 11, total_step 182240, total loss is 18.19 , inference loss is 11.66, weight deacy loss is 6.53, training accuracy is 0.250000, time 57.016 samples/sec
+epoch 11, total_step 182260, total loss is 13.92 , inference loss is 7.39, weight deacy loss is 6.53, training accuracy is 0.343750, time 57.377 samples/sec
+epoch 11, total_step 182280, total loss is 15.05 , inference loss is 8.52, weight deacy loss is 6.53, training accuracy is 0.468750, time 56.828 samples/sec
+epoch 11, total_step 182300, total loss is 14.31 , inference loss is 7.78, weight deacy loss is 6.53, training accuracy is 0.312500, time 56.930 samples/sec
+epoch 11, total_step 182320, total loss is 15.32 , inference loss is 8.79, weight deacy loss is 6.53, training accuracy is 0.406250, time 56.958 samples/sec
+epoch 11, total_step 182340, total loss is 12.46 , inference loss is 5.92, weight deacy loss is 6.53, training accuracy is 0.375000, time 55.655 samples/sec
+epoch 11, total_step 182360, total loss is 10.88 , inference loss is 4.35, weight deacy loss is 6.53, training accuracy is 0.500000, time 56.587 samples/sec
+epoch 11, total_step 182380, total loss is 14.50 , inference loss is 7.97, weight deacy loss is 6.53, training accuracy is 0.218750, time 56.047 samples/sec
+epoch 11, total_step 182400, total loss is 14.01 , inference loss is 7.47, weight deacy loss is 6.53, training accuracy is 0.375000, time 56.462 samples/sec
+epoch 11, total_step 182420, total loss is 19.91 , inference loss is 13.38, weight deacy loss is 6.53, training accuracy is 0.218750, time 57.089 samples/sec
+epoch 11, total_step 182440, total loss is 16.40 , inference loss is 9.87, weight deacy loss is 6.53, training accuracy is 0.312500, time 55.814 samples/sec
+epoch 11, total_step 182460, total loss is 17.41 , inference loss is 10.88, weight deacy loss is 6.53, training accuracy is 0.375000, time 56.722 samples/sec
+epoch 11, total_step 182480, total loss is 14.42 , inference loss is 7.89, weight deacy loss is 6.53, training accuracy is 0.343750, time 56.482 samples/sec
+epoch 11, total_step 182500, total loss is 19.11 , inference loss is 12.58, weight deacy loss is 6.53, training accuracy is 0.406250, time 57.126 samples/sec
+epoch 11, total_step 182520, total loss is 13.50 , inference loss is 6.97, weight deacy loss is 6.53, training accuracy is 0.281250, time 57.657 samples/sec
+epoch 11, total_step 182540, total loss is 14.86 , inference loss is 8.33, weight deacy loss is 6.53, training accuracy is 0.343750, time 55.732 samples/sec
+epoch 11, total_step 182560, total loss is 17.68 , inference loss is 11.15, weight deacy loss is 6.53, training accuracy is 0.468750, time 55.973 samples/sec
+epoch 11, total_step 182580, total loss is 13.70 , inference loss is 7.17, weight deacy loss is 6.53, training accuracy is 0.500000, time 56.088 samples/sec
+epoch 11, total_step 182600, total loss is 13.46 , inference loss is 6.93, weight deacy loss is 6.53, training accuracy is 0.375000, time 55.785 samples/sec
+epoch 11, total_step 182620, total loss is 13.14 , inference loss is 6.62, weight deacy loss is 6.53, training accuracy is 0.375000, time 56.669 samples/sec
+epoch 11, total_step 182640, total loss is 17.31 , inference loss is 10.78, weight deacy loss is 6.53, training accuracy is 0.281250, time 56.660 samples/sec
+epoch 11, total_step 182660, total loss is 16.29 , inference loss is 9.76, weight deacy loss is 6.53, training accuracy is 0.250000, time 56.270 samples/sec
+epoch 11, total_step 182680, total loss is 17.56 , inference loss is 11.03, weight deacy loss is 6.53, training accuracy is 0.218750, time 56.526 samples/sec
+epoch 11, total_step 182700, total loss is 10.98 , inference loss is 4.45, weight deacy loss is 6.53, training accuracy is 0.531250, time 57.078 samples/sec
+epoch 11, total_step 182720, total loss is 18.98 , inference loss is 12.45, weight deacy loss is 6.53, training accuracy is 0.187500, time 56.908 samples/sec
+epoch 11, total_step 182740, total loss is 20.80 , inference loss is 14.27, weight deacy loss is 6.53, training accuracy is 0.281250, time 57.023 samples/sec
+epoch 11, total_step 182760, total loss is 14.32 , inference loss is 7.79, weight deacy loss is 6.53, training accuracy is 0.343750, time 56.546 samples/sec
+epoch 11, total_step 182780, total loss is 19.23 , inference loss is 12.70, weight deacy loss is 6.53, training accuracy is 0.312500, time 56.585 samples/sec
+epoch 11, total_step 182800, total loss is 14.97 , inference loss is 8.44, weight deacy loss is 6.53, training accuracy is 0.250000, time 57.044 samples/sec
+epoch 11, total_step 182820, total loss is 14.25 , inference loss is 7.72, weight deacy loss is 6.53, training accuracy is 0.437500, time 56.176 samples/sec
+epoch 11, total_step 182840, total loss is 15.70 , inference loss is 9.17, weight deacy loss is 6.53, training accuracy is 0.375000, time 56.733 samples/sec
+epoch 11, total_step 182860, total loss is 19.05 , inference loss is 12.53, weight deacy loss is 6.53, training accuracy is 0.250000, time 55.745 samples/sec
+epoch 11, total_step 182880, total loss is 15.01 , inference loss is 8.49, weight deacy loss is 6.53, training accuracy is 0.375000, time 56.615 samples/sec
+epoch 11, total_step 182900, total loss is 19.75 , inference loss is 13.23, weight deacy loss is 6.53, training accuracy is 0.250000, time 57.345 samples/sec
+epoch 11, total_step 182920, total loss is 17.85 , inference loss is 11.32, weight deacy loss is 6.53, training accuracy is 0.312500, time 56.664 samples/sec
+epoch 11, total_step 182940, total loss is 13.36 , inference loss is 6.83, weight deacy loss is 6.53, training accuracy is 0.562500, time 56.456 samples/sec
+epoch 11, total_step 182960, total loss is 14.55 , inference loss is 8.03, weight deacy loss is 6.53, training accuracy is 0.468750, time 56.113 samples/sec
+epoch 11, total_step 182980, total loss is 15.12 , inference loss is 8.59, weight deacy loss is 6.53, training accuracy is 0.343750, time 56.192 samples/sec
+epoch 11, total_step 183000, total loss is 10.58 , inference loss is 4.05, weight deacy loss is 6.53, training accuracy is 0.562500, time 56.200 samples/sec
+epoch 11, total_step 183020, total loss is 14.50 , inference loss is 7.98, weight deacy loss is 6.53, training accuracy is 0.375000, time 56.491 samples/sec
+epoch 11, total_step 183040, total loss is 14.86 , inference loss is 8.34, weight deacy loss is 6.53, training accuracy is 0.312500, time 56.515 samples/sec
+epoch 11, total_step 183060, total loss is 16.59 , inference loss is 10.07, weight deacy loss is 6.52, training accuracy is 0.437500, time 55.845 samples/sec
+epoch 11, total_step 183080, total loss is 15.90 , inference loss is 9.38, weight deacy loss is 6.52, training accuracy is 0.218750, time 56.685 samples/sec
+epoch 11, total_step 183100, total loss is 16.28 , inference loss is 9.75, weight deacy loss is 6.52, training accuracy is 0.375000, time 56.927 samples/sec
+epoch 11, total_step 183120, total loss is 22.23 , inference loss is 15.70, weight deacy loss is 6.52, training accuracy is 0.281250, time 56.145 samples/sec
+epoch 11, total_step 183140, total loss is 15.25 , inference loss is 8.73, weight deacy loss is 6.52, training accuracy is 0.406250, time 56.190 samples/sec
+epoch 11, total_step 183160, total loss is 12.61 , inference loss is 6.09, weight deacy loss is 6.52, training accuracy is 0.375000, time 55.757 samples/sec
+epoch 11, total_step 183180, total loss is 13.48 , inference loss is 6.95, weight deacy loss is 6.52, training accuracy is 0.468750, time 56.494 samples/sec
+epoch 11, total_step 183200, total loss is 11.79 , inference loss is 5.27, weight deacy loss is 6.52, training accuracy is 0.406250, time 57.190 samples/sec
+epoch 11, total_step 183220, total loss is 13.82 , inference loss is 7.30, weight deacy loss is 6.52, training accuracy is 0.437500, time 55.517 samples/sec
+epoch 11, total_step 183240, total loss is 19.45 , inference loss is 12.92, weight deacy loss is 6.52, training accuracy is 0.281250, time 56.307 samples/sec
+epoch 11, total_step 183260, total loss is 15.69 , inference loss is 9.17, weight deacy loss is 6.52, training accuracy is 0.218750, time 56.291 samples/sec
+epoch 11, total_step 183280, total loss is 15.37 , inference loss is 8.85, weight deacy loss is 6.52, training accuracy is 0.437500, time 57.025 samples/sec
+epoch 11, total_step 183300, total loss is 16.74 , inference loss is 10.22, weight deacy loss is 6.52, training accuracy is 0.250000, time 56.216 samples/sec
+epoch 11, total_step 183320, total loss is 14.05 , inference loss is 7.53, weight deacy loss is 6.52, training accuracy is 0.375000, time 56.743 samples/sec
+epoch 11, total_step 183340, total loss is 17.32 , inference loss is 10.80, weight deacy loss is 6.52, training accuracy is 0.250000, time 56.699 samples/sec
+epoch 11, total_step 183360, total loss is 16.66 , inference loss is 10.13, weight deacy loss is 6.52, training accuracy is 0.343750, time 57.038 samples/sec
+epoch 11, total_step 183380, total loss is 18.72 , inference loss is 12.20, weight deacy loss is 6.52, training accuracy is 0.312500, time 56.078 samples/sec
+epoch 11, total_step 183400, total loss is 18.10 , inference loss is 11.58, weight deacy loss is 6.52, training accuracy is 0.312500, time 56.521 samples/sec
+epoch 11, total_step 183420, total loss is 15.28 , inference loss is 8.76, weight deacy loss is 6.52, training accuracy is 0.250000, time 56.218 samples/sec
+epoch 11, total_step 183440, total loss is 17.83 , inference loss is 11.31, weight deacy loss is 6.52, training accuracy is 0.343750, time 56.026 samples/sec
+epoch 11, total_step 183460, total loss is 15.27 , inference loss is 8.75, weight deacy loss is 6.52, training accuracy is 0.312500, time 56.671 samples/sec
+epoch 11, total_step 183480, total loss is 16.38 , inference loss is 9.86, weight deacy loss is 6.52, training accuracy is 0.281250, time 56.629 samples/sec
+epoch 11, total_step 183500, total loss is 16.61 , inference loss is 10.09, weight deacy loss is 6.52, training accuracy is 0.250000, time 56.549 samples/sec
+epoch 11, total_step 183520, total loss is 15.58 , inference loss is 9.06, weight deacy loss is 6.52, training accuracy is 0.250000, time 56.125 samples/sec
+epoch 11, total_step 183540, total loss is 14.14 , inference loss is 7.62, weight deacy loss is 6.52, training accuracy is 0.343750, time 55.904 samples/sec
+epoch 11, total_step 183560, total loss is 14.40 , inference loss is 7.88, weight deacy loss is 6.52, training accuracy is 0.375000, time 56.358 samples/sec
+epoch 11, total_step 183580, total loss is 15.68 , inference loss is 9.17, weight deacy loss is 6.52, training accuracy is 0.281250, time 56.106 samples/sec
+epoch 11, total_step 183600, total loss is 25.21 , inference loss is 18.70, weight deacy loss is 6.52, training accuracy is 0.125000, time 56.383 samples/sec
+epoch 11, total_step 183620, total loss is 15.81 , inference loss is 9.29, weight deacy loss is 6.52, training accuracy is 0.343750, time 56.674 samples/sec
+epoch 11, total_step 183640, total loss is 16.55 , inference loss is 10.03, weight deacy loss is 6.52, training accuracy is 0.218750, time 56.370 samples/sec
+epoch 11, total_step 183660, total loss is 13.11 , inference loss is 6.59, weight deacy loss is 6.52, training accuracy is 0.375000, time 57.616 samples/sec
+epoch 11, total_step 183680, total loss is 13.55 , inference loss is 7.03, weight deacy loss is 6.52, training accuracy is 0.437500, time 56.335 samples/sec
+epoch 11, total_step 183700, total loss is 16.37 , inference loss is 9.85, weight deacy loss is 6.52, training accuracy is 0.375000, time 56.640 samples/sec
+epoch 11, total_step 183720, total loss is 14.41 , inference loss is 7.89, weight deacy loss is 6.52, training accuracy is 0.343750, time 56.672 samples/sec
+epoch 11, total_step 183740, total loss is 16.66 , inference loss is 10.14, weight deacy loss is 6.52, training accuracy is 0.218750, time 56.445 samples/sec
+epoch 11, total_step 183760, total loss is 12.95 , inference loss is 6.44, weight deacy loss is 6.52, training accuracy is 0.343750, time 56.530 samples/sec
+epoch 11, total_step 183780, total loss is 15.59 , inference loss is 9.08, weight deacy loss is 6.52, training accuracy is 0.343750, time 56.028 samples/sec
+epoch 11, total_step 183800, total loss is 13.25 , inference loss is 6.74, weight deacy loss is 6.52, training accuracy is 0.218750, time 56.593 samples/sec
+epoch 11, total_step 183820, total loss is 20.44 , inference loss is 13.92, weight deacy loss is 6.52, training accuracy is 0.375000, time 56.700 samples/sec
+epoch 11, total_step 183840, total loss is 14.06 , inference loss is 7.54, weight deacy loss is 6.52, training accuracy is 0.375000, time 56.903 samples/sec
+epoch 11, total_step 183860, total loss is 16.02 , inference loss is 9.51, weight deacy loss is 6.52, training accuracy is 0.343750, time 56.330 samples/sec
+epoch 11, total_step 183880, total loss is 14.53 , inference loss is 8.01, weight deacy loss is 6.52, training accuracy is 0.343750, time 56.496 samples/sec
+epoch 11, total_step 183900, total loss is 13.69 , inference loss is 7.17, weight deacy loss is 6.52, training accuracy is 0.312500, time 56.252 samples/sec
+epoch 11, total_step 183920, total loss is 15.73 , inference loss is 9.21, weight deacy loss is 6.52, training accuracy is 0.218750, time 56.086 samples/sec
+epoch 11, total_step 183940, total loss is 15.91 , inference loss is 9.40, weight deacy loss is 6.52, training accuracy is 0.437500, time 56.000 samples/sec
+epoch 11, total_step 183960, total loss is 12.05 , inference loss is 5.54, weight deacy loss is 6.52, training accuracy is 0.468750, time 55.084 samples/sec
+epoch 11, total_step 183980, total loss is 13.07 , inference loss is 6.56, weight deacy loss is 6.52, training accuracy is 0.500000, time 55.905 samples/sec
+
+
+
+
+End of epoch 11
+testing verification..
+(12000, 512)
+infer time 73.262775
+best_threshold_index 116 0.9861111111111112
+best_threshold_index 113 0.9857407407407407
+best_threshold_index 113 0.9861111111111112
+best_threshold_index 112 0.9862962962962963
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 113 0.9855555555555555
+best_threshold_index 116 0.9862962962962963
+best_threshold_index 113 0.9864814814814815
+best_threshold_index 113 0.9853703703703703
+best_threshold_index 113 0.9851851851851852
+[lfw][184000]XNorm: 19.734588
+[lfw][184000]Accuracy-Flip: 0.98483+-0.00411
+testing verification..
+(14000, 512)
+infer time 85.45040399999998
+best_threshold_index 110 0.9884126984126984
+best_threshold_index 110 0.9879365079365079
+best_threshold_index 110 0.9879365079365079
+best_threshold_index 110 0.9868253968253968
+best_threshold_index 110 0.9874603174603175
+best_threshold_index 110 0.986984126984127
+best_threshold_index 110 0.9880952380952381
+best_threshold_index 110 0.9877777777777778
+best_threshold_index 110 0.9876190476190476
+best_threshold_index 110 0.9880952380952381
+[cfp_ff][184000]XNorm: 18.986046
+[cfp_ff][184000]Accuracy-Flip: 0.98771+-0.00430
+testing verification..
+(14000, 512)
+infer time 85.56074300000007
+best_threshold_index 133 0.9182539682539682
+best_threshold_index 130 0.9179365079365079
+best_threshold_index 133 0.9173015873015873
+best_threshold_index 133 0.9180952380952381
+best_threshold_index 134 0.922063492063492
+best_threshold_index 133 0.9174603174603174
+best_threshold_index 133 0.9152380952380952
+best_threshold_index 133 0.9182539682539682
+best_threshold_index 130 0.9185714285714286
+best_threshold_index 135 0.9193650793650794
+[cfp_fp][184000]XNorm: 18.993781
+[cfp_fp][184000]Accuracy-Flip: 0.91543+-0.01548
+testing verification..
+(12000, 512)
+infer time 73.53584799999999
+best_threshold_index 134 0.9044444444444445
+best_threshold_index 133 0.9011111111111111
+best_threshold_index 133 0.9061111111111111
+best_threshold_index 134 0.9057407407407407
+best_threshold_index 133 0.9077777777777778
+best_threshold_index 134 0.9037037037037037
+best_threshold_index 132 0.9061111111111111
+best_threshold_index 133 0.9087037037037037
+best_threshold_index 134 0.9051851851851852
+best_threshold_index 134 0.9055555555555556
+[agedb_30][184000]XNorm: 19.634198
+[agedb_30][184000]Accuracy-Flip: 0.90350+-0.01843
+test accuracy is: 0.9848333333333332
+epoch 12, total_step 184000, total loss is 13.38 , inference loss is 6.87, weight deacy loss is 6.52, training accuracy is 0.375000, time 54.028 samples/sec
+epoch 12, total_step 184020, total loss is 8.57 , inference loss is 2.05, weight deacy loss is 6.52, training accuracy is 0.750000, time 56.803 samples/sec
+epoch 12, total_step 184040, total loss is 11.62 , inference loss is 5.10, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.280 samples/sec
+epoch 12, total_step 184060, total loss is 12.63 , inference loss is 6.12, weight deacy loss is 6.51, training accuracy is 0.656250, time 56.012 samples/sec
+epoch 12, total_step 184080, total loss is 12.94 , inference loss is 6.42, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.394 samples/sec
+epoch 12, total_step 184100, total loss is 11.06 , inference loss is 4.55, weight deacy loss is 6.51, training accuracy is 0.718750, time 57.017 samples/sec
+epoch 12, total_step 184120, total loss is 10.44 , inference loss is 3.93, weight deacy loss is 6.51, training accuracy is 0.562500, time 57.516 samples/sec
+epoch 12, total_step 184140, total loss is 11.23 , inference loss is 4.72, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.245 samples/sec
+epoch 12, total_step 184160, total loss is 10.16 , inference loss is 3.64, weight deacy loss is 6.51, training accuracy is 0.750000, time 56.259 samples/sec
+epoch 12, total_step 184180, total loss is 9.96 , inference loss is 3.45, weight deacy loss is 6.51, training accuracy is 0.750000, time 55.714 samples/sec
+epoch 12, total_step 184200, total loss is 8.85 , inference loss is 2.34, weight deacy loss is 6.51, training accuracy is 0.812500, time 56.829 samples/sec
+epoch 12, total_step 184220, total loss is 10.20 , inference loss is 3.69, weight deacy loss is 6.51, training accuracy is 0.656250, time 56.708 samples/sec
+epoch 12, total_step 184240, total loss is 9.07 , inference loss is 2.56, weight deacy loss is 6.51, training accuracy is 0.593750, time 55.975 samples/sec
+epoch 12, total_step 184260, total loss is 11.06 , inference loss is 4.55, weight deacy loss is 6.51, training accuracy is 0.625000, time 56.295 samples/sec
+epoch 12, total_step 184280, total loss is 8.64 , inference loss is 2.13, weight deacy loss is 6.51, training accuracy is 0.781250, time 56.233 samples/sec
+epoch 12, total_step 184300, total loss is 11.15 , inference loss is 4.64, weight deacy loss is 6.51, training accuracy is 0.656250, time 56.668 samples/sec
+epoch 12, total_step 184320, total loss is 13.55 , inference loss is 7.03, weight deacy loss is 6.51, training accuracy is 0.625000, time 56.024 samples/sec
+epoch 12, total_step 184340, total loss is 10.81 , inference loss is 4.30, weight deacy loss is 6.51, training accuracy is 0.656250, time 56.056 samples/sec
+epoch 12, total_step 184360, total loss is 10.28 , inference loss is 3.77, weight deacy loss is 6.51, training accuracy is 0.625000, time 56.351 samples/sec
+epoch 12, total_step 184380, total loss is 11.39 , inference loss is 4.87, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.012 samples/sec
+epoch 12, total_step 184400, total loss is 10.53 , inference loss is 4.02, weight deacy loss is 6.51, training accuracy is 0.750000, time 57.004 samples/sec
+epoch 12, total_step 184420, total loss is 16.47 , inference loss is 9.96, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.773 samples/sec
+epoch 12, total_step 184440, total loss is 9.74 , inference loss is 3.23, weight deacy loss is 6.51, training accuracy is 0.750000, time 56.591 samples/sec
+epoch 12, total_step 184460, total loss is 11.36 , inference loss is 4.85, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.925 samples/sec
+epoch 12, total_step 184480, total loss is 11.41 , inference loss is 4.90, weight deacy loss is 6.51, training accuracy is 0.625000, time 55.847 samples/sec
+epoch 12, total_step 184500, total loss is 11.04 , inference loss is 4.52, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.644 samples/sec
+epoch 12, total_step 184520, total loss is 11.95 , inference loss is 5.44, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.708 samples/sec
+epoch 12, total_step 184540, total loss is 11.28 , inference loss is 4.77, weight deacy loss is 6.51, training accuracy is 0.656250, time 56.288 samples/sec
+epoch 12, total_step 184560, total loss is 12.26 , inference loss is 5.75, weight deacy loss is 6.51, training accuracy is 0.718750, time 57.168 samples/sec
+epoch 12, total_step 184580, total loss is 13.96 , inference loss is 7.45, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.644 samples/sec
+epoch 12, total_step 184600, total loss is 13.53 , inference loss is 7.02, weight deacy loss is 6.51, training accuracy is 0.531250, time 57.147 samples/sec
+epoch 12, total_step 184620, total loss is 8.41 , inference loss is 1.90, weight deacy loss is 6.51, training accuracy is 0.718750, time 56.971 samples/sec
+epoch 12, total_step 184640, total loss is 13.00 , inference loss is 6.49, weight deacy loss is 6.51, training accuracy is 0.656250, time 56.717 samples/sec
+epoch 12, total_step 184660, total loss is 19.53 , inference loss is 13.01, weight deacy loss is 6.51, training accuracy is 0.406250, time 56.892 samples/sec
+epoch 12, total_step 184680, total loss is 15.08 , inference loss is 8.57, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.653 samples/sec
+epoch 12, total_step 184700, total loss is 10.16 , inference loss is 3.65, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.751 samples/sec
+epoch 12, total_step 184720, total loss is 11.60 , inference loss is 5.09, weight deacy loss is 6.51, training accuracy is 0.718750, time 56.355 samples/sec
+epoch 12, total_step 184740, total loss is 14.35 , inference loss is 7.84, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.751 samples/sec
+epoch 12, total_step 184760, total loss is 10.09 , inference loss is 3.57, weight deacy loss is 6.51, training accuracy is 0.718750, time 55.935 samples/sec
+epoch 12, total_step 184780, total loss is 13.17 , inference loss is 6.66, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.180 samples/sec
+epoch 12, total_step 184800, total loss is 18.79 , inference loss is 12.27, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.578 samples/sec
+epoch 12, total_step 184820, total loss is 13.47 , inference loss is 6.95, weight deacy loss is 6.51, training accuracy is 0.406250, time 56.080 samples/sec
+epoch 12, total_step 184840, total loss is 12.19 , inference loss is 5.68, weight deacy loss is 6.51, training accuracy is 0.625000, time 56.609 samples/sec
+epoch 12, total_step 184860, total loss is 17.29 , inference loss is 10.77, weight deacy loss is 6.51, training accuracy is 0.375000, time 56.212 samples/sec
+epoch 12, total_step 184880, total loss is 13.55 , inference loss is 7.04, weight deacy loss is 6.51, training accuracy is 0.593750, time 57.286 samples/sec
+epoch 12, total_step 184900, total loss is 11.86 , inference loss is 5.34, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.876 samples/sec
+epoch 12, total_step 184920, total loss is 10.96 , inference loss is 4.44, weight deacy loss is 6.51, training accuracy is 0.656250, time 56.837 samples/sec
+epoch 12, total_step 184940, total loss is 14.00 , inference loss is 7.49, weight deacy loss is 6.51, training accuracy is 0.437500, time 55.968 samples/sec
+epoch 12, total_step 184960, total loss is 13.54 , inference loss is 7.03, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.774 samples/sec
+epoch 12, total_step 184980, total loss is 11.64 , inference loss is 5.13, weight deacy loss is 6.51, training accuracy is 0.500000, time 57.380 samples/sec
+epoch 12, total_step 185000, total loss is 14.74 , inference loss is 8.23, weight deacy loss is 6.51, training accuracy is 0.656250, time 56.965 samples/sec
+epoch 12, total_step 185020, total loss is 13.49 , inference loss is 6.97, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.812 samples/sec
+epoch 12, total_step 185040, total loss is 15.91 , inference loss is 9.39, weight deacy loss is 6.52, training accuracy is 0.437500, time 56.456 samples/sec
+epoch 12, total_step 185060, total loss is 11.45 , inference loss is 4.93, weight deacy loss is 6.52, training accuracy is 0.625000, time 56.513 samples/sec
+epoch 12, total_step 185080, total loss is 15.19 , inference loss is 8.67, weight deacy loss is 6.52, training accuracy is 0.437500, time 56.889 samples/sec
+epoch 12, total_step 185100, total loss is 15.23 , inference loss is 8.71, weight deacy loss is 6.52, training accuracy is 0.468750, time 55.881 samples/sec
+epoch 12, total_step 185120, total loss is 11.80 , inference loss is 5.28, weight deacy loss is 6.52, training accuracy is 0.625000, time 56.955 samples/sec
+epoch 12, total_step 185140, total loss is 15.36 , inference loss is 8.84, weight deacy loss is 6.52, training accuracy is 0.562500, time 56.848 samples/sec
+epoch 12, total_step 185160, total loss is 14.71 , inference loss is 8.19, weight deacy loss is 6.52, training accuracy is 0.468750, time 56.577 samples/sec
+epoch 12, total_step 185180, total loss is 13.10 , inference loss is 6.58, weight deacy loss is 6.52, training accuracy is 0.593750, time 57.103 samples/sec
+epoch 12, total_step 185200, total loss is 13.15 , inference loss is 6.64, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.943 samples/sec
+epoch 12, total_step 185220, total loss is 13.05 , inference loss is 6.54, weight deacy loss is 6.52, training accuracy is 0.531250, time 56.675 samples/sec
+epoch 12, total_step 185240, total loss is 15.15 , inference loss is 8.63, weight deacy loss is 6.52, training accuracy is 0.625000, time 57.053 samples/sec
+epoch 12, total_step 185260, total loss is 13.47 , inference loss is 6.96, weight deacy loss is 6.52, training accuracy is 0.468750, time 56.278 samples/sec
+epoch 12, total_step 185280, total loss is 16.75 , inference loss is 10.23, weight deacy loss is 6.52, training accuracy is 0.343750, time 55.909 samples/sec
+epoch 12, total_step 185300, total loss is 9.67 , inference loss is 3.15, weight deacy loss is 6.52, training accuracy is 0.625000, time 57.149 samples/sec
+epoch 12, total_step 185320, total loss is 10.66 , inference loss is 4.14, weight deacy loss is 6.52, training accuracy is 0.625000, time 56.452 samples/sec
+epoch 12, total_step 185340, total loss is 13.41 , inference loss is 6.90, weight deacy loss is 6.52, training accuracy is 0.593750, time 55.808 samples/sec
+epoch 12, total_step 185360, total loss is 17.45 , inference loss is 10.93, weight deacy loss is 6.52, training accuracy is 0.468750, time 55.428 samples/sec
+epoch 12, total_step 185380, total loss is 12.75 , inference loss is 6.23, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.920 samples/sec
+epoch 12, total_step 185400, total loss is 19.34 , inference loss is 12.82, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.608 samples/sec
+epoch 12, total_step 185420, total loss is 10.27 , inference loss is 3.75, weight deacy loss is 6.52, training accuracy is 0.562500, time 56.855 samples/sec
+epoch 12, total_step 185440, total loss is 10.14 , inference loss is 3.63, weight deacy loss is 6.52, training accuracy is 0.687500, time 55.777 samples/sec
+epoch 12, total_step 185460, total loss is 15.42 , inference loss is 8.90, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.418 samples/sec
+epoch 12, total_step 185480, total loss is 12.03 , inference loss is 5.52, weight deacy loss is 6.52, training accuracy is 0.625000, time 57.048 samples/sec
+epoch 12, total_step 185500, total loss is 16.66 , inference loss is 10.14, weight deacy loss is 6.52, training accuracy is 0.531250, time 56.199 samples/sec
+epoch 12, total_step 185520, total loss is 10.87 , inference loss is 4.35, weight deacy loss is 6.52, training accuracy is 0.656250, time 56.143 samples/sec
+epoch 12, total_step 185540, total loss is 14.01 , inference loss is 7.49, weight deacy loss is 6.52, training accuracy is 0.531250, time 56.514 samples/sec
+epoch 12, total_step 185560, total loss is 13.66 , inference loss is 7.14, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.233 samples/sec
+epoch 12, total_step 185580, total loss is 12.82 , inference loss is 6.30, weight deacy loss is 6.52, training accuracy is 0.500000, time 55.577 samples/sec
+epoch 12, total_step 185600, total loss is 12.34 , inference loss is 5.82, weight deacy loss is 6.52, training accuracy is 0.593750, time 55.671 samples/sec
+epoch 12, total_step 185620, total loss is 15.35 , inference loss is 8.83, weight deacy loss is 6.52, training accuracy is 0.406250, time 56.730 samples/sec
+epoch 12, total_step 185640, total loss is 13.69 , inference loss is 7.17, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.517 samples/sec
+epoch 12, total_step 185660, total loss is 10.16 , inference loss is 3.64, weight deacy loss is 6.52, training accuracy is 0.531250, time 56.200 samples/sec
+epoch 12, total_step 185680, total loss is 10.47 , inference loss is 3.96, weight deacy loss is 6.52, training accuracy is 0.718750, time 55.957 samples/sec
+epoch 12, total_step 185700, total loss is 9.78 , inference loss is 3.26, weight deacy loss is 6.52, training accuracy is 0.593750, time 55.289 samples/sec
+epoch 12, total_step 185720, total loss is 11.69 , inference loss is 5.17, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.722 samples/sec
+epoch 12, total_step 185740, total loss is 14.45 , inference loss is 7.93, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.503 samples/sec
+epoch 12, total_step 185760, total loss is 16.22 , inference loss is 9.70, weight deacy loss is 6.52, training accuracy is 0.437500, time 56.527 samples/sec
+epoch 12, total_step 185780, total loss is 10.08 , inference loss is 3.56, weight deacy loss is 6.52, training accuracy is 0.718750, time 57.144 samples/sec
+epoch 12, total_step 185800, total loss is 12.10 , inference loss is 5.59, weight deacy loss is 6.52, training accuracy is 0.531250, time 57.241 samples/sec
+epoch 12, total_step 185820, total loss is 11.94 , inference loss is 5.42, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.357 samples/sec
+epoch 12, total_step 185840, total loss is 14.74 , inference loss is 8.22, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.664 samples/sec
+epoch 12, total_step 185860, total loss is 13.30 , inference loss is 6.79, weight deacy loss is 6.52, training accuracy is 0.562500, time 57.044 samples/sec
+epoch 12, total_step 185880, total loss is 11.49 , inference loss is 4.97, weight deacy loss is 6.52, training accuracy is 0.687500, time 56.548 samples/sec
+epoch 12, total_step 185900, total loss is 15.34 , inference loss is 8.82, weight deacy loss is 6.52, training accuracy is 0.562500, time 56.088 samples/sec
+epoch 12, total_step 185920, total loss is 13.70 , inference loss is 7.18, weight deacy loss is 6.52, training accuracy is 0.468750, time 55.524 samples/sec
+epoch 12, total_step 185940, total loss is 12.75 , inference loss is 6.23, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.232 samples/sec
+epoch 12, total_step 185960, total loss is 10.87 , inference loss is 4.35, weight deacy loss is 6.52, training accuracy is 0.687500, time 56.084 samples/sec
+epoch 12, total_step 185980, total loss is 17.31 , inference loss is 10.80, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.310 samples/sec
+testing verification..
+(12000, 512)
+infer time 72.41242799999996
+best_threshold_index 120 0.9879629629629629
+best_threshold_index 120 0.9872222222222222
+best_threshold_index 121 0.9877777777777778
+best_threshold_index 120 0.9868518518518519
+best_threshold_index 120 0.9877777777777778
+best_threshold_index 120 0.9864814814814815
+best_threshold_index 120 0.9879629629629629
+best_threshold_index 120 0.987037037037037
+best_threshold_index 120 0.9859259259259259
+best_threshold_index 120 0.9868518518518519
+[lfw][186000]XNorm: 21.277325
+[lfw][186000]Accuracy-Flip: 0.98700+-0.00586
+testing verification..
+(14000, 512)
+infer time 85.81587899999998
+best_threshold_index 117 0.9855555555555555
+best_threshold_index 124 0.9846031746031746
+best_threshold_index 124 0.9852380952380952
+best_threshold_index 124 0.9841269841269841
+best_threshold_index 124 0.9855555555555555
+best_threshold_index 124 0.9841269841269841
+best_threshold_index 124 0.9858730158730159
+best_threshold_index 124 0.9852380952380952
+best_threshold_index 124 0.9852380952380952
+best_threshold_index 124 0.9858730158730159
+[cfp_ff][186000]XNorm: 20.352212
+[cfp_ff][186000]Accuracy-Flip: 0.98443+-0.00637
+testing verification..
+(14000, 512)
+infer time 85.66637499999997
+best_threshold_index 139 0.9255555555555556
+best_threshold_index 139 0.9234920634920635
+best_threshold_index 140 0.9223809523809524
+best_threshold_index 139 0.9249206349206349
+best_threshold_index 140 0.9255555555555556
+best_threshold_index 140 0.9214285714285714
+best_threshold_index 140 0.9204761904761904
+best_threshold_index 139 0.9204761904761904
+best_threshold_index 143 0.9230158730158731
+best_threshold_index 140 0.9247619047619048
+[cfp_fp][186000]XNorm: 20.998159
+[cfp_fp][186000]Accuracy-Flip: 0.92129+-0.01708
+testing verification..
+(12000, 512)
+infer time 73.53886800000001
+best_threshold_index 142 0.9074074074074074
+best_threshold_index 142 0.9070370370370371
+best_threshold_index 141 0.9114814814814814
+best_threshold_index 142 0.9094444444444445
+best_threshold_index 142 0.9146296296296297
+best_threshold_index 141 0.9087037037037037
+best_threshold_index 141 0.9094444444444445
+best_threshold_index 141 0.9125925925925926
+best_threshold_index 141 0.9120370370370371
+best_threshold_index 142 0.9107407407407407
+[agedb_30][186000]XNorm: 22.444455
+[agedb_30][186000]Accuracy-Flip: 0.90933+-0.02065
+test accuracy is: 0.9870000000000001
+epoch 12, total_step 186000, total loss is 11.24 , inference loss is 4.73, weight deacy loss is 6.52, training accuracy is 0.562500, time 53.469 samples/sec
+epoch 12, total_step 186020, total loss is 10.71 , inference loss is 4.19, weight deacy loss is 6.52, training accuracy is 0.687500, time 56.866 samples/sec
+epoch 12, total_step 186040, total loss is 12.30 , inference loss is 5.78, weight deacy loss is 6.52, training accuracy is 0.562500, time 56.591 samples/sec
+epoch 12, total_step 186060, total loss is 11.17 , inference loss is 4.65, weight deacy loss is 6.52, training accuracy is 0.625000, time 56.325 samples/sec
+epoch 12, total_step 186080, total loss is 15.12 , inference loss is 8.60, weight deacy loss is 6.52, training accuracy is 0.437500, time 56.083 samples/sec
+epoch 12, total_step 186100, total loss is 13.02 , inference loss is 6.50, weight deacy loss is 6.52, training accuracy is 0.625000, time 56.319 samples/sec
+epoch 12, total_step 186120, total loss is 12.38 , inference loss is 5.86, weight deacy loss is 6.52, training accuracy is 0.468750, time 56.545 samples/sec
+epoch 12, total_step 186140, total loss is 11.13 , inference loss is 4.61, weight deacy loss is 6.52, training accuracy is 0.656250, time 56.547 samples/sec
+epoch 12, total_step 186160, total loss is 12.04 , inference loss is 5.52, weight deacy loss is 6.52, training accuracy is 0.625000, time 56.298 samples/sec
+epoch 12, total_step 186180, total loss is 11.57 , inference loss is 5.05, weight deacy loss is 6.52, training accuracy is 0.687500, time 56.779 samples/sec
+epoch 12, total_step 186200, total loss is 15.04 , inference loss is 8.52, weight deacy loss is 6.52, training accuracy is 0.437500, time 56.807 samples/sec
+epoch 12, total_step 186220, total loss is 18.04 , inference loss is 11.52, weight deacy loss is 6.52, training accuracy is 0.375000, time 56.934 samples/sec
+epoch 12, total_step 186240, total loss is 13.18 , inference loss is 6.66, weight deacy loss is 6.52, training accuracy is 0.593750, time 57.703 samples/sec
+epoch 12, total_step 186260, total loss is 11.43 , inference loss is 4.91, weight deacy loss is 6.52, training accuracy is 0.593750, time 56.671 samples/sec
+epoch 12, total_step 186280, total loss is 12.04 , inference loss is 5.52, weight deacy loss is 6.52, training accuracy is 0.687500, time 57.772 samples/sec
+epoch 12, total_step 186300, total loss is 10.82 , inference loss is 4.30, weight deacy loss is 6.52, training accuracy is 0.593750, time 55.722 samples/sec
+epoch 12, total_step 186320, total loss is 9.07 , inference loss is 2.56, weight deacy loss is 6.52, training accuracy is 0.812500, time 57.306 samples/sec
+epoch 12, total_step 186340, total loss is 8.29 , inference loss is 1.77, weight deacy loss is 6.52, training accuracy is 0.718750, time 56.759 samples/sec
+epoch 12, total_step 186360, total loss is 12.39 , inference loss is 5.87, weight deacy loss is 6.52, training accuracy is 0.593750, time 56.701 samples/sec
+epoch 12, total_step 186380, total loss is 9.36 , inference loss is 2.84, weight deacy loss is 6.52, training accuracy is 0.687500, time 56.235 samples/sec
+epoch 12, total_step 186400, total loss is 11.47 , inference loss is 4.95, weight deacy loss is 6.52, training accuracy is 0.531250, time 56.830 samples/sec
+epoch 12, total_step 186420, total loss is 15.73 , inference loss is 9.22, weight deacy loss is 6.52, training accuracy is 0.531250, time 56.162 samples/sec
+epoch 12, total_step 186440, total loss is 15.00 , inference loss is 8.49, weight deacy loss is 6.52, training accuracy is 0.437500, time 56.297 samples/sec
+epoch 12, total_step 186460, total loss is 13.26 , inference loss is 6.74, weight deacy loss is 6.52, training accuracy is 0.687500, time 56.276 samples/sec
+epoch 12, total_step 186480, total loss is 11.67 , inference loss is 5.16, weight deacy loss is 6.52, training accuracy is 0.593750, time 57.200 samples/sec
+epoch 12, total_step 186500, total loss is 13.34 , inference loss is 6.83, weight deacy loss is 6.52, training accuracy is 0.562500, time 57.074 samples/sec
+epoch 12, total_step 186520, total loss is 12.75 , inference loss is 6.23, weight deacy loss is 6.52, training accuracy is 0.687500, time 56.606 samples/sec
+epoch 12, total_step 186540, total loss is 10.34 , inference loss is 3.83, weight deacy loss is 6.52, training accuracy is 0.500000, time 56.668 samples/sec
+epoch 12, total_step 186560, total loss is 11.51 , inference loss is 5.00, weight deacy loss is 6.52, training accuracy is 0.531250, time 56.775 samples/sec
+epoch 12, total_step 186580, total loss is 11.55 , inference loss is 5.04, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.891 samples/sec
+epoch 12, total_step 186600, total loss is 10.79 , inference loss is 4.27, weight deacy loss is 6.51, training accuracy is 0.718750, time 55.826 samples/sec
+epoch 12, total_step 186620, total loss is 10.03 , inference loss is 3.52, weight deacy loss is 6.51, training accuracy is 0.781250, time 56.186 samples/sec
+epoch 12, total_step 186640, total loss is 9.74 , inference loss is 3.22, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.699 samples/sec
+epoch 12, total_step 186660, total loss is 13.04 , inference loss is 6.53, weight deacy loss is 6.51, training accuracy is 0.625000, time 56.101 samples/sec
+epoch 12, total_step 186680, total loss is 14.56 , inference loss is 8.04, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.250 samples/sec
+epoch 12, total_step 186700, total loss is 9.15 , inference loss is 2.63, weight deacy loss is 6.51, training accuracy is 0.718750, time 57.069 samples/sec
+epoch 12, total_step 186720, total loss is 11.35 , inference loss is 4.84, weight deacy loss is 6.51, training accuracy is 0.593750, time 55.926 samples/sec
+epoch 12, total_step 186740, total loss is 12.86 , inference loss is 6.35, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.239 samples/sec
+epoch 12, total_step 186760, total loss is 12.85 , inference loss is 6.34, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.348 samples/sec
+epoch 12, total_step 186780, total loss is 9.88 , inference loss is 3.37, weight deacy loss is 6.51, training accuracy is 0.750000, time 56.697 samples/sec
+epoch 12, total_step 186800, total loss is 9.86 , inference loss is 3.34, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.461 samples/sec
+epoch 12, total_step 186820, total loss is 9.44 , inference loss is 2.93, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.067 samples/sec
+epoch 12, total_step 186840, total loss is 10.74 , inference loss is 4.23, weight deacy loss is 6.51, training accuracy is 0.781250, time 56.185 samples/sec
+epoch 12, total_step 186860, total loss is 15.38 , inference loss is 8.87, weight deacy loss is 6.51, training accuracy is 0.531250, time 55.744 samples/sec
+epoch 12, total_step 186880, total loss is 9.08 , inference loss is 2.57, weight deacy loss is 6.51, training accuracy is 0.718750, time 56.513 samples/sec
+epoch 12, total_step 186900, total loss is 8.95 , inference loss is 2.43, weight deacy loss is 6.51, training accuracy is 0.750000, time 56.031 samples/sec
+epoch 12, total_step 186920, total loss is 13.55 , inference loss is 7.04, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.019 samples/sec
+epoch 12, total_step 186940, total loss is 11.68 , inference loss is 5.17, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.950 samples/sec
+epoch 12, total_step 186960, total loss is 11.20 , inference loss is 4.68, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.303 samples/sec
+epoch 12, total_step 186980, total loss is 13.65 , inference loss is 7.14, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.557 samples/sec
+epoch 12, total_step 187000, total loss is 13.04 , inference loss is 6.53, weight deacy loss is 6.51, training accuracy is 0.625000, time 56.685 samples/sec
+epoch 12, total_step 187020, total loss is 12.41 , inference loss is 5.90, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.963 samples/sec
+epoch 12, total_step 187040, total loss is 15.04 , inference loss is 8.53, weight deacy loss is 6.51, training accuracy is 0.406250, time 56.686 samples/sec
+epoch 12, total_step 187060, total loss is 13.01 , inference loss is 6.49, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.796 samples/sec
+epoch 12, total_step 187080, total loss is 10.13 , inference loss is 3.62, weight deacy loss is 6.51, training accuracy is 0.781250, time 55.978 samples/sec
+epoch 12, total_step 187100, total loss is 12.21 , inference loss is 5.70, weight deacy loss is 6.51, training accuracy is 0.625000, time 56.831 samples/sec
+epoch 12, total_step 187120, total loss is 9.96 , inference loss is 3.44, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.862 samples/sec
+epoch 12, total_step 187140, total loss is 11.31 , inference loss is 4.80, weight deacy loss is 6.51, training accuracy is 0.562500, time 55.958 samples/sec
+epoch 12, total_step 187160, total loss is 12.78 , inference loss is 6.27, weight deacy loss is 6.51, training accuracy is 0.718750, time 56.580 samples/sec
+epoch 12, total_step 187180, total loss is 13.46 , inference loss is 6.95, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.696 samples/sec
+epoch 12, total_step 187200, total loss is 15.54 , inference loss is 9.02, weight deacy loss is 6.51, training accuracy is 0.375000, time 56.396 samples/sec
+epoch 12, total_step 187220, total loss is 14.29 , inference loss is 7.78, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.427 samples/sec
+epoch 12, total_step 187240, total loss is 13.80 , inference loss is 7.29, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.705 samples/sec
+epoch 12, total_step 187260, total loss is 14.79 , inference loss is 8.27, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.348 samples/sec
+epoch 12, total_step 187280, total loss is 10.11 , inference loss is 3.60, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.788 samples/sec
+epoch 12, total_step 187300, total loss is 12.41 , inference loss is 5.89, weight deacy loss is 6.51, training accuracy is 0.343750, time 56.578 samples/sec
+epoch 12, total_step 187320, total loss is 14.73 , inference loss is 8.22, weight deacy loss is 6.51, training accuracy is 0.343750, time 56.475 samples/sec
+epoch 12, total_step 187340, total loss is 16.68 , inference loss is 10.17, weight deacy loss is 6.51, training accuracy is 0.406250, time 56.419 samples/sec
+epoch 12, total_step 187360, total loss is 14.73 , inference loss is 8.22, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.689 samples/sec
+epoch 12, total_step 187380, total loss is 14.05 , inference loss is 7.54, weight deacy loss is 6.51, training accuracy is 0.437500, time 56.676 samples/sec
+epoch 12, total_step 187400, total loss is 13.58 , inference loss is 7.07, weight deacy loss is 6.51, training accuracy is 0.437500, time 56.459 samples/sec
+epoch 12, total_step 187420, total loss is 14.95 , inference loss is 8.43, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.907 samples/sec
+epoch 12, total_step 187440, total loss is 10.93 , inference loss is 4.42, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.738 samples/sec
+epoch 12, total_step 187460, total loss is 10.08 , inference loss is 3.57, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.893 samples/sec
+epoch 12, total_step 187480, total loss is 13.77 , inference loss is 7.26, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.446 samples/sec
+epoch 12, total_step 187500, total loss is 13.94 , inference loss is 7.42, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.587 samples/sec
+epoch 12, total_step 187520, total loss is 13.80 , inference loss is 7.29, weight deacy loss is 6.51, training accuracy is 0.562500, time 55.848 samples/sec
+epoch 12, total_step 187540, total loss is 14.37 , inference loss is 7.86, weight deacy loss is 6.51, training accuracy is 0.437500, time 56.879 samples/sec
+epoch 12, total_step 187560, total loss is 14.07 , inference loss is 7.56, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.782 samples/sec
+epoch 12, total_step 187580, total loss is 14.13 , inference loss is 7.62, weight deacy loss is 6.51, training accuracy is 0.437500, time 56.430 samples/sec
+epoch 12, total_step 187600, total loss is 11.03 , inference loss is 4.52, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.147 samples/sec
+epoch 12, total_step 187620, total loss is 18.27 , inference loss is 11.76, weight deacy loss is 6.51, training accuracy is 0.406250, time 55.500 samples/sec
+epoch 12, total_step 187640, total loss is 15.72 , inference loss is 9.21, weight deacy loss is 6.51, training accuracy is 0.406250, time 56.224 samples/sec
+epoch 12, total_step 187660, total loss is 10.98 , inference loss is 4.47, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.936 samples/sec
+epoch 12, total_step 187680, total loss is 17.25 , inference loss is 10.74, weight deacy loss is 6.51, training accuracy is 0.406250, time 56.219 samples/sec
+epoch 12, total_step 187700, total loss is 11.85 , inference loss is 5.34, weight deacy loss is 6.51, training accuracy is 0.406250, time 55.575 samples/sec
+epoch 12, total_step 187720, total loss is 16.19 , inference loss is 9.68, weight deacy loss is 6.51, training accuracy is 0.437500, time 57.234 samples/sec
+epoch 12, total_step 187740, total loss is 11.06 , inference loss is 4.55, weight deacy loss is 6.51, training accuracy is 0.687500, time 57.066 samples/sec
+epoch 12, total_step 187760, total loss is 16.39 , inference loss is 9.88, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.466 samples/sec
+epoch 12, total_step 187780, total loss is 18.78 , inference loss is 12.27, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.917 samples/sec
+epoch 12, total_step 187800, total loss is 12.07 , inference loss is 5.56, weight deacy loss is 6.51, training accuracy is 0.562500, time 55.086 samples/sec
+epoch 12, total_step 187820, total loss is 14.54 , inference loss is 8.03, weight deacy loss is 6.51, training accuracy is 0.531250, time 57.321 samples/sec
+epoch 12, total_step 187840, total loss is 11.98 , inference loss is 5.47, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.972 samples/sec
+epoch 12, total_step 187860, total loss is 12.44 , inference loss is 5.93, weight deacy loss is 6.51, training accuracy is 0.656250, time 55.552 samples/sec
+epoch 12, total_step 187880, total loss is 12.55 , inference loss is 6.04, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.557 samples/sec
+epoch 12, total_step 187900, total loss is 13.47 , inference loss is 6.96, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.708 samples/sec
+epoch 12, total_step 187920, total loss is 12.93 , inference loss is 6.42, weight deacy loss is 6.51, training accuracy is 0.687500, time 56.032 samples/sec
+epoch 12, total_step 187940, total loss is 15.97 , inference loss is 9.46, weight deacy loss is 6.51, training accuracy is 0.343750, time 55.617 samples/sec
+epoch 12, total_step 187960, total loss is 19.88 , inference loss is 13.37, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.108 samples/sec
+epoch 12, total_step 187980, total loss is 12.63 , inference loss is 6.12, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.601 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.2055060000001
+best_threshold_index 121 0.9877777777777778
+best_threshold_index 121 0.9868518518518519
+best_threshold_index 121 0.9872222222222222
+best_threshold_index 121 0.9881481481481481
+best_threshold_index 121 0.9875925925925926
+best_threshold_index 121 0.9872222222222222
+best_threshold_index 121 0.9879629629629629
+best_threshold_index 121 0.9874074074074074
+best_threshold_index 121 0.9862962962962963
+best_threshold_index 121 0.9868518518518519
+[lfw][188000]XNorm: 21.544435
+[lfw][188000]Accuracy-Flip: 0.98733+-0.00484
+testing verification..
+(14000, 512)
+infer time 85.5788470000001
+best_threshold_index 125 0.9876190476190476
+best_threshold_index 126 0.9857142857142858
+best_threshold_index 127 0.986984126984127
+best_threshold_index 126 0.9857142857142858
+best_threshold_index 126 0.9874603174603175
+best_threshold_index 126 0.9857142857142858
+best_threshold_index 126 0.986031746031746
+best_threshold_index 126 0.9858730158730159
+best_threshold_index 126 0.9852380952380952
+best_threshold_index 128 0.986984126984127
+[cfp_ff][188000]XNorm: 20.945405
+[cfp_ff][188000]Accuracy-Flip: 0.98529+-0.00793
+testing verification..
+(14000, 512)
+infer time 85.694674
+best_threshold_index 142 0.9238095238095239
+best_threshold_index 142 0.9241269841269841
+best_threshold_index 142 0.9228571428571428
+best_threshold_index 142 0.9252380952380952
+best_threshold_index 142 0.9258730158730158
+best_threshold_index 142 0.923968253968254
+best_threshold_index 142 0.9226984126984127
+best_threshold_index 150 0.9230158730158731
+best_threshold_index 142 0.9231746031746032
+best_threshold_index 142 0.9241269841269841
+[cfp_fp][188000]XNorm: 20.837037
+[cfp_fp][188000]Accuracy-Flip: 0.92214+-0.00860
+testing verification..
+(12000, 512)
+infer time 73.60295900000003
+best_threshold_index 142 0.899074074074074
+best_threshold_index 142 0.8972222222222223
+best_threshold_index 142 0.8988888888888888
+best_threshold_index 142 0.8992592592592593
+best_threshold_index 142 0.9027777777777778
+best_threshold_index 142 0.897962962962963
+best_threshold_index 142 0.8996296296296297
+best_threshold_index 142 0.9046296296296297
+best_threshold_index 142 0.9018518518518519
+best_threshold_index 142 0.9003703703703704
+[agedb_30][188000]XNorm: 22.342971
+[agedb_30][188000]Accuracy-Flip: 0.90017+-0.01953
+test accuracy is: 0.9873333333333333
+epoch 12, total_step 188000, total loss is 12.52 , inference loss is 6.01, weight deacy loss is 6.51, training accuracy is 0.406250, time 54.604 samples/sec
+epoch 12, total_step 188020, total loss is 12.63 , inference loss is 6.12, weight deacy loss is 6.51, training accuracy is 0.437500, time 55.570 samples/sec
+epoch 12, total_step 188040, total loss is 16.89 , inference loss is 10.38, weight deacy loss is 6.51, training accuracy is 0.375000, time 56.569 samples/sec
+epoch 12, total_step 188060, total loss is 13.71 , inference loss is 7.20, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.655 samples/sec
+epoch 12, total_step 188080, total loss is 13.14 , inference loss is 6.64, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.822 samples/sec
+epoch 12, total_step 188100, total loss is 13.92 , inference loss is 7.41, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.663 samples/sec
+epoch 12, total_step 188120, total loss is 10.17 , inference loss is 3.66, weight deacy loss is 6.51, training accuracy is 0.625000, time 56.529 samples/sec
+epoch 12, total_step 188140, total loss is 16.20 , inference loss is 9.69, weight deacy loss is 6.51, training accuracy is 0.468750, time 55.787 samples/sec
+epoch 12, total_step 188160, total loss is 14.65 , inference loss is 8.15, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.505 samples/sec
+epoch 12, total_step 188180, total loss is 11.22 , inference loss is 4.71, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.168 samples/sec
+epoch 12, total_step 188200, total loss is 16.73 , inference loss is 10.22, weight deacy loss is 6.51, training accuracy is 0.437500, time 56.411 samples/sec
+epoch 12, total_step 188220, total loss is 12.89 , inference loss is 6.39, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.583 samples/sec
+epoch 12, total_step 188240, total loss is 17.40 , inference loss is 10.89, weight deacy loss is 6.51, training accuracy is 0.281250, time 56.898 samples/sec
+epoch 12, total_step 188260, total loss is 15.42 , inference loss is 8.91, weight deacy loss is 6.51, training accuracy is 0.500000, time 55.639 samples/sec
+epoch 12, total_step 188280, total loss is 15.14 , inference loss is 8.63, weight deacy loss is 6.51, training accuracy is 0.562500, time 56.556 samples/sec
+epoch 12, total_step 188300, total loss is 12.53 , inference loss is 6.02, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.346 samples/sec
+epoch 12, total_step 188320, total loss is 12.53 , inference loss is 6.02, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.595 samples/sec
+epoch 12, total_step 188340, total loss is 12.40 , inference loss is 5.90, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.203 samples/sec
+epoch 12, total_step 188360, total loss is 9.92 , inference loss is 3.41, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.537 samples/sec
+epoch 12, total_step 188380, total loss is 14.37 , inference loss is 7.86, weight deacy loss is 6.51, training accuracy is 0.593750, time 57.698 samples/sec
+epoch 12, total_step 188400, total loss is 17.29 , inference loss is 10.78, weight deacy loss is 6.51, training accuracy is 0.437500, time 56.772 samples/sec
+epoch 12, total_step 188420, total loss is 18.24 , inference loss is 11.73, weight deacy loss is 6.51, training accuracy is 0.375000, time 55.690 samples/sec
+epoch 12, total_step 188440, total loss is 11.85 , inference loss is 5.34, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.151 samples/sec
+epoch 12, total_step 188460, total loss is 11.66 , inference loss is 5.16, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.774 samples/sec
+epoch 12, total_step 188480, total loss is 14.68 , inference loss is 8.17, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.335 samples/sec
+epoch 12, total_step 188500, total loss is 12.71 , inference loss is 6.21, weight deacy loss is 6.51, training accuracy is 0.500000, time 54.156 samples/sec
+epoch 12, total_step 188520, total loss is 18.69 , inference loss is 12.18, weight deacy loss is 6.51, training accuracy is 0.468750, time 57.082 samples/sec
+epoch 12, total_step 188540, total loss is 12.88 , inference loss is 6.37, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.969 samples/sec
+epoch 12, total_step 188560, total loss is 12.48 , inference loss is 5.97, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.922 samples/sec
+epoch 12, total_step 188580, total loss is 15.13 , inference loss is 8.63, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.704 samples/sec
+epoch 12, total_step 188600, total loss is 15.87 , inference loss is 9.37, weight deacy loss is 6.51, training accuracy is 0.437500, time 55.988 samples/sec
+epoch 12, total_step 188620, total loss is 17.01 , inference loss is 10.51, weight deacy loss is 6.51, training accuracy is 0.562500, time 55.683 samples/sec
+epoch 12, total_step 188640, total loss is 14.86 , inference loss is 8.35, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.108 samples/sec
+epoch 12, total_step 188660, total loss is 16.84 , inference loss is 10.33, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.266 samples/sec
+epoch 12, total_step 188680, total loss is 13.29 , inference loss is 6.79, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.142 samples/sec
+epoch 12, total_step 188700, total loss is 13.30 , inference loss is 6.79, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.310 samples/sec
+epoch 12, total_step 188720, total loss is 14.07 , inference loss is 7.57, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.621 samples/sec
+epoch 12, total_step 188740, total loss is 12.28 , inference loss is 5.78, weight deacy loss is 6.51, training accuracy is 0.531250, time 56.019 samples/sec
+epoch 12, total_step 188760, total loss is 12.19 , inference loss is 5.69, weight deacy loss is 6.51, training accuracy is 0.593750, time 57.004 samples/sec
+epoch 12, total_step 188780, total loss is 11.18 , inference loss is 4.67, weight deacy loss is 6.51, training accuracy is 0.562500, time 57.126 samples/sec
+epoch 12, total_step 188800, total loss is 17.80 , inference loss is 11.30, weight deacy loss is 6.51, training accuracy is 0.218750, time 56.500 samples/sec
+epoch 12, total_step 188820, total loss is 10.66 , inference loss is 4.15, weight deacy loss is 6.51, training accuracy is 0.562500, time 55.794 samples/sec
+epoch 12, total_step 188840, total loss is 15.18 , inference loss is 8.68, weight deacy loss is 6.51, training accuracy is 0.468750, time 56.569 samples/sec
+epoch 12, total_step 188860, total loss is 14.86 , inference loss is 8.36, weight deacy loss is 6.51, training accuracy is 0.406250, time 56.290 samples/sec
+epoch 12, total_step 188880, total loss is 12.15 , inference loss is 5.65, weight deacy loss is 6.51, training accuracy is 0.593750, time 56.212 samples/sec
+epoch 12, total_step 188900, total loss is 9.62 , inference loss is 3.11, weight deacy loss is 6.51, training accuracy is 0.500000, time 56.663 samples/sec
+epoch 12, total_step 188920, total loss is 15.93 , inference loss is 9.42, weight deacy loss is 6.50, training accuracy is 0.562500, time 56.149 samples/sec
+epoch 12, total_step 188940, total loss is 14.54 , inference loss is 8.03, weight deacy loss is 6.50, training accuracy is 0.593750, time 56.353 samples/sec
+epoch 12, total_step 188960, total loss is 13.58 , inference loss is 7.07, weight deacy loss is 6.50, training accuracy is 0.531250, time 56.443 samples/sec
+epoch 12, total_step 188980, total loss is 15.37 , inference loss is 8.86, weight deacy loss is 6.50, training accuracy is 0.531250, time 57.099 samples/sec
+epoch 12, total_step 189000, total loss is 13.97 , inference loss is 7.46, weight deacy loss is 6.50, training accuracy is 0.437500, time 56.728 samples/sec
+epoch 12, total_step 189020, total loss is 14.08 , inference loss is 7.58, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.102 samples/sec
+epoch 12, total_step 189040, total loss is 14.74 , inference loss is 8.23, weight deacy loss is 6.50, training accuracy is 0.531250, time 56.311 samples/sec
+epoch 12, total_step 189060, total loss is 12.48 , inference loss is 5.98, weight deacy loss is 6.50, training accuracy is 0.531250, time 56.425 samples/sec
+epoch 12, total_step 189080, total loss is 12.26 , inference loss is 5.76, weight deacy loss is 6.50, training accuracy is 0.562500, time 57.067 samples/sec
+epoch 12, total_step 189100, total loss is 20.12 , inference loss is 13.61, weight deacy loss is 6.50, training accuracy is 0.468750, time 56.505 samples/sec
+epoch 12, total_step 189120, total loss is 14.75 , inference loss is 8.25, weight deacy loss is 6.50, training accuracy is 0.343750, time 56.805 samples/sec
+epoch 12, total_step 189140, total loss is 12.56 , inference loss is 6.06, weight deacy loss is 6.50, training accuracy is 0.500000, time 56.657 samples/sec
+epoch 12, total_step 189160, total loss is 11.91 , inference loss is 5.41, weight deacy loss is 6.50, training accuracy is 0.531250, time 57.384 samples/sec
+epoch 12, total_step 189180, total loss is 14.92 , inference loss is 8.41, weight deacy loss is 6.50, training accuracy is 0.375000, time 57.374 samples/sec
+epoch 12, total_step 189200, total loss is 12.62 , inference loss is 6.12, weight deacy loss is 6.50, training accuracy is 0.468750, time 56.700 samples/sec
+epoch 12, total_step 189220, total loss is 12.66 , inference loss is 6.15, weight deacy loss is 6.50, training accuracy is 0.562500, time 56.577 samples/sec
+epoch 12, total_step 189240, total loss is 12.39 , inference loss is 5.89, weight deacy loss is 6.50, training accuracy is 0.562500, time 56.205 samples/sec
+epoch 12, total_step 189260, total loss is 18.43 , inference loss is 11.93, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.684 samples/sec
+epoch 12, total_step 189280, total loss is 15.54 , inference loss is 9.03, weight deacy loss is 6.50, training accuracy is 0.468750, time 55.312 samples/sec
+epoch 12, total_step 189300, total loss is 12.00 , inference loss is 5.49, weight deacy loss is 6.50, training accuracy is 0.531250, time 55.644 samples/sec
+epoch 12, total_step 189320, total loss is 13.29 , inference loss is 6.79, weight deacy loss is 6.50, training accuracy is 0.500000, time 55.908 samples/sec
+epoch 12, total_step 189340, total loss is 14.89 , inference loss is 8.39, weight deacy loss is 6.50, training accuracy is 0.375000, time 57.157 samples/sec
+epoch 12, total_step 189360, total loss is 14.57 , inference loss is 8.06, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.810 samples/sec
+epoch 12, total_step 189380, total loss is 13.05 , inference loss is 6.55, weight deacy loss is 6.50, training accuracy is 0.406250, time 55.794 samples/sec
+epoch 12, total_step 189400, total loss is 13.04 , inference loss is 6.54, weight deacy loss is 6.50, training accuracy is 0.500000, time 57.477 samples/sec
+epoch 12, total_step 189420, total loss is 8.80 , inference loss is 2.30, weight deacy loss is 6.50, training accuracy is 0.718750, time 56.489 samples/sec
+epoch 12, total_step 189440, total loss is 11.83 , inference loss is 5.32, weight deacy loss is 6.50, training accuracy is 0.531250, time 56.570 samples/sec
+epoch 12, total_step 189460, total loss is 17.73 , inference loss is 11.23, weight deacy loss is 6.50, training accuracy is 0.343750, time 56.849 samples/sec
+epoch 12, total_step 189480, total loss is 12.25 , inference loss is 5.75, weight deacy loss is 6.50, training accuracy is 0.500000, time 56.543 samples/sec
+epoch 12, total_step 189500, total loss is 15.53 , inference loss is 9.03, weight deacy loss is 6.50, training accuracy is 0.312500, time 55.883 samples/sec
+epoch 12, total_step 189520, total loss is 16.44 , inference loss is 9.94, weight deacy loss is 6.50, training accuracy is 0.312500, time 56.328 samples/sec
+epoch 12, total_step 189540, total loss is 15.25 , inference loss is 8.75, weight deacy loss is 6.50, training accuracy is 0.593750, time 56.471 samples/sec
+epoch 12, total_step 189560, total loss is 16.55 , inference loss is 10.05, weight deacy loss is 6.50, training accuracy is 0.468750, time 57.529 samples/sec
+epoch 12, total_step 189580, total loss is 13.37 , inference loss is 6.87, weight deacy loss is 6.50, training accuracy is 0.625000, time 55.907 samples/sec
+epoch 12, total_step 189600, total loss is 17.23 , inference loss is 10.73, weight deacy loss is 6.50, training accuracy is 0.531250, time 56.888 samples/sec
+epoch 12, total_step 189620, total loss is 11.69 , inference loss is 5.19, weight deacy loss is 6.50, training accuracy is 0.531250, time 55.514 samples/sec
+epoch 12, total_step 189640, total loss is 13.63 , inference loss is 7.13, weight deacy loss is 6.50, training accuracy is 0.468750, time 56.794 samples/sec
+epoch 12, total_step 189660, total loss is 13.80 , inference loss is 7.30, weight deacy loss is 6.50, training accuracy is 0.437500, time 56.597 samples/sec
+epoch 12, total_step 189680, total loss is 11.60 , inference loss is 5.10, weight deacy loss is 6.50, training accuracy is 0.468750, time 56.006 samples/sec
+epoch 12, total_step 189700, total loss is 15.72 , inference loss is 9.21, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.025 samples/sec
+epoch 12, total_step 189720, total loss is 13.62 , inference loss is 7.12, weight deacy loss is 6.50, training accuracy is 0.468750, time 57.004 samples/sec
+epoch 12, total_step 189740, total loss is 14.10 , inference loss is 7.60, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.157 samples/sec
+epoch 12, total_step 189760, total loss is 18.27 , inference loss is 11.77, weight deacy loss is 6.50, training accuracy is 0.312500, time 56.431 samples/sec
+epoch 12, total_step 189780, total loss is 12.55 , inference loss is 6.05, weight deacy loss is 6.50, training accuracy is 0.562500, time 56.614 samples/sec
+epoch 12, total_step 189800, total loss is 15.51 , inference loss is 9.01, weight deacy loss is 6.50, training accuracy is 0.593750, time 56.639 samples/sec
+epoch 12, total_step 189820, total loss is 14.14 , inference loss is 7.64, weight deacy loss is 6.50, training accuracy is 0.437500, time 56.234 samples/sec
+epoch 12, total_step 189840, total loss is 14.53 , inference loss is 8.03, weight deacy loss is 6.50, training accuracy is 0.437500, time 56.424 samples/sec
+epoch 12, total_step 189860, total loss is 16.82 , inference loss is 10.32, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.068 samples/sec
+epoch 12, total_step 189880, total loss is 13.02 , inference loss is 6.52, weight deacy loss is 6.50, training accuracy is 0.343750, time 57.518 samples/sec
+epoch 12, total_step 189900, total loss is 12.65 , inference loss is 6.16, weight deacy loss is 6.50, training accuracy is 0.468750, time 56.741 samples/sec
+epoch 12, total_step 189920, total loss is 15.76 , inference loss is 9.26, weight deacy loss is 6.50, training accuracy is 0.343750, time 56.015 samples/sec
+epoch 12, total_step 189940, total loss is 17.41 , inference loss is 10.91, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.942 samples/sec
+epoch 12, total_step 189960, total loss is 9.94 , inference loss is 3.44, weight deacy loss is 6.50, training accuracy is 0.656250, time 57.123 samples/sec
+epoch 12, total_step 189980, total loss is 15.93 , inference loss is 9.43, weight deacy loss is 6.50, training accuracy is 0.343750, time 56.571 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.22790799999999
+best_threshold_index 118 0.9890740740740741
+best_threshold_index 118 0.9883333333333333
+best_threshold_index 118 0.9890740740740741
+best_threshold_index 117 0.9885185185185185
+best_threshold_index 118 0.9885185185185185
+best_threshold_index 118 0.9877777777777778
+best_threshold_index 118 0.9887037037037038
+best_threshold_index 118 0.9881481481481481
+best_threshold_index 118 0.9874074074074074
+best_threshold_index 118 0.9877777777777778
+[lfw][190000]XNorm: 22.062870
+[lfw][190000]Accuracy-Flip: 0.98800+-0.00499
+testing verification..
+(14000, 512)
+infer time 85.550921
+best_threshold_index 121 0.9863492063492063
+best_threshold_index 121 0.9866666666666667
+best_threshold_index 123 0.9865079365079366
+best_threshold_index 121 0.986031746031746
+best_threshold_index 120 0.9873015873015873
+best_threshold_index 120 0.9858730158730159
+best_threshold_index 121 0.9868253968253968
+best_threshold_index 121 0.9861904761904762
+best_threshold_index 121 0.9858730158730159
+best_threshold_index 121 0.986984126984127
+[cfp_ff][190000]XNorm: 21.306549
+[cfp_ff][190000]Accuracy-Flip: 0.98571+-0.00429
+testing verification..
+(14000, 512)
+infer time 85.64785099999989
+best_threshold_index 139 0.9238095238095239
+best_threshold_index 139 0.9249206349206349
+best_threshold_index 139 0.9226984126984127
+best_threshold_index 139 0.9250793650793651
+best_threshold_index 138 0.9273015873015873
+best_threshold_index 139 0.9223809523809524
+best_threshold_index 139 0.9222222222222223
+best_threshold_index 138 0.9247619047619048
+best_threshold_index 139 0.9236507936507936
+best_threshold_index 139 0.9247619047619048
+[cfp_fp][190000]XNorm: 22.130150
+[cfp_fp][190000]Accuracy-Flip: 0.92343+-0.01394
+testing verification..
+(12000, 512)
+infer time 72.91035699999995
+best_threshold_index 138 0.912962962962963
+best_threshold_index 139 0.9118518518518518
+best_threshold_index 138 0.9135185185185185
+best_threshold_index 139 0.915
+best_threshold_index 138 0.9174074074074074
+best_threshold_index 139 0.9124074074074074
+best_threshold_index 139 0.9131481481481482
+best_threshold_index 141 0.9183333333333333
+best_threshold_index 141 0.9151851851851852
+best_threshold_index 139 0.9153703703703704
+[agedb_30][190000]XNorm: 22.716146
+[agedb_30][190000]Accuracy-Flip: 0.91150+-0.01905
+test accuracy is: 0.9880000000000001
+epoch 12, total_step 190000, total loss is 10.08 , inference loss is 3.58, weight deacy loss is 6.50, training accuracy is 0.562500, time 55.200 samples/sec
+epoch 12, total_step 190020, total loss is 15.75 , inference loss is 9.25, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.633 samples/sec
+epoch 12, total_step 190040, total loss is 13.74 , inference loss is 7.24, weight deacy loss is 6.50, training accuracy is 0.437500, time 56.579 samples/sec
+epoch 12, total_step 190060, total loss is 13.97 , inference loss is 7.47, weight deacy loss is 6.50, training accuracy is 0.468750, time 56.868 samples/sec
+epoch 12, total_step 190080, total loss is 12.13 , inference loss is 5.63, weight deacy loss is 6.50, training accuracy is 0.375000, time 56.435 samples/sec
+epoch 12, total_step 190100, total loss is 12.31 , inference loss is 5.82, weight deacy loss is 6.50, training accuracy is 0.500000, time 56.742 samples/sec
+epoch 12, total_step 190120, total loss is 13.69 , inference loss is 7.20, weight deacy loss is 6.50, training accuracy is 0.562500, time 56.694 samples/sec
+epoch 12, total_step 190140, total loss is 13.46 , inference loss is 6.96, weight deacy loss is 6.50, training accuracy is 0.375000, time 56.876 samples/sec
+epoch 12, total_step 190160, total loss is 11.87 , inference loss is 5.37, weight deacy loss is 6.50, training accuracy is 0.375000, time 56.881 samples/sec
+epoch 12, total_step 190180, total loss is 13.73 , inference loss is 7.24, weight deacy loss is 6.50, training accuracy is 0.531250, time 56.525 samples/sec
+epoch 12, total_step 190200, total loss is 12.68 , inference loss is 6.18, weight deacy loss is 6.50, training accuracy is 0.468750, time 55.646 samples/sec
+epoch 12, total_step 190220, total loss is 15.61 , inference loss is 9.12, weight deacy loss is 6.50, training accuracy is 0.468750, time 56.948 samples/sec
+epoch 12, total_step 190240, total loss is 14.69 , inference loss is 8.19, weight deacy loss is 6.50, training accuracy is 0.687500, time 56.372 samples/sec
+epoch 12, total_step 190260, total loss is 13.63 , inference loss is 7.14, weight deacy loss is 6.50, training accuracy is 0.500000, time 57.008 samples/sec
+epoch 12, total_step 190280, total loss is 18.52 , inference loss is 12.02, weight deacy loss is 6.50, training accuracy is 0.437500, time 56.355 samples/sec
+epoch 12, total_step 190300, total loss is 14.59 , inference loss is 8.10, weight deacy loss is 6.50, training accuracy is 0.500000, time 56.703 samples/sec
+epoch 12, total_step 190320, total loss is 12.38 , inference loss is 5.88, weight deacy loss is 6.50, training accuracy is 0.562500, time 56.266 samples/sec
+epoch 12, total_step 190340, total loss is 12.50 , inference loss is 6.01, weight deacy loss is 6.50, training accuracy is 0.625000, time 56.500 samples/sec
+epoch 12, total_step 190360, total loss is 14.06 , inference loss is 7.57, weight deacy loss is 6.50, training accuracy is 0.437500, time 55.978 samples/sec
+epoch 12, total_step 190380, total loss is 13.24 , inference loss is 6.74, weight deacy loss is 6.50, training accuracy is 0.406250, time 56.944 samples/sec
+epoch 12, total_step 190400, total loss is 12.80 , inference loss is 6.31, weight deacy loss is 6.50, training accuracy is 0.468750, time 57.424 samples/sec
+epoch 12, total_step 190420, total loss is 14.56 , inference loss is 8.07, weight deacy loss is 6.50, training accuracy is 0.468750, time 55.579 samples/sec
+epoch 12, total_step 190440, total loss is 13.34 , inference loss is 6.84, weight deacy loss is 6.50, training accuracy is 0.500000, time 57.044 samples/sec
+epoch 12, total_step 190460, total loss is 17.69 , inference loss is 11.20, weight deacy loss is 6.50, training accuracy is 0.500000, time 56.624 samples/sec
+epoch 12, total_step 190480, total loss is 12.36 , inference loss is 5.87, weight deacy loss is 6.50, training accuracy is 0.593750, time 56.126 samples/sec
+epoch 12, total_step 190500, total loss is 13.91 , inference loss is 7.41, weight deacy loss is 6.50, training accuracy is 0.500000, time 56.587 samples/sec
+epoch 12, total_step 190520, total loss is 15.46 , inference loss is 8.97, weight deacy loss is 6.49, training accuracy is 0.437500, time 57.156 samples/sec
+epoch 12, total_step 190540, total loss is 13.63 , inference loss is 7.14, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.658 samples/sec
+epoch 12, total_step 190560, total loss is 12.46 , inference loss is 5.97, weight deacy loss is 6.49, training accuracy is 0.593750, time 55.591 samples/sec
+epoch 12, total_step 190580, total loss is 12.65 , inference loss is 6.16, weight deacy loss is 6.49, training accuracy is 0.312500, time 56.695 samples/sec
+epoch 12, total_step 190600, total loss is 12.18 , inference loss is 5.69, weight deacy loss is 6.49, training accuracy is 0.375000, time 56.388 samples/sec
+epoch 12, total_step 190620, total loss is 13.61 , inference loss is 7.11, weight deacy loss is 6.49, training accuracy is 0.343750, time 56.325 samples/sec
+epoch 12, total_step 190640, total loss is 13.56 , inference loss is 7.06, weight deacy loss is 6.49, training accuracy is 0.625000, time 56.002 samples/sec
+epoch 12, total_step 190660, total loss is 14.06 , inference loss is 7.56, weight deacy loss is 6.49, training accuracy is 0.437500, time 57.133 samples/sec
+epoch 12, total_step 190680, total loss is 12.98 , inference loss is 6.48, weight deacy loss is 6.49, training accuracy is 0.531250, time 56.516 samples/sec
+epoch 12, total_step 190700, total loss is 15.83 , inference loss is 9.34, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.441 samples/sec
+epoch 12, total_step 190720, total loss is 11.99 , inference loss is 5.49, weight deacy loss is 6.49, training accuracy is 0.375000, time 56.634 samples/sec
+epoch 12, total_step 190740, total loss is 11.73 , inference loss is 5.24, weight deacy loss is 6.49, training accuracy is 0.593750, time 57.061 samples/sec
+epoch 12, total_step 190760, total loss is 13.91 , inference loss is 7.42, weight deacy loss is 6.49, training accuracy is 0.562500, time 56.631 samples/sec
+epoch 12, total_step 190780, total loss is 16.39 , inference loss is 9.90, weight deacy loss is 6.49, training accuracy is 0.406250, time 55.823 samples/sec
+epoch 12, total_step 190800, total loss is 14.53 , inference loss is 8.03, weight deacy loss is 6.49, training accuracy is 0.375000, time 56.674 samples/sec
+epoch 12, total_step 190820, total loss is 11.87 , inference loss is 5.38, weight deacy loss is 6.49, training accuracy is 0.500000, time 55.647 samples/sec
+epoch 12, total_step 190840, total loss is 15.77 , inference loss is 9.28, weight deacy loss is 6.49, training accuracy is 0.406250, time 56.943 samples/sec
+epoch 12, total_step 190860, total loss is 12.22 , inference loss is 5.73, weight deacy loss is 6.49, training accuracy is 0.562500, time 55.809 samples/sec
+epoch 12, total_step 190880, total loss is 16.25 , inference loss is 9.76, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.108 samples/sec
+epoch 12, total_step 190900, total loss is 13.01 , inference loss is 6.52, weight deacy loss is 6.49, training accuracy is 0.500000, time 56.739 samples/sec
+epoch 12, total_step 190920, total loss is 12.27 , inference loss is 5.77, weight deacy loss is 6.49, training accuracy is 0.531250, time 56.588 samples/sec
+epoch 12, total_step 190940, total loss is 13.31 , inference loss is 6.82, weight deacy loss is 6.49, training accuracy is 0.437500, time 56.111 samples/sec
+epoch 12, total_step 190960, total loss is 9.91 , inference loss is 3.42, weight deacy loss is 6.49, training accuracy is 0.562500, time 57.216 samples/sec
+epoch 12, total_step 190980, total loss is 12.12 , inference loss is 5.63, weight deacy loss is 6.49, training accuracy is 0.562500, time 56.109 samples/sec
+epoch 12, total_step 191000, total loss is 18.56 , inference loss is 12.07, weight deacy loss is 6.49, training accuracy is 0.406250, time 56.258 samples/sec
+epoch 12, total_step 191020, total loss is 10.37 , inference loss is 3.88, weight deacy loss is 6.49, training accuracy is 0.562500, time 56.565 samples/sec
+epoch 12, total_step 191040, total loss is 12.07 , inference loss is 5.58, weight deacy loss is 6.49, training accuracy is 0.500000, time 56.140 samples/sec
+epoch 12, total_step 191060, total loss is 13.96 , inference loss is 7.46, weight deacy loss is 6.49, training accuracy is 0.437500, time 56.839 samples/sec
+epoch 12, total_step 191080, total loss is 18.03 , inference loss is 11.54, weight deacy loss is 6.49, training accuracy is 0.375000, time 56.414 samples/sec
+epoch 12, total_step 191100, total loss is 13.02 , inference loss is 6.53, weight deacy loss is 6.49, training accuracy is 0.562500, time 56.528 samples/sec
+epoch 12, total_step 191120, total loss is 13.82 , inference loss is 7.33, weight deacy loss is 6.49, training accuracy is 0.500000, time 56.241 samples/sec
+epoch 12, total_step 191140, total loss is 10.98 , inference loss is 4.49, weight deacy loss is 6.49, training accuracy is 0.437500, time 55.041 samples/sec
+epoch 12, total_step 191160, total loss is 15.41 , inference loss is 8.92, weight deacy loss is 6.49, training accuracy is 0.562500, time 56.716 samples/sec
+epoch 12, total_step 191180, total loss is 14.29 , inference loss is 7.80, weight deacy loss is 6.49, training accuracy is 0.531250, time 55.936 samples/sec
+epoch 12, total_step 191200, total loss is 12.40 , inference loss is 5.91, weight deacy loss is 6.49, training accuracy is 0.531250, time 57.231 samples/sec
+epoch 12, total_step 191220, total loss is 12.77 , inference loss is 6.28, weight deacy loss is 6.49, training accuracy is 0.562500, time 56.684 samples/sec
+epoch 12, total_step 191240, total loss is 13.27 , inference loss is 6.78, weight deacy loss is 6.49, training accuracy is 0.656250, time 56.739 samples/sec
+epoch 12, total_step 191260, total loss is 13.78 , inference loss is 7.29, weight deacy loss is 6.49, training accuracy is 0.593750, time 57.085 samples/sec
+epoch 12, total_step 191280, total loss is 13.08 , inference loss is 6.59, weight deacy loss is 6.49, training accuracy is 0.468750, time 57.137 samples/sec
+epoch 12, total_step 191300, total loss is 13.97 , inference loss is 7.48, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.603 samples/sec
+epoch 12, total_step 191320, total loss is 14.87 , inference loss is 8.38, weight deacy loss is 6.49, training accuracy is 0.531250, time 57.417 samples/sec
+epoch 12, total_step 191340, total loss is 15.59 , inference loss is 9.10, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.668 samples/sec
+epoch 12, total_step 191360, total loss is 9.78 , inference loss is 3.28, weight deacy loss is 6.49, training accuracy is 0.593750, time 56.533 samples/sec
+epoch 12, total_step 191380, total loss is 13.62 , inference loss is 7.13, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.730 samples/sec
+epoch 12, total_step 191400, total loss is 11.81 , inference loss is 5.32, weight deacy loss is 6.49, training accuracy is 0.531250, time 56.472 samples/sec
+epoch 12, total_step 191420, total loss is 13.75 , inference loss is 7.26, weight deacy loss is 6.49, training accuracy is 0.375000, time 57.449 samples/sec
+epoch 12, total_step 191440, total loss is 17.35 , inference loss is 10.86, weight deacy loss is 6.49, training accuracy is 0.437500, time 56.097 samples/sec
+epoch 12, total_step 191460, total loss is 17.62 , inference loss is 11.13, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.400 samples/sec
+epoch 12, total_step 191480, total loss is 10.45 , inference loss is 3.96, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.775 samples/sec
+epoch 12, total_step 191500, total loss is 14.58 , inference loss is 8.09, weight deacy loss is 6.49, training accuracy is 0.375000, time 55.439 samples/sec
+epoch 12, total_step 191520, total loss is 11.62 , inference loss is 5.14, weight deacy loss is 6.49, training accuracy is 0.593750, time 56.147 samples/sec
+epoch 12, total_step 191540, total loss is 12.54 , inference loss is 6.05, weight deacy loss is 6.49, training accuracy is 0.500000, time 56.723 samples/sec
+epoch 12, total_step 191560, total loss is 15.05 , inference loss is 8.56, weight deacy loss is 6.49, training accuracy is 0.562500, time 56.794 samples/sec
+epoch 12, total_step 191580, total loss is 10.95 , inference loss is 4.46, weight deacy loss is 6.49, training accuracy is 0.500000, time 56.792 samples/sec
+epoch 12, total_step 191600, total loss is 10.80 , inference loss is 4.31, weight deacy loss is 6.49, training accuracy is 0.593750, time 56.437 samples/sec
+epoch 12, total_step 191620, total loss is 11.78 , inference loss is 5.29, weight deacy loss is 6.49, training accuracy is 0.437500, time 56.280 samples/sec
+epoch 12, total_step 191640, total loss is 13.85 , inference loss is 7.36, weight deacy loss is 6.49, training accuracy is 0.375000, time 57.280 samples/sec
+epoch 12, total_step 191660, total loss is 10.48 , inference loss is 3.99, weight deacy loss is 6.49, training accuracy is 0.593750, time 56.642 samples/sec
+epoch 12, total_step 191680, total loss is 13.07 , inference loss is 6.58, weight deacy loss is 6.49, training accuracy is 0.531250, time 56.364 samples/sec
+epoch 12, total_step 191700, total loss is 12.69 , inference loss is 6.20, weight deacy loss is 6.49, training accuracy is 0.437500, time 56.515 samples/sec
+epoch 12, total_step 191720, total loss is 13.25 , inference loss is 6.77, weight deacy loss is 6.49, training accuracy is 0.625000, time 56.927 samples/sec
+epoch 12, total_step 191740, total loss is 13.98 , inference loss is 7.49, weight deacy loss is 6.49, training accuracy is 0.531250, time 56.075 samples/sec
+epoch 12, total_step 191760, total loss is 10.90 , inference loss is 4.41, weight deacy loss is 6.49, training accuracy is 0.656250, time 57.040 samples/sec
+epoch 12, total_step 191780, total loss is 16.05 , inference loss is 9.56, weight deacy loss is 6.49, training accuracy is 0.531250, time 56.540 samples/sec
+epoch 12, total_step 191800, total loss is 15.91 , inference loss is 9.43, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.470 samples/sec
+epoch 12, total_step 191820, total loss is 12.04 , inference loss is 5.56, weight deacy loss is 6.49, training accuracy is 0.625000, time 55.666 samples/sec
+epoch 12, total_step 191840, total loss is 15.06 , inference loss is 8.58, weight deacy loss is 6.49, training accuracy is 0.468750, time 56.244 samples/sec
+epoch 12, total_step 191860, total loss is 14.79 , inference loss is 8.30, weight deacy loss is 6.49, training accuracy is 0.406250, time 56.294 samples/sec
+epoch 12, total_step 191880, total loss is 16.02 , inference loss is 9.54, weight deacy loss is 6.49, training accuracy is 0.437500, time 57.027 samples/sec
+epoch 12, total_step 191900, total loss is 13.89 , inference loss is 7.40, weight deacy loss is 6.49, training accuracy is 0.406250, time 56.736 samples/sec
+epoch 12, total_step 191920, total loss is 13.69 , inference loss is 7.20, weight deacy loss is 6.49, training accuracy is 0.500000, time 56.500 samples/sec
+epoch 12, total_step 191940, total loss is 11.80 , inference loss is 5.31, weight deacy loss is 6.49, training accuracy is 0.531250, time 56.502 samples/sec
+epoch 12, total_step 191960, total loss is 15.82 , inference loss is 9.33, weight deacy loss is 6.49, training accuracy is 0.468750, time 55.253 samples/sec
+epoch 12, total_step 191980, total loss is 12.06 , inference loss is 5.58, weight deacy loss is 6.49, training accuracy is 0.500000, time 57.480 samples/sec
+testing verification..
+(12000, 512)
+infer time 73.42206199999993
+best_threshold_index 113 0.9864814814814815
+best_threshold_index 113 0.9853703703703703
+best_threshold_index 112 0.9859259259259259
+best_threshold_index 113 0.9872222222222222
+best_threshold_index 114 0.9859259259259259
+best_threshold_index 113 0.985
+best_threshold_index 112 0.9857407407407407
+best_threshold_index 113 0.9848148148148148
+best_threshold_index 113 0.9844444444444445
+best_threshold_index 113 0.9844444444444445
+[lfw][192000]XNorm: 21.589511
+[lfw][192000]Accuracy-Flip: 0.98467+-0.00792
+testing verification..
+(14000, 512)
+infer time 85.227685
+best_threshold_index 125 0.9853968253968254
+best_threshold_index 125 0.9846031746031746
+best_threshold_index 125 0.9855555555555555
+best_threshold_index 124 0.9844444444444445
+best_threshold_index 122 0.9850793650793651
+best_threshold_index 123 0.9846031746031746
+best_threshold_index 125 0.9852380952380952
+best_threshold_index 125 0.9844444444444445
+best_threshold_index 125 0.9850793650793651
+best_threshold_index 125 0.9855555555555555
+[cfp_ff][192000]XNorm: 20.944467
+[cfp_ff][192000]Accuracy-Flip: 0.98400+-0.00388
+testing verification..
+(14000, 512)
+infer time 85.28094899999998
+best_threshold_index 143 0.9276190476190476
+best_threshold_index 143 0.9266666666666666
+best_threshold_index 143 0.9273015873015873
+best_threshold_index 143 0.9295238095238095
+best_threshold_index 143 0.9306349206349206
+best_threshold_index
\ No newline at end of file
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/figures/if_npu.log b/TensorFlow/contrib/cv/InsightFace_TF/figures/if_npu.log
new file mode 100644
index 0000000000000000000000000000000000000000..1da22546aa58ca438175db51bf107bcb564995b1
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/figures/if_npu.log
@@ -0,0 +1,18141 @@
+nohup: ignoring input
+WARNING:tensorflow:From /usr/local/Ascend/tfplugin/latest/python/site-packages/npu_bridge/estimator/npu/npu_optimizer.py:128: The name tf.train.Optimizer is deprecated. Please use tf.compat.v1.train.Optimizer instead.
+
+WARNING:tensorflow:From /usr/local/python3.7.5/lib/python3.7/site-packages/tensorlayer/layers.py:28: The name tf.GraphKeys is deprecated. Please use tf.compat.v1.GraphKeys instead.
+
+WARNING:tensorflow:From /usr/local/python3.7.5/lib/python3.7/site-packages/tensorlayer/layers.py:4030: The name tf.nn.max_pool is deprecated. Please use tf.nn.max_pool2d instead.
+
+WARNING:tensorflow:From train_nets.py:51: The name tf.assign_add is deprecated. Please use tf.compat.v1.assign_add instead.
+
+WARNING:tensorflow:From train_nets.py:52: The name tf.placeholder is deprecated. Please use tf.compat.v1.placeholder instead.
+
+WARNING:tensorflow:From /usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/autograph/converters/directives.py:119: The name tf.FixedLenFeature is deprecated. Please use tf.io.FixedLenFeature instead.
+
+WARNING:tensorflow:From /usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/autograph/converters/directives.py:119: The name tf.parse_single_example is deprecated. Please use tf.io.parse_single_example instead.
+
+2022-11-30 23:57:30.309709: W tensorflow/stream_executor/platform/default/dso_loader.cc:55] Could not load dynamic library 'libcuda.so.1'; dlerror: libcuda.so.1: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /usr/local/python3.7.5/lib/python3.7/site-packages/cv2/../../lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/python3.7.5/lib:/usr/local/Ascend/ascend-toolkit/latest/lib64:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/opskernel:/usr/local/Ascend/ascend-toolkit/latest/lib64/plugin/nnengine:/usr/local/Ascend/driver/lib64:/usr/local/Ascend/driver/lib64/common:/usr/local/Ascend/driver/lib64/driver:/usr/local/python3.7.5/lib:
+2022-11-30 23:57:30.309768: E tensorflow/stream_executor/cuda/cuda_driver.cc:318] failed call to cuInit: UNKNOWN ERROR (303)
+2022-11-30 23:57:30.309799: I tensorflow/stream_executor/cuda/cuda_diagnostics.cc:156] kernel driver does not appear to be running on this host (ubuntu): /proc/driver/nvidia/version does not exist
+WARNING:tensorflow:From train_nets.py:67: DatasetV1.make_initializable_iterator (from tensorflow.python.data.ops.dataset_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Use `for ... in dataset:` to iterate over a dataset. If using `tf.estimator`, return the `Dataset` object directly from your input function. As a last resort, you can use `tf.compat.v1.data.make_initializable_iterator(dataset)`.
+WARNING:tensorflow:
+The TensorFlow contrib module will not be included in TensorFlow 2.0.
+For more information, please see:
+ * https://github.com/tensorflow/community/blob/master/rfcs/20180907-contrib-sunset.md
+ * https://github.com/tensorflow/addons
+ * https://github.com/tensorflow/io (for I/O related ops)
+If you depend on functionality not listed there, please file an issue.
+
+WARNING:tensorflow:From /home/test_user06/dave/transfer/InsightFace_TF_npu_20221130111637/nets/L_Resnet_E_IR_fix_issue9.py:295: The name tf.variable_scope is deprecated. Please use tf.compat.v1.variable_scope instead.
+
+WARNING:tensorflow:From /usr/local/python3.7.5/lib/python3.7/site-packages/tensorlayer/layers.py:288: The name tf.get_variable_scope is deprecated. Please use tf.compat.v1.get_variable_scope instead.
+
+WARNING:tensorflow:From /usr/local/python3.7.5/lib/python3.7/site-packages/tensorlayer/layers.py:1385: The name tf.get_variable is deprecated. Please use tf.compat.v1.get_variable instead.
+
+tfrecords_f path : ./datasets/tfrecords_webface/tran.tfrecords
+parse_function :
+let look at the next_element:
+32
+begin db lfw convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+(12000, 112, 112, 3)
+ [TL] InputLayer resnet_v1_100/input_layer: (?, 112, 112, 3)
+ [TL] Conv2dLayer resnet_v1_100/conv1: shape:[3, 3, 3, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/bn0: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/bn0/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/bn0/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/prelu0: channel_shared:False
+ [TL] PoolLayer resnet_v1_100/block1/unit_1/bottleneck_v1/shortcut: ksize:[1, 1, 1, 1] strides:[1, 2, 2, 1] padding:SAME pool:max_pool
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block1/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_1/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_2/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_3/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 64, 128] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block2/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_1/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_2/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_3/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_4/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_5/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_6/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_7/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_8/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_9/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_10/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_11/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_12/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_13/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 128, 256] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1: shape:[3, 3, 128, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block3/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_1/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_2/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_3/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_4/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_5/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_6/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_7/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_8/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_9/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_10/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_11/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_12/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_13/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_14/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_15/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_16/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_17/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_18/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_19/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_20/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_21/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_22/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_23/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_24/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_25/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_26/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_27/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_28/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_29/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_30/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 256, 512] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1: shape:[3, 3, 256, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block4/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:TrueWARNING:tensorflow:From /home/test_user06/dave/transfer/InsightFace_TF_npu_20221130111637/losses/face_losses.py:21: calling norm (from tensorflow.python.ops.linalg_ops) with keep_dims is deprecated and will be removed in a future version.
+Instructions for updating:
+keep_dims is deprecated, use keepdims instead
+WARNING:tensorflow:From /home/test_user06/dave/transfer/InsightFace_TF_npu_20221130111637/losses/face_losses.py:22: div (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Deprecated in favor of operator or tf.math.divide.
+WARNING:tensorflow:From /home/test_user06/dave/transfer/InsightFace_TF_npu_20221130111637/losses/face_losses.py:41: where (from tensorflow.python.ops.array_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Use tf.where in 2.0, which has the same broadcast rule as np.where
+
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_1/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_2/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_3/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/E_BN1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/E_BN1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/E_BN1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ReshapeLayer resnet_v1_100/E_Reshapelayer: (?, 25088)
+ [TL] DenseLayer resnet_v1_100/E_DenseLayer: 512 identity
+ [TL] BatchNormLayer resnet_v1_100/E_BN2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/E_BN2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/E_BN2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] InputLayer resnet_v1_100/input_layer: (?, 112, 112, 3)
+ [TL] Conv2dLayer resnet_v1_100/conv1: shape:[3, 3, 3, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/bn0: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/prelu0: channel_shared:False
+ [TL] PoolLayer resnet_v1_100/block1/unit_1/bottleneck_v1/shortcut: ksize:[1, 1, 1, 1] strides:[1, 2, 2, 1] padding:SAME pool:max_pool
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block1/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_1/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_2/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_3/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 64, 128] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block2/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_1/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_2/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_3/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_4/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_5/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_6/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_7/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_8/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_9/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_10/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_11/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_12/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_13/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 128, 256] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1: shape:[3, 3, 128, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block3/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_1/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_2/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_3/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_4/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_5/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_6/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_7/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_8/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_9/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_10/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_11/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_12/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_13/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_14/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_15/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_16/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_17/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_18/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_19/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_20/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_21/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_22/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_23/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_24/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_25/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_26/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_27/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_28/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_29/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_30/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 256, 512] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1: shape:[3, 3, 256, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block4/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_1/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_2/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_3/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/E_BN1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ReshapeLayer resnet_v1_100/E_Reshapelayer: (?, 25088)
+ [TL] DenseLayer resnet_v1_100/E_DenseLayer: 512 identityWARNING:tensorflow:From /usr/local/python3.7.5/lib/python3.7/site-packages/tensorlayer/layers.py:169: The name tf.trainable_variables is deprecated. Please use tf.compat.v1.trainable_variables instead.
+
+
+ [TL] BatchNormLayer resnet_v1_100/E_BN2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [*] geting variables with W_conv2d
+ got 0: resnet_v1_100/conv1/W_conv2d:0 (3, 3, 3, 64)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 2: resnet_v1_100/block1/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 3: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 4: resnet_v1_100/block1/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 5: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 6: resnet_v1_100/block1/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 7: resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 64, 128)
+ got 8: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 128)
+ got 9: resnet_v1_100/block2/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 10: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 11: resnet_v1_100/block2/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 12: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 13: resnet_v1_100/block2/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 14: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 15: resnet_v1_100/block2/unit_4/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 16: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 17: resnet_v1_100/block2/unit_5/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 18: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 19: resnet_v1_100/block2/unit_6/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 20: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 21: resnet_v1_100/block2/unit_7/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 22: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 23: resnet_v1_100/block2/unit_8/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 24: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 25: resnet_v1_100/block2/unit_9/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 26: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 27: resnet_v1_100/block2/unit_10/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 28: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 29: resnet_v1_100/block2/unit_11/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 30: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 31: resnet_v1_100/block2/unit_12/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 32: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 33: resnet_v1_100/block2/unit_13/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 34: resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 128, 256)
+ got 35: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 256)
+ got 36: resnet_v1_100/block3/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 37: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 38: resnet_v1_100/block3/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 39: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 40: resnet_v1_100/block3/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 41: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 42: resnet_v1_100/block3/unit_4/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 43: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 44: resnet_v1_100/block3/unit_5/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 45: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 46: resnet_v1_100/block3/unit_6/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 47: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 48: resnet_v1_100/block3/unit_7/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 49: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 50: resnet_v1_100/block3/unit_8/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 51: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 52: resnet_v1_100/block3/unit_9/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 53: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 54: resnet_v1_100/block3/unit_10/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 55: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 56: resnet_v1_100/block3/unit_11/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 57: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 58: resnet_v1_100/block3/unit_12/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 59: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 60: resnet_v1_100/block3/unit_13/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 61: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 62: resnet_v1_100/block3/unit_14/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 63: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 64: resnet_v1_100/block3/unit_15/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 65: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 66: resnet_v1_100/block3/unit_16/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 67: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 68: resnet_v1_100/block3/unit_17/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 69: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 70: resnet_v1_100/block3/unit_18/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 71: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 72: resnet_v1_100/block3/unit_19/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 73: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 74: resnet_v1_100/block3/unit_20/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 75: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 76: resnet_v1_100/block3/unit_21/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 77: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 78: resnet_v1_100/block3/unit_22/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 79: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 80: resnet_v1_100/block3/unit_23/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 81: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 82: resnet_v1_100/block3/unit_24/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 83: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 84: resnet_v1_100/block3/unit_25/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 85: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 86: resnet_v1_100/block3/unit_26/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 87: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 88: resnet_v1_100/block3/unit_27/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 89: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 90: resnet_v1_100/block3/unit_28/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 91: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 92: resnet_v1_100/block3/unit_29/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 93: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 94: resnet_v1_100/block3/unit_30/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 95: resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 256, 512)
+ got 96: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 512)
+ got 97: resnet_v1_100/block4/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ got 98: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 512, 512)
+ got 99: resnet_v1_100/block4/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ got 100: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 512, 512)
+ got 101: resnet_v1_100/block4/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ [*] geting variables with resnet_v1_50/E_DenseLayer/W
+ [*] geting variables with embedding_weights
+ got 0: arcface_loss/embedding_weights:0 (512, 85164)
+ [*] geting variables with gamma
+ got 0: resnet_v1_100/bn0/gamma:0 (64,)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 2: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 3: resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 4: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 5: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 6: resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 7: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 8: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 9: resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 10: resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (128,)
+ got 11: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 12: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 13: resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 14: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 15: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 16: resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 17: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 18: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 19: resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 20: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 21: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 22: resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 23: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 24: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 25: resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 26: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 27: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 28: resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 29: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 30: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 31: resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 32: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 33: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 34: resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 35: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 36: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 37: resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 38: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 39: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 40: resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 41: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 42: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 43: resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 44: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 45: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 46: resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 47: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 48: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 49: resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 50: resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (256,)
+ got 51: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 52: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 53: resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 54: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 55: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 56: resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 57: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 58: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 59: resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 60: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 61: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 62: resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 63: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 64: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 65: resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 66: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 67: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 68: resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 69: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 70: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 71: resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 72: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 73: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 74: resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 75: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 76: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 77: resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 78: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 79: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 80: resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 81: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 82: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 83: resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 84: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 85: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 86: resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 87: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 88: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 89: resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 90: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 91: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 92: resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 93: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 94: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 95: resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 96: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 97: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 98: resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 99: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 100: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 101: resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 102: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 103: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 104: resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 105: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 106: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 107: resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 108: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 109: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 110: resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 111: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 112: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 113: resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 114: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 115: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 116: resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 117: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 118: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 119: resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 120: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 121: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 122: resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 123: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 124: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 125: resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 126: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 127: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 128: resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 129: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 130: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 131: resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 132: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 133: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 134: resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 135: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 136: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 137: resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 138: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 139: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 140: resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 141: resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (512,)
+ got 142: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 143: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 144: resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 145: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (512,)
+ got 146: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 147: resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 148: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (512,)
+ got 149: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 150: resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 151: resnet_v1_100/E_BN1/gamma:0 (512,)
+ [*] geting variables with alphas
+ got 0: resnet_v1_100/prelu0/alphas:0 (64,)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 2: resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 3: resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 4: resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 5: resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 6: resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 7: resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 8: resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 9: resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 10: resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 11: resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 12: resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 13: resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 14: resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 15: resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 16: resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 17: resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 18: resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 19: resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 20: resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 21: resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 22: resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 23: resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 24: resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 25: resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 26: resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 27: resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 28: resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 29: resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 30: resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 31: resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer/alphas:0 (256,)WARNING:tensorflow:From train_nets.py:118: The name tf.train.piecewise_constant is deprecated. Please use tf.compat.v1.train.piecewise_constant instead.
+
+WARNING:tensorflow:From train_nets.py:120: The name tf.train.MomentumOptimizer is deprecated. Please use tf.compat.v1.train.MomentumOptimizer instead.
+
+WARNING:tensorflow:From train_nets.py:123: The name tf.get_collection is deprecated. Please use tf.compat.v1.get_collection instead.
+
+WARNING:tensorflow:From train_nets.py:131: The name tf.ConfigProto is deprecated. Please use tf.compat.v1.ConfigProto instead.
+
+WARNING:tensorflow:From train_nets.py:134: The name tf.Session is deprecated. Please use tf.compat.v1.Session instead.
+
+2022-12-01 00:01:04.831203: I tensorflow/core/platform/cpu_feature_guard.cc:142] Your CPU supports instructions that this TensorFlow binary was not compiled to use: AVX2 AVX512F FMA
+2022-12-01 00:01:04.867231: I tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency: 2600000000 Hz
+2022-12-01 00:01:04.873637: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x55e0e7866050 initialized for platform Host (this does not guarantee that XLA will be used). Devices:
+2022-12-01 00:01:04.873693: I tensorflow/compiler/xla/service/service.cc:176] StreamExecutor device (0): Host, Default Version
+WARNING:tensorflow:From train_nets.py:136: The name tf.summary.FileWriter is deprecated. Please use tf.compat.v1.summary.FileWriter instead.
+
+WARNING:tensorflow:From train_nets.py:141: The name tf.summary.histogram is deprecated. Please use tf.compat.v1.summary.histogram instead.
+
+WARNING:tensorflow:From train_nets.py:146: The name tf.summary.scalar is deprecated. Please use tf.compat.v1.summary.scalar instead.
+
+WARNING:tensorflow:From train_nets.py:151: The name tf.summary.merge is deprecated. Please use tf.compat.v1.summary.merge instead.
+
+WARNING:tensorflow:From train_nets.py:153: The name tf.train.Saver is deprecated. Please use tf.compat.v1.train.Saver instead.
+
+WARNING:tensorflow:From train_nets.py:155: The name tf.global_variables_initializer is deprecated. Please use tf.compat.v1.global_variables_initializer instead.
+
+2022-12-01 00:01:15.663888: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:15.670381: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:15.670533: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:15.670549: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:15.670559: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:16.884690: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:16.884772: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:16.884784: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:16.884789: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:16.893457: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:16.893522: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:16.893564: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:16.893574: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:16.893578: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:16.893672: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:16.893708: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:16.893714: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:16.893719: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:16.893770: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+
+ got 32: resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 33: resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 34: resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 35: resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 36: resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 37: resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 38: resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 39: resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 40: resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 41: resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 42: resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 43: resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 44: resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 45: resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 46: resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 47: resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer/alphas:0 (512,)
+ got 48: resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer/alphas:0 (512,)
+ got 49: resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer/alphas:0 (512,)
+[640000, 960000, 1280000]
+2022-12-01 00:01:35.401739: I tf_adapter/kernels/geop_npu.cc:854] The model has been compiled on the Ascend AI processor, current graph id is: 1
+2022-12-01 00:01:55.383488: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.383630: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.383700: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.383711: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:55.383717: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:55.384121: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.384190: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.384200: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:55.384208: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:55.395218: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.395278: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.395330: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.395342: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:55.395349: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:55.395447: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.395494: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:55.395500: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:55.395505: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:55.395569: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.258483: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.258624: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.258676: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.258685: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:56.258690: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:56.259146: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.259204: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.259211: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:56.259215: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:56.262415: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.262459: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.262508: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.262515: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:56.262519: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:56.262606: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.262650: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:01:56.262656: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:56.262659: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:01:56.262698: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:06.263893: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7480 of 10000
+2022-12-01 00:02:09.775751: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+2022-12-01 00:02:23.358866: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:23.375295: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:23.375401: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:23.375413: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:02:23.375427: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:02:29.970795: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:29.970885: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:29.970898: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:02:29.970903: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:02:29.981158: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:29.981215: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:29.981259: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:29.981268: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:02:29.981273: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:02:29.981387: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:29.981426: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:29.981432: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:02:29.981437: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:02:29.981491: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:02:37.058312: I tf_adapter/kernels/geop_npu.cc:854] The model has been compiled on the Ascend AI processor, current graph id is: 11
+2022-12-01 00:05:50.449880: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:50.477049: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:50.477196: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:50.477217: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:05:50.477226: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:05:56.688123: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:56.688284: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:56.688307: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:05:56.688318: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:05:57.395126: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:57.396227: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:57.396311: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:57.396326: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:05:57.396335: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:05:57.397619: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:57.397696: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:05:57.397708: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:05:57.397716: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:05:57.398994: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:06:06.431427: I tf_adapter/kernels/geop_npu.cc:854] The model has been compiled on the Ascend AI processor, current graph id is: 21
+epoch 0, total_step 20, total loss is 58.00 , inference loss is 43.53, weight deacy loss is 14.47, training accuracy is 0.000000, time 229.218 samples/sec
+epoch 0, total_step 40, total loss is 59.88 , inference loss is 45.40, weight deacy loss is 14.48, training accuracy is 0.000000, time 225.196 samples/sec
+epoch 0, total_step 60, total loss is 57.18 , inference loss is 42.70, weight deacy loss is 14.48, training accuracy is 0.000000, time 231.740 samples/sec
+epoch 0, total_step 80, total loss is 53.08 , inference loss is 38.61, weight deacy loss is 14.48, training accuracy is 0.000000, time 215.144 samples/sec
+epoch 0, total_step 100, total loss is 52.97 , inference loss is 38.50, weight deacy loss is 14.48, training accuracy is 0.000000, time 230.647 samples/sec
+epoch 0, total_step 120, total loss is 53.69 , inference loss is 39.22, weight deacy loss is 14.48, training accuracy is 0.000000, time 205.789 samples/sec
+epoch 0, total_step 140, total loss is 50.71 , inference loss is 36.23, weight deacy loss is 14.47, training accuracy is 0.000000, time 209.200 samples/sec
+epoch 0, total_step 160, total loss is 52.98 , inference loss is 38.51, weight deacy loss is 14.47, training accuracy is 0.000000, time 211.722 samples/sec
+epoch 0, total_step 180, total loss is 51.75 , inference loss is 37.28, weight deacy loss is 14.47, training accuracy is 0.000000, time 230.534 samples/sec
+epoch 0, total_step 200, total loss is 54.64 , inference loss is 40.17, weight deacy loss is 14.47, training accuracy is 0.000000, time 217.277 samples/sec
+epoch 0, total_step 220, total loss is 51.94 , inference loss is 37.47, weight deacy loss is 14.47, training accuracy is 0.000000, time 224.246 samples/sec
+epoch 0, total_step 240, total loss is 50.74 , inference loss is 36.27, weight deacy loss is 14.47, training accuracy is 0.000000, time 225.463 samples/sec
+epoch 0, total_step 260, total loss is 51.68 , inference loss is 37.21, weight deacy loss is 14.47, training accuracy is 0.000000, time 307.132 samples/sec
+epoch 0, total_step 280, total loss is 49.30 , inference loss is 34.83, weight deacy loss is 14.46, training accuracy is 0.000000, time 307.546 samples/sec
+epoch 0, total_step 300, total loss is 50.46 , inference loss is 36.00, weight deacy loss is 14.46, training accuracy is 0.031250, time 11.770 samples/sec
+epoch 0, total_step 320, total loss is 47.11 , inference loss is 32.65, weight deacy loss is 14.46, training accuracy is 0.031250, time 231.697 samples/sec
+epoch 0, total_step 340, total loss is 50.18 , inference loss is 35.72, weight deacy loss is 14.46, training accuracy is 0.031250, time 234.566 samples/sec
+epoch 0, total_step 360, total loss is 45.88 , inference loss is 31.43, weight deacy loss is 14.46, training accuracy is 0.031250, time 216.966 samples/sec
+epoch 0, total_step 380, total loss is 49.33 , inference loss is 34.88, weight deacy loss is 14.46, training accuracy is 0.000000, time 211.972 samples/sec
+epoch 0, total_step 400, total loss is 49.87 , inference loss is 35.41, weight deacy loss is 14.46, training accuracy is 0.000000, time 230.757 samples/sec
+epoch 0, total_step 420, total loss is 48.85 , inference loss is 34.40, weight deacy loss is 14.45, training accuracy is 0.000000, time 226.519 samples/sec
+epoch 0, total_step 440, total loss is 50.17 , inference loss is 35.71, weight deacy loss is 14.45, training accuracy is 0.000000, time 227.768 samples/sec
+epoch 0, total_step 460, total loss is 45.53 , inference loss is 31.08, weight deacy loss is 14.45, training accuracy is 0.031250, time 308.002 samples/sec
+epoch 0, total_step 480, total loss is 48.24 , inference loss is 33.79, weight deacy loss is 14.45, training accuracy is 0.000000, time 230.973 samples/sec
+epoch 0, total_step 500, total loss is 48.08 , inference loss is 33.63, weight deacy loss is 14.45, training accuracy is 0.000000, time 225.139 samples/sec
+epoch 0, total_step 520, total loss is 49.79 , inference loss is 35.34, weight deacy loss is 14.45, training accuracy is 0.031250, time 223.400 samples/sec
+epoch 0, total_step 540, total loss is 48.29 , inference loss is 33.85, weight deacy loss is 14.44, training accuracy is 0.000000, time 195.154 samples/sec
+epoch 0, total_step 560, total loss is 41.78 , inference loss is 27.34, weight deacy loss is 14.44, training accuracy is 0.062500, time 232.673 samples/sec
+epoch 0, total_step 580, total loss is 50.05 , inference loss is 35.61, weight deacy loss is 14.44, training accuracy is 0.000000, time 238.076 samples/sec
+epoch 0, total_step 600, total loss is 45.07 , inference loss is 30.63, weight deacy loss is 14.44, training accuracy is 0.000000, time 11.912 samples/sec
+epoch 0, total_step 620, total loss is 47.14 , inference loss is 32.70, weight deacy loss is 14.44, training accuracy is 0.000000, time 219.179 samples/sec
+epoch 0, total_step 640, total loss is 47.14 , inference loss is 32.70, weight deacy loss is 14.44, training accuracy is 0.031250, time 226.400 samples/sec
+epoch 0, total_step 660, total loss is 46.03 , inference loss is 31.60, weight deacy loss is 14.43, training accuracy is 0.062500, time 214.944 samples/sec
+epoch 0, total_step 680, total loss is 46.66 , inference loss is 32.23, weight deacy loss is 14.43, training accuracy is 0.031250, time 214.748 samples/sec
+epoch 0, total_step 700, total loss is 42.60 , inference loss is 28.17, weight deacy loss is 14.43, training accuracy is 0.000000, time 232.199 samples/sec
+epoch 0, total_step 720, total loss is 46.23 , inference loss is 31.80, weight deacy loss is 14.43, training accuracy is 0.000000, time 307.828 samples/sec
+epoch 0, total_step 740, total loss is 49.02 , inference loss is 34.59, weight deacy loss is 14.43, training accuracy is 0.000000, time 217.507 samples/sec
+epoch 0, total_step 760, total loss is 47.08 , inference loss is 32.65, weight deacy loss is 14.43, training accuracy is 0.000000, time 208.583 samples/sec
+epoch 0, total_step 780, total loss is 49.48 , inference loss is 35.05, weight deacy loss is 14.42, training accuracy is 0.000000, time 216.031 samples/sec
+epoch 0, total_step 800, total loss is 44.07 , inference loss is 29.65, weight deacy loss is 14.42, training accuracy is 0.000000, time 219.188 samples/sec
+epoch 0, total_step 820, total loss is 44.64 , inference loss is 30.22, weight deacy loss is 14.42, training accuracy is 0.000000, time 223.743 samples/sec
+epoch 0, total_step 840, total loss is 48.95 , inference loss is 34.53, weight deacy loss is 14.42, training accuracy is 0.000000, time 216.013 samples/sec
+epoch 0, total_step 860, total loss is 49.86 , inference loss is 35.44, weight deacy loss is 14.42, training accuracy is 0.000000, time 217.496 samples/sec
+epoch 0, total_step 880, total loss is 48.14 , inference loss is 33.72, weight deacy loss is 14.42, training accuracy is 0.000000, time 229.196 samples/sec
+epoch 0, total_step 900, total loss is 48.90 , inference loss is 34.49, weight deacy loss is 14.41, training accuracy is 0.000000, time 12.869 samples/sec
+epoch 0, total_step 920, total loss is 45.39 , inference loss is 30.98, weight deacy loss is 14.41, training accuracy is 0.031250, time 217.772 samples/sec
+epoch 0, total_step 940, total loss is 46.97 , inference loss is 32.56, weight deacy loss is 14.41, training accuracy is 0.000000, time 221.886 samples/sec
+epoch 0, total_step 960, total loss is 48.81 , inference loss is 34.40, weight deacy loss is 14.41, training accuracy is 0.000000, time 231.360 samples/sec
+epoch 0, total_step 980, total loss is 47.86 , inference loss is 33.45, weight deacy loss is 14.41, training accuracy is 0.000000, time 307.514 samples/sec
+epoch 0, total_step 1000, total loss is 48.09 , inference loss is 33.68, weight deacy loss is 14.41, training accuracy is 0.000000, time 227.155 samples/sec
+epoch 0, total_step 1020, total loss is 44.16 , inference loss is 29.76, weight deacy loss is 14.40, training accuracy is 0.031250, time 226.580 samples/sec
+epoch 0, total_step 1040, total loss is 46.34 , inference loss is 31.94, weight deacy loss is 14.40, training accuracy is 0.031250, time 231.529 samples/sec
+epoch 0, total_step 1060, total loss is 49.14 , inference loss is 34.74, weight deacy loss is 14.40, training accuracy is 0.000000, time 188.304 samples/sec2022-12-01 00:14:51.865527: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:51.867125: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:51.867195: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:51.867205: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:14:51.867210: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:14:52.089229: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:52.089308: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:52.089318: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:14:52.089323: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:14:52.118131: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:52.118199: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:52.118241: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:52.118249: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:14:52.118264: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:14:52.118360: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:52.118395: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:52.118401: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:14:52.118405: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:14:52.118452: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:14:52.951850: I tf_adapter/kernels/geop_npu.cc:854] The model has been compiled on the Ascend AI processor, current graph id is: 31
+2022-12-01 00:16:08.326466: I tf_adapter/kernels/geop_npu.cc:854] The model has been compiled on the Ascend AI processor, current graph id is: 11
+
+epoch 0, total_step 1080, total loss is 46.10 , inference loss is 31.70, weight deacy loss is 14.40, training accuracy is 0.000000, time 228.987 samples/sec
+epoch 0, total_step 1100, total loss is 49.23 , inference loss is 34.83, weight deacy loss is 14.40, training accuracy is 0.031250, time 230.325 samples/sec
+epoch 0, total_step 1120, total loss is 46.71 , inference loss is 32.32, weight deacy loss is 14.40, training accuracy is 0.000000, time 203.241 samples/sec
+epoch 0, total_step 1140, total loss is 48.76 , inference loss is 34.37, weight deacy loss is 14.39, training accuracy is 0.000000, time 220.196 samples/sec
+epoch 0, total_step 1160, total loss is 45.98 , inference loss is 31.58, weight deacy loss is 14.39, training accuracy is 0.031250, time 231.192 samples/sec
+epoch 0, total_step 1180, total loss is 47.62 , inference loss is 33.23, weight deacy loss is 14.39, training accuracy is 0.000000, time 182.470 samples/sec
+epoch 0, total_step 1200, total loss is 44.98 , inference loss is 30.59, weight deacy loss is 14.39, training accuracy is 0.031250, time 13.806 samples/sec
+epoch 0, total_step 1220, total loss is 46.43 , inference loss is 32.04, weight deacy loss is 14.39, training accuracy is 0.000000, time 206.836 samples/sec
+epoch 0, total_step 1240, total loss is 45.82 , inference loss is 31.43, weight deacy loss is 14.39, training accuracy is 0.000000, time 308.028 samples/sec
+epoch 0, total_step 1260, total loss is 45.33 , inference loss is 30.95, weight deacy loss is 14.38, training accuracy is 0.031250, time 201.451 samples/sec
+epoch 0, total_step 1280, total loss is 50.14 , inference loss is 35.76, weight deacy loss is 14.38, training accuracy is 0.000000, time 241.663 samples/sec
+epoch 0, total_step 1300, total loss is 44.44 , inference loss is 30.06, weight deacy loss is 14.38, training accuracy is 0.000000, time 228.695 samples/sec
+epoch 0, total_step 1320, total loss is 46.58 , inference loss is 32.20, weight deacy loss is 14.38, training accuracy is 0.031250, time 239.792 samples/sec
+epoch 0, total_step 1340, total loss is 43.13 , inference loss is 28.75, weight deacy loss is 14.38, training accuracy is 0.062500, time 213.824 samples/sec
+epoch 0, total_step 1360, total loss is 42.82 , inference loss is 28.45, weight deacy loss is 14.38, training accuracy is 0.031250, time 210.143 samples/sec
+epoch 0, total_step 1380, total loss is 43.23 , inference loss is 28.85, weight deacy loss is 14.37, training accuracy is 0.062500, time 229.791 samples/sec
+epoch 0, total_step 1400, total loss is 44.59 , inference loss is 30.21, weight deacy loss is 14.37, training accuracy is 0.031250, time 230.106 samples/sec
+epoch 0, total_step 1420, total loss is 50.86 , inference loss is 36.49, weight deacy loss is 14.37, training accuracy is 0.000000, time 219.408 samples/sec
+epoch 0, total_step 1440, total loss is 46.86 , inference loss is 32.49, weight deacy loss is 14.37, training accuracy is 0.000000, time 237.366 samples/sec
+epoch 0, total_step 1460, total loss is 47.11 , inference loss is 32.74, weight deacy loss is 14.37, training accuracy is 0.000000, time 227.834 samples/sec
+epoch 0, total_step 1480, total loss is 49.36 , inference loss is 34.99, weight deacy loss is 14.37, training accuracy is 0.000000, time 243.561 samples/sec
+epoch 0, total_step 1500, total loss is 44.19 , inference loss is 29.83, weight deacy loss is 14.36, training accuracy is 0.000000, time 11.230 samples/sec
+epoch 0, total_step 1520, total loss is 43.71 , inference loss is 29.34, weight deacy loss is 14.36, training accuracy is 0.000000, time 224.357 samples/sec
+epoch 0, total_step 1540, total loss is 47.66 , inference loss is 33.30, weight deacy loss is 14.36, training accuracy is 0.000000, time 230.488 samples/sec
+epoch 0, total_step 1560, total loss is 46.10 , inference loss is 31.74, weight deacy loss is 14.36, training accuracy is 0.031250, time 230.514 samples/sec
+epoch 0, total_step 1580, total loss is 47.08 , inference loss is 32.72, weight deacy loss is 14.36, training accuracy is 0.000000, time 234.027 samples/sec
+epoch 0, total_step 1600, total loss is 43.86 , inference loss is 29.51, weight deacy loss is 14.36, training accuracy is 0.000000, time 242.328 samples/sec
+epoch 0, total_step 1620, total loss is 44.44 , inference loss is 30.09, weight deacy loss is 14.35, training accuracy is 0.000000, time 229.108 samples/sec
+epoch 0, total_step 1640, total loss is 43.77 , inference loss is 29.42, weight deacy loss is 14.35, training accuracy is 0.062500, time 222.003 samples/sec
+epoch 0, total_step 1660, total loss is 43.45 , inference loss is 29.10, weight deacy loss is 14.35, training accuracy is 0.031250, time 220.195 samples/sec
+epoch 0, total_step 1680, total loss is 41.13 , inference loss is 26.78, weight deacy loss is 14.35, training accuracy is 0.062500, time 234.592 samples/sec
+epoch 0, total_step 1700, total loss is 48.73 , inference loss is 34.38, weight deacy loss is 14.35, training accuracy is 0.000000, time 218.548 samples/sec
+epoch 0, total_step 1720, total loss is 41.62 , inference loss is 27.28, weight deacy loss is 14.35, training accuracy is 0.031250, time 216.976 samples/sec
+epoch 0, total_step 1740, total loss is 48.37 , inference loss is 34.03, weight deacy loss is 14.35, training accuracy is 0.000000, time 228.147 samples/sec
+epoch 0, total_step 1760, total loss is 46.83 , inference loss is 32.49, weight deacy loss is 14.34, training accuracy is 0.031250, time 219.310 samples/sec
+epoch 0, total_step 1780, total loss is 44.90 , inference loss is 30.56, weight deacy loss is 14.34, training accuracy is 0.031250, time 229.947 samples/sec
+epoch 0, total_step 1800, total loss is 43.59 , inference loss is 29.25, weight deacy loss is 14.34, training accuracy is 0.000000, time 14.094 samples/sec
+epoch 0, total_step 1820, total loss is 39.39 , inference loss is 25.05, weight deacy loss is 14.34, training accuracy is 0.062500, time 222.939 samples/sec
+epoch 0, total_step 1840, total loss is 49.11 , inference loss is 34.77, weight deacy loss is 14.34, training accuracy is 0.000000, time 234.387 samples/sec
+epoch 0, total_step 1860, total loss is 46.39 , inference loss is 32.05, weight deacy loss is 14.34, training accuracy is 0.031250, time 228.654 samples/sec
+epoch 0, total_step 1880, total loss is 45.34 , inference loss is 31.01, weight deacy loss is 14.33, training accuracy is 0.062500, time 223.622 samples/sec
+epoch 0, total_step 1900, total loss is 44.55 , inference loss is 30.22, weight deacy loss is 14.33, training accuracy is 0.062500, time 231.079 samples/sec
+epoch 0, total_step 1920, total loss is 45.71 , inference loss is 31.38, weight deacy loss is 14.33, training accuracy is 0.000000, time 308.354 samples/sec
+epoch 0, total_step 1940, total loss is 42.37 , inference loss is 28.04, weight deacy loss is 14.33, training accuracy is 0.125000, time 307.054 samples/sec
+epoch 0, total_step 1960, total loss is 42.83 , inference loss is 28.51, weight deacy loss is 14.33, training accuracy is 0.031250, time 229.505 samples/sec
+epoch 0, total_step 1980, total loss is 39.57 , inference loss is 25.24, weight deacy loss is 14.33, training accuracy is 0.031250, time 229.005 samples/sec
+testing verification..
+(12000, 512)
+infer time 51.93786499999995
+best_threshold_index 134 0.8825925925925926
+best_threshold_index 129 0.8833333333333333
+best_threshold_index 134 0.8803703703703704
+best_threshold_index 134 0.8818518518518519
+best_threshold_index 134 0.8868518518518519
+best_threshold_index 134 0.8803703703703704
+best_threshold_index 134 0.8792592592592593
+best_threshold_index 132 0.8818518518518519
+best_threshold_index 134 0.8816666666666667
+best_threshold_index 134 0.8809259259259259
+[lfw][2000]XNorm: 20.567539
+[lfw][2000]Accuracy-Flip: 0.87933+-0.01886
+test accuracy is: 0.8793333333333335
+epoch 0, total_step 2000, total loss is 45.31 , inference loss is 30.99, weight deacy loss is 14.32, training accuracy is 0.000000, time 0.189 samples/sec
+epoch 0, total_step 2020, total loss is 45.69 , inference loss is 31.36, weight deacy loss is 14.32, training accuracy is 0.031250, time 225.188 samples/sec
+epoch 0, total_step 2040, total loss is 43.91 , inference loss is 29.59, weight deacy loss is 14.32, training accuracy is 0.062500, time 236.274 samples/sec2022-12-01 00:19:17.618028: I tf_adapter/kernels/geop_npu.cc:854] The model has been compiled on the Ascend AI processor, current graph id is: 21
+
+epoch 0, total_step 2060, total loss is 41.72 , inference loss is 27.40, weight deacy loss is 14.32, training accuracy is 0.062500, time 224.270 samples/sec
+epoch 0, total_step 2080, total loss is 41.25 , inference loss is 26.93, weight deacy loss is 14.32, training accuracy is 0.093750, time 214.745 samples/sec
+epoch 0, total_step 2100, total loss is 43.23 , inference loss is 28.92, weight deacy loss is 14.32, training accuracy is 0.062500, time 11.888 samples/sec
+epoch 0, total_step 2120, total loss is 44.38 , inference loss is 30.06, weight deacy loss is 14.31, training accuracy is 0.031250, time 181.738 samples/sec
+epoch 0, total_step 2140, total loss is 41.66 , inference loss is 27.35, weight deacy loss is 14.31, training accuracy is 0.062500, time 211.725 samples/sec
+epoch 0, total_step 2160, total loss is 45.85 , inference loss is 31.54, weight deacy loss is 14.31, training accuracy is 0.000000, time 209.117 samples/sec
+epoch 0, total_step 2180, total loss is 38.56 , inference loss is 24.25, weight deacy loss is 14.31, training accuracy is 0.125000, time 223.834 samples/sec
+epoch 0, total_step 2200, total loss is 43.72 , inference loss is 29.41, weight deacy loss is 14.31, training accuracy is 0.031250, time 180.056 samples/sec
+epoch 0, total_step 2220, total loss is 39.60 , inference loss is 25.29, weight deacy loss is 14.31, training accuracy is 0.031250, time 215.383 samples/sec
+epoch 0, total_step 2240, total loss is 41.40 , inference loss is 27.10, weight deacy loss is 14.30, training accuracy is 0.093750, time 219.656 samples/sec
+epoch 0, total_step 2260, total loss is 41.64 , inference loss is 27.34, weight deacy loss is 14.30, training accuracy is 0.031250, time 224.060 samples/sec
+epoch 0, total_step 2280, total loss is 40.44 , inference loss is 26.14, weight deacy loss is 14.30, training accuracy is 0.125000, time 292.367 samples/sec
+epoch 0, total_step 2300, total loss is 43.51 , inference loss is 29.21, weight deacy loss is 14.30, training accuracy is 0.062500, time 225.071 samples/sec
+epoch 0, total_step 2320, total loss is 39.30 , inference loss is 25.00, weight deacy loss is 14.30, training accuracy is 0.031250, time 200.273 samples/sec
+epoch 0, total_step 2340, total loss is 41.71 , inference loss is 27.41, weight deacy loss is 14.30, training accuracy is 0.031250, time 204.280 samples/sec
+epoch 0, total_step 2360, total loss is 42.22 , inference loss is 27.92, weight deacy loss is 14.29, training accuracy is 0.093750, time 220.689 samples/sec
+epoch 0, total_step 2380, total loss is 41.50 , inference loss is 27.20, weight deacy loss is 14.29, training accuracy is 0.031250, time 225.219 samples/sec
+epoch 0, total_step 2400, total loss is 38.91 , inference loss is 24.62, weight deacy loss is 14.29, training accuracy is 0.031250, time 12.530 samples/sec
+epoch 0, total_step 2420, total loss is 40.06 , inference loss is 25.77, weight deacy loss is 14.29, training accuracy is 0.031250, time 207.353 samples/sec
+epoch 0, total_step 2440, total loss is 40.25 , inference loss is 25.96, weight deacy loss is 14.29, training accuracy is 0.062500, time 225.263 samples/sec
+epoch 0, total_step 2460, total loss is 41.07 , inference loss is 26.78, weight deacy loss is 14.29, training accuracy is 0.093750, time 225.432 samples/sec
+epoch 0, total_step 2480, total loss is 37.21 , inference loss is 22.92, weight deacy loss is 14.28, training accuracy is 0.062500, time 226.318 samples/sec
+epoch 0, total_step 2500, total loss is 43.92 , inference loss is 29.64, weight deacy loss is 14.28, training accuracy is 0.000000, time 227.330 samples/sec
+epoch 0, total_step 2520, total loss is 40.73 , inference loss is 26.45, weight deacy loss is 14.28, training accuracy is 0.062500, time 231.076 samples/sec
+epoch 0, total_step 2540, total loss is 39.79 , inference loss is 25.51, weight deacy loss is 14.28, training accuracy is 0.062500, time 298.846 samples/sec
+epoch 0, total_step 2560, total loss is 42.07 , inference loss is 27.79, weight deacy loss is 14.28, training accuracy is 0.031250, time 216.565 samples/sec
+epoch 0, total_step 2580, total loss is 41.09 , inference loss is 26.81, weight deacy loss is 14.28, training accuracy is 0.031250, time 219.927 samples/sec
+epoch 0, total_step 2600, total loss is 40.69 , inference loss is 26.41, weight deacy loss is 14.28, training accuracy is 0.000000, time 224.814 samples/sec
+epoch 0, total_step 2620, total loss is 40.73 , inference loss is 26.46, weight deacy loss is 14.27, training accuracy is 0.093750, time 209.706 samples/sec
+epoch 0, total_step 2640, total loss is 38.24 , inference loss is 23.97, weight deacy loss is 14.27, training accuracy is 0.062500, time 216.848 samples/sec
+epoch 0, total_step 2660, total loss is 33.58 , inference loss is 19.31, weight deacy loss is 14.27, training accuracy is 0.062500, time 228.096 samples/sec
+epoch 0, total_step 2680, total loss is 40.68 , inference loss is 26.41, weight deacy loss is 14.27, training accuracy is 0.000000, time 217.333 samples/sec
+epoch 0, total_step 2700, total loss is 41.59 , inference loss is 27.33, weight deacy loss is 14.27, training accuracy is 0.000000, time 12.260 samples/sec
+epoch 0, total_step 2720, total loss is 43.80 , inference loss is 29.53, weight deacy loss is 14.27, training accuracy is 0.000000, time 206.909 samples/sec
+epoch 0, total_step 2740, total loss is 40.66 , inference loss is 26.40, weight deacy loss is 14.26, training accuracy is 0.000000, time 228.168 samples/sec
+epoch 0, total_step 2760, total loss is 37.56 , inference loss is 23.29, weight deacy loss is 14.26, training accuracy is 0.062500, time 223.200 samples/sec
+epoch 0, total_step 2780, total loss is 41.92 , inference loss is 27.66, weight deacy loss is 14.26, training accuracy is 0.000000, time 225.122 samples/sec
+epoch 0, total_step 2800, total loss is 39.63 , inference loss is 25.38, weight deacy loss is 14.26, training accuracy is 0.031250, time 299.030 samples/sec
+epoch 0, total_step 2820, total loss is 38.68 , inference loss is 24.43, weight deacy loss is 14.26, training accuracy is 0.062500, time 227.212 samples/sec
+epoch 0, total_step 2840, total loss is 35.11 , inference loss is 20.85, weight deacy loss is 14.26, training accuracy is 0.031250, time 209.412 samples/sec
+epoch 0, total_step 2860, total loss is 42.69 , inference loss is 28.44, weight deacy loss is 14.25, training accuracy is 0.000000, time 227.119 samples/sec
+epoch 0, total_step 2880, total loss is 36.18 , inference loss is 21.92, weight deacy loss is 14.25, training accuracy is 0.031250, time 224.495 samples/sec
+epoch 0, total_step 2900, total loss is 42.99 , inference loss is 28.74, weight deacy loss is 14.25, training accuracy is 0.031250, time 218.760 samples/sec
+epoch 0, total_step 2920, total loss is 40.33 , inference loss is 26.08, weight deacy loss is 14.25, training accuracy is 0.000000, time 213.844 samples/sec
+epoch 0, total_step 2940, total loss is 43.78 , inference loss is 29.53, weight deacy loss is 14.25, training accuracy is 0.000000, time 204.397 samples/sec
+epoch 0, total_step 2960, total loss is 37.61 , inference loss is 23.36, weight deacy loss is 14.25, training accuracy is 0.062500, time 243.074 samples/sec
+epoch 0, total_step 2980, total loss is 38.01 , inference loss is 23.76, weight deacy loss is 14.24, training accuracy is 0.093750, time 224.376 samples/sec
+epoch 0, total_step 3000, total loss is 37.69 , inference loss is 23.44, weight deacy loss is 14.24, training accuracy is 0.125000, time 13.115 samples/sec
+epoch 0, total_step 3020, total loss is 49.66 , inference loss is 35.41, weight deacy loss is 14.24, training accuracy is 0.000000, time 206.125 samples/sec
+epoch 0, total_step 3040, total loss is 38.15 , inference loss is 23.91, weight deacy loss is 14.24, training accuracy is 0.031250, time 291.849 samples/sec
+epoch 0, total_step 3060, total loss is 37.27 , inference loss is 23.03, weight deacy loss is 14.24, training accuracy is 0.156250, time 210.583 samples/sec
+epoch 0, total_step 3080, total loss is 43.70 , inference loss is 29.46, weight deacy loss is 14.24, training accuracy is 0.000000, time 219.806 samples/sec
+epoch 0, total_step 3100, total loss is 41.47 , inference loss is 27.24, weight deacy loss is 14.24, training accuracy is 0.000000, time 224.165 samples/sec
+epoch 0, total_step 3120, total loss is 37.51 , inference loss is 23.28, weight deacy loss is 14.23, training accuracy is 0.062500, time 233.688 samples/sec
+epoch 0, total_step 3140, total loss is 39.36 , inference loss is 25.13, weight deacy loss is 14.23, training accuracy is 0.093750, time 212.830 samples/sec
+epoch 0, total_step 3160, total loss is 45.02 , inference loss is 30.79, weight deacy loss is 14.23, training accuracy is 0.031250, time 223.794 samples/sec
+epoch 0, total_step 3180, total loss is 38.34 , inference loss is 24.11, weight deacy loss is 14.23, training accuracy is 0.093750, time 199.600 samples/sec
+epoch 0, total_step 3200, total loss is 40.26 , inference loss is 26.03, weight deacy loss is 14.23, training accuracy is 0.031250, time 197.144 samples/sec
+epoch 0, total_step 3220, total loss is 47.33 , inference loss is 33.11, weight deacy loss is 14.23, training accuracy is 0.062500, time 225.237 samples/sec
+epoch 0, total_step 3240, total loss is 42.50 , inference loss is 28.28, weight deacy loss is 14.22, training accuracy is 0.031250, time 218.475 samples/sec
+epoch 0, total_step 3260, total loss is 41.35 , inference loss is 27.13, weight deacy loss is 14.22, training accuracy is 0.062500, time 222.349 samples/sec
+epoch 0, total_step 3280, total loss is 43.42 , inference loss is 29.19, weight deacy loss is 14.22, training accuracy is 0.125000, time 231.593 samples/sec
+epoch 0, total_step 3300, total loss is 41.34 , inference loss is 27.12, weight deacy loss is 14.22, training accuracy is 0.031250, time 12.561 samples/sec
+epoch 0, total_step 3320, total loss is 42.28 , inference loss is 28.06, weight deacy loss is 14.22, training accuracy is 0.062500, time 226.227 samples/sec
+epoch 0, total_step 3340, total loss is 44.45 , inference loss is 30.23, weight deacy loss is 14.22, training accuracy is 0.031250, time 207.744 samples/sec
+epoch 0, total_step 3360, total loss is 37.99 , inference loss is 23.77, weight deacy loss is 14.22, training accuracy is 0.125000, time 245.806 samples/sec
+epoch 0, total_step 3380, total loss is 40.78 , inference loss is 26.57, weight deacy loss is 14.21, training accuracy is 0.093750, time 214.675 samples/sec
+epoch 0, total_step 3400, total loss is 42.62 , inference loss is 28.41, weight deacy loss is 14.21, training accuracy is 0.000000, time 217.273 samples/sec
+epoch 0, total_step 3420, total loss is 42.98 , inference loss is 28.77, weight deacy loss is 14.21, training accuracy is 0.000000, time 299.034 samples/sec
+epoch 0, total_step 3440, total loss is 46.09 , inference loss is 31.88, weight deacy loss is 14.21, training accuracy is 0.000000, time 216.514 samples/sec
+epoch 0, total_step 3460, total loss is 40.49 , inference loss is 26.28, weight deacy loss is 14.21, training accuracy is 0.031250, time 205.028 samples/sec
+epoch 0, total_step 3480, total loss is 40.61 , inference loss is 26.40, weight deacy loss is 14.21, training accuracy is 0.000000, time 210.952 samples/sec
+epoch 0, total_step 3500, total loss is 40.73 , inference loss is 26.53, weight deacy loss is 14.21, training accuracy is 0.062500, time 221.954 samples/sec
+epoch 0, total_step 3520, total loss is 39.07 , inference loss is 24.87, weight deacy loss is 14.20, training accuracy is 0.000000, time 214.756 samples/sec
+epoch 0, total_step 3540, total loss is 41.13 , inference loss is 26.93, weight deacy loss is 14.20, training accuracy is 0.031250, time 222.537 samples/sec
+epoch 0, total_step 3560, total loss is 43.88 , inference loss is 29.68, weight deacy loss is 14.20, training accuracy is 0.062500, time 227.478 samples/sec
+epoch 0, total_step 3580, total loss is 43.54 , inference loss is 29.35, weight deacy loss is 14.20, training accuracy is 0.000000, time 297.400 samples/sec
+epoch 0, total_step 3600, total loss is 45.86 , inference loss is 31.67, weight deacy loss is 14.20, training accuracy is 0.062500, time 11.549 samples/sec
+epoch 0, total_step 3620, total loss is 40.71 , inference loss is 26.52, weight deacy loss is 14.20, training accuracy is 0.125000, time 219.476 samples/sec
+epoch 0, total_step 3640, total loss is 41.10 , inference loss is 26.91, weight deacy loss is 14.19, training accuracy is 0.000000, time 210.690 samples/sec
+epoch 0, total_step 3660, total loss is 45.96 , inference loss is 31.76, weight deacy loss is 14.19, training accuracy is 0.000000, time 206.760 samples/sec
+epoch 0, total_step 3680, total loss is 40.38 , inference loss is 26.19, weight deacy loss is 14.19, training accuracy is 0.031250, time 219.798 samples/sec
+epoch 0, total_step 3700, total loss is 43.34 , inference loss is 29.15, weight deacy loss is 14.19, training accuracy is 0.000000, time 232.858 samples/sec
+epoch 0, total_step 3720, total loss is 41.00 , inference loss is 26.81, weight deacy loss is 14.19, training accuracy is 0.062500, time 220.474 samples/sec
+epoch 0, total_step 3740, total loss is 42.76 , inference loss is 28.57, weight deacy loss is 14.19, training accuracy is 0.062500, time 207.040 samples/sec
+epoch 0, total_step 3760, total loss is 45.45 , inference loss is 31.26, weight deacy loss is 14.19, training accuracy is 0.031250, time 296.075 samples/sec
+epoch 0, total_step 3780, total loss is 47.40 , inference loss is 33.22, weight deacy loss is 14.18, training accuracy is 0.031250, time 208.643 samples/sec
+epoch 0, total_step 3800, total loss is 38.92 , inference loss is 24.74, weight deacy loss is 14.18, training accuracy is 0.031250, time 212.285 samples/sec
+epoch 0, total_step 3820, total loss is 43.97 , inference loss is 29.79, weight deacy loss is 14.18, training accuracy is 0.000000, time 219.559 samples/sec
+epoch 0, total_step 3840, total loss is 47.78 , inference loss is 33.60, weight deacy loss is 14.18, training accuracy is 0.031250, time 222.359 samples/sec
+epoch 0, total_step 3860, total loss is 38.08 , inference loss is 23.90, weight deacy loss is 14.18, training accuracy is 0.000000, time 218.014 samples/sec
+epoch 0, total_step 3880, total loss is 42.05 , inference loss is 27.88, weight deacy loss is 14.18, training accuracy is 0.000000, time 230.337 samples/sec
+epoch 0, total_step 3900, total loss is 40.32 , inference loss is 26.15, weight deacy loss is 14.18, training accuracy is 0.000000, time 13.813 samples/sec
+epoch 0, total_step 3920, total loss is 45.62 , inference loss is 31.45, weight deacy loss is 14.17, training accuracy is 0.062500, time 210.422 samples/sec
+epoch 0, total_step 3940, total loss is 42.70 , inference loss is 28.53, weight deacy loss is 14.17, training accuracy is 0.000000, time 224.680 samples/sec
+epoch 0, total_step 3960, total loss is 41.88 , inference loss is 27.70, weight deacy loss is 14.17, training accuracy is 0.000000, time 218.602 samples/sec
+epoch 0, total_step 3980, total loss is 41.24 , inference loss is 27.08, weight deacy loss is 14.17, training accuracy is 0.062500, time 247.280 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.909249999999986
+best_threshold_index 125 0.9033333333333333
+best_threshold_index 125 0.9046296296296297
+best_threshold_index 125 0.9072222222222223
+best_threshold_index 127 0.9042592592592592
+best_threshold_index 125 0.9070370370370371
+best_threshold_index 125 0.902962962962963
+best_threshold_index 125 0.9024074074074074
+best_threshold_index 125 0.9044444444444445
+best_threshold_index 125 0.9044444444444445
+best_threshold_index 125 0.9027777777777778
+[lfw][4000]XNorm: 23.632429
+[lfw][4000]Accuracy-Flip: 0.90350+-0.01427
+test accuracy is: 0.9035
+epoch 0, total_step 4000, total loss is 41.63 , inference loss is 27.46, weight deacy loss is 14.17, training accuracy is 0.031250, time 14.296 samples/sec
+epoch 0, total_step 4020, total loss is 44.52 , inference loss is 30.35, weight deacy loss is 14.17, training accuracy is 0.000000, time 209.118 samples/sec
+epoch 0, total_step 4040, total loss is 45.43 , inference loss is 31.27, weight deacy loss is 14.16, training accuracy is 0.000000, time 215.253 samples/sec
+epoch 0, total_step 4060, total loss is 40.85 , inference loss is 26.68, weight deacy loss is 14.16, training accuracy is 0.000000, time 225.616 samples/sec
+epoch 0, total_step 4080, total loss is 45.11 , inference loss is 30.95, weight deacy loss is 14.16, training accuracy is 0.031250, time 211.405 samples/sec
+epoch 0, total_step 4100, total loss is 42.31 , inference loss is 28.15, weight deacy loss is 14.16, training accuracy is 0.000000, time 213.748 samples/sec
+epoch 0, total_step 4120, total loss is 40.69 , inference loss is 26.54, weight deacy loss is 14.16, training accuracy is 0.031250, time 296.876 samples/sec
+epoch 0, total_step 4140, total loss is 45.83 , inference loss is 31.67, weight deacy loss is 14.16, training accuracy is 0.000000, time 198.025 samples/sec
+epoch 0, total_step 4160, total loss is 49.67 , inference loss is 35.51, weight deacy loss is 14.16, training accuracy is 0.000000, time 202.778 samples/sec
+epoch 0, total_step 4180, total loss is 43.88 , inference loss is 29.73, weight deacy loss is 14.15, training accuracy is 0.031250, time 216.163 samples/sec
+epoch 0, total_step 4200, total loss is 39.21 , inference loss is 25.06, weight deacy loss is 14.15, training accuracy is 0.093750, time 14.133 samples/sec
+epoch 0, total_step 4220, total loss is 45.65 , inference loss is 31.50, weight deacy loss is 14.15, training accuracy is 0.000000, time 224.851 samples/sec
+epoch 0, total_step 4240, total loss is 41.53 , inference loss is 27.38, weight deacy loss is 14.15, training accuracy is 0.062500, time 226.193 samples/sec
+epoch 0, total_step 4260, total loss is 42.28 , inference loss is 28.14, weight deacy loss is 14.15, training accuracy is 0.031250, time 234.638 samples/sec
+epoch 0, total_step 4280, total loss is 43.84 , inference loss is 29.70, weight deacy loss is 14.15, training accuracy is 0.000000, time 226.598 samples/sec
+epoch 0, total_step 4300, total loss is 42.78 , inference loss is 28.64, weight deacy loss is 14.14, training accuracy is 0.000000, time 205.427 samples/sec
+epoch 0, total_step 4320, total loss is 41.45 , inference loss is 27.31, weight deacy loss is 14.14, training accuracy is 0.031250, time 224.803 samples/sec
+epoch 0, total_step 4340, total loss is 47.12 , inference loss is 32.98, weight deacy loss is 14.14, training accuracy is 0.031250, time 233.593 samples/sec
+epoch 0, total_step 4360, total loss is 40.26 , inference loss is 26.12, weight deacy loss is 14.14, training accuracy is 0.000000, time 205.880 samples/sec
+epoch 0, total_step 4380, total loss is 43.57 , inference loss is 29.43, weight deacy loss is 14.14, training accuracy is 0.062500, time 183.881 samples/sec
+epoch 0, total_step 4400, total loss is 43.79 , inference loss is 29.66, weight deacy loss is 14.14, training accuracy is 0.031250, time 196.014 samples/sec
+epoch 0, total_step 4420, total loss is 39.47 , inference loss is 25.33, weight deacy loss is 14.14, training accuracy is 0.000000, time 221.336 samples/sec
+epoch 0, total_step 4440, total loss is 41.98 , inference loss is 27.85, weight deacy loss is 14.13, training accuracy is 0.093750, time 224.512 samples/sec
+epoch 0, total_step 4460, total loss is 44.34 , inference loss is 30.20, weight deacy loss is 14.13, training accuracy is 0.031250, time 213.172 samples/sec
+epoch 0, total_step 4480, total loss is 38.82 , inference loss is 24.69, weight deacy loss is 14.13, training accuracy is 0.093750, time 220.888 samples/sec
+epoch 0, total_step 4500, total loss is 43.31 , inference loss is 29.18, weight deacy loss is 14.13, training accuracy is 0.031250, time 13.786 samples/sec
+epoch 0, total_step 4520, total loss is 43.02 , inference loss is 28.89, weight deacy loss is 14.13, training accuracy is 0.062500, time 225.666 samples/sec
+epoch 0, total_step 4540, total loss is 42.30 , inference loss is 28.18, weight deacy loss is 14.13, training accuracy is 0.125000, time 202.215 samples/sec
+epoch 0, total_step 4560, total loss is 41.37 , inference loss is 27.24, weight deacy loss is 14.13, training accuracy is 0.031250, time 209.784 samples/sec
+epoch 0, total_step 4580, total loss is 42.31 , inference loss is 28.19, weight deacy loss is 14.12, training accuracy is 0.031250, time 215.815 samples/sec
+epoch 0, total_step 4600, total loss is 40.34 , inference loss is 26.22, weight deacy loss is 14.12, training accuracy is 0.062500, time 208.654 samples/sec
+epoch 0, total_step 4620, total loss is 38.83 , inference loss is 24.71, weight deacy loss is 14.12, training accuracy is 0.062500, time 221.953 samples/sec
+epoch 0, total_step 4640, total loss is 45.55 , inference loss is 31.43, weight deacy loss is 14.12, training accuracy is 0.031250, time 227.162 samples/sec
+epoch 0, total_step 4660, total loss is 42.25 , inference loss is 28.14, weight deacy loss is 14.12, training accuracy is 0.031250, time 299.394 samples/sec
+epoch 0, total_step 4680, total loss is 39.96 , inference loss is 25.84, weight deacy loss is 14.12, training accuracy is 0.031250, time 240.752 samples/sec
+epoch 0, total_step 4700, total loss is 41.16 , inference loss is 27.05, weight deacy loss is 14.11, training accuracy is 0.031250, time 218.059 samples/sec
+epoch 0, total_step 4720, total loss is 45.69 , inference loss is 31.58, weight deacy loss is 14.11, training accuracy is 0.031250, time 207.106 samples/sec
+epoch 0, total_step 4740, total loss is 43.53 , inference loss is 29.42, weight deacy loss is 14.11, training accuracy is 0.062500, time 221.839 samples/sec
+epoch 0, total_step 4760, total loss is 39.11 , inference loss is 25.00, weight deacy loss is 14.11, training accuracy is 0.031250, time 233.259 samples/sec
+epoch 0, total_step 4780, total loss is 47.71 , inference loss is 33.60, weight deacy loss is 14.11, training accuracy is 0.000000, time 215.213 samples/sec
+epoch 0, total_step 4800, total loss is 37.75 , inference loss is 23.64, weight deacy loss is 14.11, training accuracy is 0.062500, time 14.606 samples/sec
+epoch 0, total_step 4820, total loss is 44.17 , inference loss is 30.06, weight deacy loss is 14.11, training accuracy is 0.000000, time 222.950 samples/sec
+epoch 0, total_step 4840, total loss is 42.47 , inference loss is 28.36, weight deacy loss is 14.10, training accuracy is 0.062500, time 227.889 samples/sec
+epoch 0, total_step 4860, total loss is 40.97 , inference loss is 26.87, weight deacy loss is 14.10, training accuracy is 0.031250, time 221.765 samples/sec
+epoch 0, total_step 4880, total loss is 43.24 , inference loss is 29.14, weight deacy loss is 14.10, training accuracy is 0.000000, time 209.463 samples/sec
+epoch 0, total_step 4900, total loss is 42.58 , inference loss is 28.48, weight deacy loss is 14.10, training accuracy is 0.000000, time 212.143 samples/sec
+epoch 0, total_step 4920, total loss is 37.33 , inference loss is 23.23, weight deacy loss is 14.10, training accuracy is 0.062500, time 295.078 samples/sec
+epoch 0, total_step 4940, total loss is 40.16 , inference loss is 26.07, weight deacy loss is 14.10, training accuracy is 0.000000, time 215.808 samples/sec
+epoch 0, total_step 4960, total loss is 46.05 , inference loss is 31.95, weight deacy loss is 14.09, training accuracy is 0.000000, time 227.842 samples/sec
+epoch 0, total_step 4980, total loss is 43.89 , inference loss is 29.80, weight deacy loss is 14.09, training accuracy is 0.031250, time 220.430 samples/sec
+epoch 0, total_step 5000, total loss is 40.78 , inference loss is 26.68, weight deacy loss is 14.09, training accuracy is 0.000000, time 220.020 samples/sec
+epoch 0, total_step 5020, total loss is 38.53 , inference loss is 24.44, weight deacy loss is 14.09, training accuracy is 0.000000, time 209.777 samples/sec
+epoch 0, total_step 5040, total loss is 38.18 , inference loss is 24.09, weight deacy loss is 14.09, training accuracy is 0.000000, time 203.837 samples/sec
+epoch 0, total_step 5060, total loss is 42.36 , inference loss is 28.28, weight deacy loss is 14.09, training accuracy is 0.000000, time 212.750 samples/sec
+epoch 0, total_step 5080, total loss is 39.95 , inference loss is 25.87, weight deacy loss is 14.09, training accuracy is 0.000000, time 206.061 samples/sec
+epoch 0, total_step 5100, total loss is 37.79 , inference loss is 23.71, weight deacy loss is 14.08, training accuracy is 0.031250, time 14.185 samples/sec
+epoch 0, total_step 5120, total loss is 40.18 , inference loss is 26.10, weight deacy loss is 14.08, training accuracy is 0.031250, time 229.401 samples/sec
+epoch 0, total_step 5140, total loss is 39.39 , inference loss is 25.31, weight deacy loss is 14.08, training accuracy is 0.031250, time 209.700 samples/sec
+epoch 0, total_step 5160, total loss is 44.20 , inference loss is 30.12, weight deacy loss is 14.08, training accuracy is 0.031250, time 218.358 samples/sec
+epoch 0, total_step 5180, total loss is 38.69 , inference loss is 24.62, weight deacy loss is 14.08, training accuracy is 0.031250, time 298.923 samples/sec
+epoch 0, total_step 5200, total loss is 40.82 , inference loss is 26.75, weight deacy loss is 14.08, training accuracy is 0.031250, time 215.454 samples/sec
+epoch 0, total_step 5220, total loss is 37.15 , inference loss is 23.07, weight deacy loss is 14.07, training accuracy is 0.093750, time 213.797 samples/sec
+epoch 0, total_step 5240, total loss is 39.51 , inference loss is 25.43, weight deacy loss is 14.07, training accuracy is 0.031250, time 212.770 samples/sec
+epoch 0, total_step 5260, total loss is 42.75 , inference loss is 28.68, weight deacy loss is 14.07, training accuracy is 0.093750, time 215.258 samples/sec
+epoch 0, total_step 5280, total loss is 45.13 , inference loss is 31.06, weight deacy loss is 14.07, training accuracy is 0.000000, time 212.453 samples/sec
+epoch 0, total_step 5300, total loss is 39.38 , inference loss is 25.31, weight deacy loss is 14.07, training accuracy is 0.031250, time 214.693 samples/sec
+epoch 0, total_step 5320, total loss is 41.43 , inference loss is 27.36, weight deacy loss is 14.07, training accuracy is 0.093750, time 218.108 samples/sec
+epoch 0, total_step 5340, total loss is 39.35 , inference loss is 25.28, weight deacy loss is 14.07, training accuracy is 0.093750, time 220.084 samples/sec
+epoch 0, total_step 5360, total loss is 42.31 , inference loss is 28.24, weight deacy loss is 14.06, training accuracy is 0.062500, time 295.879 samples/sec
+epoch 0, total_step 5380, total loss is 41.35 , inference loss is 27.29, weight deacy loss is 14.06, training accuracy is 0.000000, time 298.990 samples/sec
+epoch 0, total_step 5400, total loss is 44.59 , inference loss is 30.53, weight deacy loss is 14.06, training accuracy is 0.000000, time 12.528 samples/sec
+epoch 0, total_step 5420, total loss is 38.20 , inference loss is 24.14, weight deacy loss is 14.06, training accuracy is 0.031250, time 219.593 samples/sec
+epoch 0, total_step 5440, total loss is 43.31 , inference loss is 29.25, weight deacy loss is 14.06, training accuracy is 0.093750, time 214.928 samples/sec
+epoch 0, total_step 5460, total loss is 38.39 , inference loss is 24.34, weight deacy loss is 14.06, training accuracy is 0.062500, time 206.321 samples/sec
+epoch 0, total_step 5480, total loss is 40.89 , inference loss is 26.83, weight deacy loss is 14.05, training accuracy is 0.031250, time 222.915 samples/sec
+epoch 0, total_step 5500, total loss is 41.15 , inference loss is 27.10, weight deacy loss is 14.05, training accuracy is 0.031250, time 226.260 samples/sec
+epoch 0, total_step 5520, total loss is 42.91 , inference loss is 28.86, weight deacy loss is 14.05, training accuracy is 0.031250, time 239.511 samples/sec
+epoch 0, total_step 5540, total loss is 44.24 , inference loss is 30.19, weight deacy loss is 14.05, training accuracy is 0.031250, time 206.267 samples/sec
+epoch 0, total_step 5560, total loss is 41.14 , inference loss is 27.10, weight deacy loss is 14.05, training accuracy is 0.031250, time 209.505 samples/sec
+epoch 0, total_step 5580, total loss is 34.79 , inference loss is 20.74, weight deacy loss is 14.05, training accuracy is 0.093750, time 218.110 samples/sec
+epoch 0, total_step 5600, total loss is 41.54 , inference loss is 27.49, weight deacy loss is 14.05, training accuracy is 0.031250, time 223.013 samples/sec
+epoch 0, total_step 5620, total loss is 39.16 , inference loss is 25.12, weight deacy loss is 14.04, training accuracy is 0.031250, time 216.339 samples/sec
+epoch 0, total_step 5640, total loss is 40.50 , inference loss is 26.46, weight deacy loss is 14.04, training accuracy is 0.093750, time 296.288 samples/sec
+epoch 0, total_step 5660, total loss is 35.88 , inference loss is 21.84, weight deacy loss is 14.04, training accuracy is 0.093750, time 206.429 samples/sec
+epoch 0, total_step 5680, total loss is 43.46 , inference loss is 29.42, weight deacy loss is 14.04, training accuracy is 0.000000, time 205.225 samples/sec
+epoch 0, total_step 5700, total loss is 39.13 , inference loss is 25.09, weight deacy loss is 14.04, training accuracy is 0.062500, time 14.092 samples/sec
+epoch 0, total_step 5720, total loss is 37.08 , inference loss is 23.05, weight deacy loss is 14.04, training accuracy is 0.156250, time 220.009 samples/sec
+epoch 0, total_step 5740, total loss is 40.34 , inference loss is 26.30, weight deacy loss is 14.03, training accuracy is 0.031250, time 214.957 samples/sec
+epoch 0, total_step 5760, total loss is 42.75 , inference loss is 28.72, weight deacy loss is 14.03, training accuracy is 0.031250, time 212.549 samples/sec
+epoch 0, total_step 5780, total loss is 41.76 , inference loss is 27.73, weight deacy loss is 14.03, training accuracy is 0.000000, time 208.070 samples/sec
+epoch 0, total_step 5800, total loss is 44.89 , inference loss is 30.86, weight deacy loss is 14.03, training accuracy is 0.000000, time 212.568 samples/sec
+epoch 0, total_step 5820, total loss is 46.20 , inference loss is 32.17, weight deacy loss is 14.03, training accuracy is 0.062500, time 212.274 samples/sec
+epoch 0, total_step 5840, total loss is 41.72 , inference loss is 27.69, weight deacy loss is 14.03, training accuracy is 0.000000, time 205.575 samples/sec
+epoch 0, total_step 5860, total loss is 44.11 , inference loss is 30.09, weight deacy loss is 14.03, training accuracy is 0.062500, time 230.538 samples/sec
+epoch 0, total_step 5880, total loss is 36.02 , inference loss is 21.99, weight deacy loss is 14.02, training accuracy is 0.093750, time 203.219 samples/sec
+epoch 0, total_step 5900, total loss is 42.94 , inference loss is 28.91, weight deacy loss is 14.02, training accuracy is 0.000000, time 297.743 samples/sec
+epoch 0, total_step 5920, total loss is 40.43 , inference loss is 26.41, weight deacy loss is 14.02, training accuracy is 0.062500, time 210.597 samples/sec
+epoch 0, total_step 5940, total loss is 42.04 , inference loss is 28.02, weight deacy loss is 14.02, training accuracy is 0.062500, time 220.860 samples/sec
+epoch 0, total_step 5960, total loss is 38.47 , inference loss is 24.45, weight deacy loss is 14.02, training accuracy is 0.093750, time 238.188 samples/sec
+epoch 0, total_step 5980, total loss is 36.82 , inference loss is 22.81, weight deacy loss is 14.02, training accuracy is 0.031250, time 238.908 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.707121000000004
+best_threshold_index 123 0.9301851851851852
+best_threshold_index 123 0.9285185185185185
+best_threshold_index 123 0.9305555555555556
+best_threshold_index 123 0.9312962962962963
+best_threshold_index 123 0.9346296296296296
+best_threshold_index 123 0.9283333333333333
+best_threshold_index 123 0.9296296296296296
+best_threshold_index 123 0.9311111111111111
+best_threshold_index 123 0.9285185185185185
+best_threshold_index 123 0.9288888888888889
+[lfw][6000]XNorm: 22.421414
+[lfw][6000]Accuracy-Flip: 0.93017+-0.01634
+test accuracy is: 0.9301666666666666
+epoch 0, total_step 6000, total loss is 39.18 , inference loss is 25.17, weight deacy loss is 14.02, training accuracy is 0.031250, time 17.873 samples/sec
+epoch 0, total_step 6020, total loss is 38.36 , inference loss is 24.34, weight deacy loss is 14.01, training accuracy is 0.000000, time 217.374 samples/sec
+epoch 0, total_step 6040, total loss is 41.46 , inference loss is 27.45, weight deacy loss is 14.01, training accuracy is 0.062500, time 219.496 samples/sec
+epoch 0, total_step 6060, total loss is 38.54 , inference loss is 24.53, weight deacy loss is 14.01, training accuracy is 0.031250, time 205.114 samples/sec
+epoch 0, total_step 6080, total loss is 45.05 , inference loss is 31.04, weight deacy loss is 14.01, training accuracy is 0.000000, time 212.164 samples/sec
+epoch 0, total_step 6100, total loss is 41.62 , inference loss is 27.61, weight deacy loss is 14.01, training accuracy is 0.125000, time 214.256 samples/sec
+epoch 0, total_step 6120, total loss is 40.57 , inference loss is 26.57, weight deacy loss is 14.01, training accuracy is 0.000000, time 230.122 samples/sec
+epoch 0, total_step 6140, total loss is 39.94 , inference loss is 25.94, weight deacy loss is 14.01, training accuracy is 0.062500, time 214.281 samples/sec
+epoch 0, total_step 6160, total loss is 38.88 , inference loss is 24.88, weight deacy loss is 14.00, training accuracy is 0.062500, time 228.936 samples/sec
+epoch 0, total_step 6180, total loss is 35.48 , inference loss is 21.48, weight deacy loss is 14.00, training accuracy is 0.062500, time 222.271 samples/sec
+epoch 0, total_step 6200, total loss is 41.25 , inference loss is 27.25, weight deacy loss is 14.00, training accuracy is 0.031250, time 214.419 samples/sec
+epoch 0, total_step 6220, total loss is 36.37 , inference loss is 22.38, weight deacy loss is 14.00, training accuracy is 0.093750, time 217.969 samples/sec
+epoch 0, total_step 6240, total loss is 42.49 , inference loss is 28.49, weight deacy loss is 14.00, training accuracy is 0.062500, time 213.052 samples/sec
+epoch 0, total_step 6260, total loss is 38.78 , inference loss is 24.79, weight deacy loss is 14.00, training accuracy is 0.000000, time 226.295 samples/sec
+epoch 0, total_step 6280, total loss is 41.40 , inference loss is 27.41, weight deacy loss is 13.99, training accuracy is 0.062500, time 299.740 samples/sec
+epoch 0, total_step 6300, total loss is 36.03 , inference loss is 22.04, weight deacy loss is 13.99, training accuracy is 0.031250, time 14.494 samples/sec
+epoch 0, total_step 6320, total loss is 39.95 , inference loss is 25.96, weight deacy loss is 13.99, training accuracy is 0.031250, time 230.482 samples/sec
+epoch 0, total_step 6340, total loss is 48.28 , inference loss is 34.29, weight deacy loss is 13.99, training accuracy is 0.000000, time 231.918 samples/sec
+epoch 0, total_step 6360, total loss is 38.58 , inference loss is 24.59, weight deacy loss is 13.99, training accuracy is 0.000000, time 215.541 samples/sec
+epoch 0, total_step 6380, total loss is 39.61 , inference loss is 25.62, weight deacy loss is 13.99, training accuracy is 0.093750, time 236.565 samples/sec
+epoch 0, total_step 6400, total loss is 39.16 , inference loss is 25.18, weight deacy loss is 13.99, training accuracy is 0.000000, time 216.468 samples/sec
+epoch 0, total_step 6420, total loss is 39.24 , inference loss is 25.25, weight deacy loss is 13.98, training accuracy is 0.062500, time 215.040 samples/sec
+epoch 0, total_step 6440, total loss is 39.75 , inference loss is 25.77, weight deacy loss is 13.98, training accuracy is 0.000000, time 209.036 samples/sec
+epoch 0, total_step 6460, total loss is 38.36 , inference loss is 24.38, weight deacy loss is 13.98, training accuracy is 0.031250, time 224.697 samples/sec
+epoch 0, total_step 6480, total loss is 37.98 , inference loss is 24.00, weight deacy loss is 13.98, training accuracy is 0.062500, time 218.077 samples/sec
+epoch 0, total_step 6500, total loss is 41.50 , inference loss is 27.52, weight deacy loss is 13.98, training accuracy is 0.062500, time 224.861 samples/sec
+epoch 0, total_step 6520, total loss is 46.43 , inference loss is 32.46, weight deacy loss is 13.98, training accuracy is 0.000000, time 226.436 samples/sec
+epoch 0, total_step 6540, total loss is 40.20 , inference loss is 26.22, weight deacy loss is 13.97, training accuracy is 0.031250, time 300.276 samples/sec
+epoch 0, total_step 6560, total loss is 41.10 , inference loss is 27.13, weight deacy loss is 13.97, training accuracy is 0.000000, time 209.611 samples/sec
+epoch 0, total_step 6580, total loss is 41.09 , inference loss is 27.12, weight deacy loss is 13.97, training accuracy is 0.000000, time 211.669 samples/sec
+epoch 0, total_step 6600, total loss is 40.32 , inference loss is 26.35, weight deacy loss is 13.97, training accuracy is 0.031250, time 14.139 samples/sec
+epoch 0, total_step 6620, total loss is 39.42 , inference loss is 25.45, weight deacy loss is 13.97, training accuracy is 0.031250, time 212.100 samples/sec
+epoch 0, total_step 6640, total loss is 35.62 , inference loss is 21.65, weight deacy loss is 13.97, training accuracy is 0.031250, time 226.869 samples/sec
+epoch 0, total_step 6660, total loss is 38.30 , inference loss is 24.33, weight deacy loss is 13.97, training accuracy is 0.062500, time 227.951 samples/sec
+epoch 0, total_step 6680, total loss is 46.84 , inference loss is 32.88, weight deacy loss is 13.96, training accuracy is 0.031250, time 218.861 samples/sec
+epoch 0, total_step 6700, total loss is 37.17 , inference loss is 23.21, weight deacy loss is 13.96, training accuracy is 0.093750, time 218.478 samples/sec
+epoch 0, total_step 6720, total loss is 40.66 , inference loss is 26.70, weight deacy loss is 13.96, training accuracy is 0.062500, time 226.624 samples/sec
+epoch 0, total_step 6740, total loss is 41.87 , inference loss is 27.91, weight deacy loss is 13.96, training accuracy is 0.062500, time 226.302 samples/sec
+epoch 0, total_step 6760, total loss is 37.88 , inference loss is 23.92, weight deacy loss is 13.96, training accuracy is 0.062500, time 222.095 samples/sec
+epoch 0, total_step 6780, total loss is 45.01 , inference loss is 31.06, weight deacy loss is 13.96, training accuracy is 0.031250, time 210.818 samples/sec
+epoch 0, total_step 6800, total loss is 35.82 , inference loss is 21.87, weight deacy loss is 13.96, training accuracy is 0.125000, time 226.305 samples/sec
+epoch 0, total_step 6820, total loss is 41.69 , inference loss is 27.73, weight deacy loss is 13.95, training accuracy is 0.000000, time 295.836 samples/sec
+epoch 0, total_step 6840, total loss is 38.81 , inference loss is 24.86, weight deacy loss is 13.95, training accuracy is 0.062500, time 235.014 samples/sec
+epoch 0, total_step 6860, total loss is 35.92 , inference loss is 21.97, weight deacy loss is 13.95, training accuracy is 0.062500, time 230.830 samples/sec
+epoch 0, total_step 6880, total loss is 43.40 , inference loss is 29.45, weight deacy loss is 13.95, training accuracy is 0.062500, time 213.040 samples/sec
+epoch 0, total_step 6900, total loss is 42.88 , inference loss is 28.94, weight deacy loss is 13.95, training accuracy is 0.031250, time 14.962 samples/sec
+epoch 0, total_step 6920, total loss is 43.53 , inference loss is 29.58, weight deacy loss is 13.95, training accuracy is 0.031250, time 200.684 samples/sec
+epoch 0, total_step 6940, total loss is 43.80 , inference loss is 29.86, weight deacy loss is 13.94, training accuracy is 0.031250, time 220.170 samples/sec
+epoch 0, total_step 6960, total loss is 38.84 , inference loss is 24.90, weight deacy loss is 13.94, training accuracy is 0.000000, time 229.006 samples/sec
+epoch 0, total_step 6980, total loss is 37.27 , inference loss is 23.33, weight deacy loss is 13.94, training accuracy is 0.000000, time 214.967 samples/sec
+epoch 0, total_step 7000, total loss is 41.38 , inference loss is 27.44, weight deacy loss is 13.94, training accuracy is 0.031250, time 214.863 samples/sec
+epoch 0, total_step 7020, total loss is 38.35 , inference loss is 24.41, weight deacy loss is 13.94, training accuracy is 0.000000, time 219.088 samples/sec
+epoch 0, total_step 7040, total loss is 41.59 , inference loss is 27.65, weight deacy loss is 13.94, training accuracy is 0.000000, time 215.101 samples/sec
+epoch 0, total_step 7060, total loss is 42.18 , inference loss is 28.24, weight deacy loss is 13.94, training accuracy is 0.000000, time 211.802 samples/sec
+epoch 0, total_step 7080, total loss is 38.11 , inference loss is 24.17, weight deacy loss is 13.93, training accuracy is 0.062500, time 213.564 samples/sec
+epoch 0, total_step 7100, total loss is 39.86 , inference loss is 25.93, weight deacy loss is 13.93, training accuracy is 0.031250, time 215.417 samples/sec
+epoch 0, total_step 7120, total loss is 37.48 , inference loss is 23.55, weight deacy loss is 13.93, training accuracy is 0.062500, time 207.270 samples/sec
+epoch 0, total_step 7140, total loss is 38.86 , inference loss is 24.93, weight deacy loss is 13.93, training accuracy is 0.000000, time 296.833 samples/sec
+epoch 0, total_step 7160, total loss is 36.69 , inference loss is 22.76, weight deacy loss is 13.93, training accuracy is 0.000000, time 184.154 samples/sec
+epoch 0, total_step 7180, total loss is 37.95 , inference loss is 24.02, weight deacy loss is 13.93, training accuracy is 0.125000, time 207.785 samples/sec
+epoch 0, total_step 7200, total loss is 37.53 , inference loss is 23.61, weight deacy loss is 13.93, training accuracy is 0.000000, time 14.409 samples/sec
+epoch 0, total_step 7220, total loss is 41.28 , inference loss is 27.36, weight deacy loss is 13.92, training accuracy is 0.031250, time 297.364 samples/sec
+epoch 0, total_step 7240, total loss is 40.45 , inference loss is 26.53, weight deacy loss is 13.92, training accuracy is 0.031250, time 225.403 samples/sec
+epoch 0, total_step 7260, total loss is 39.05 , inference loss is 25.13, weight deacy loss is 13.92, training accuracy is 0.031250, time 215.956 samples/sec
+epoch 0, total_step 7280, total loss is 35.46 , inference loss is 21.54, weight deacy loss is 13.92, training accuracy is 0.031250, time 220.105 samples/sec
+epoch 0, total_step 7300, total loss is 36.11 , inference loss is 22.19, weight deacy loss is 13.92, training accuracy is 0.031250, time 226.563 samples/sec
+epoch 0, total_step 7320, total loss is 34.12 , inference loss is 20.20, weight deacy loss is 13.92, training accuracy is 0.031250, time 223.895 samples/sec
+epoch 0, total_step 7340, total loss is 35.87 , inference loss is 21.96, weight deacy loss is 13.92, training accuracy is 0.000000, time 216.301 samples/sec
+epoch 0, total_step 7360, total loss is 39.92 , inference loss is 26.01, weight deacy loss is 13.91, training accuracy is 0.000000, time 215.860 samples/sec
+epoch 0, total_step 7380, total loss is 37.65 , inference loss is 23.74, weight deacy loss is 13.91, training accuracy is 0.031250, time 225.527 samples/sec
+epoch 0, total_step 7400, total loss is 41.86 , inference loss is 27.95, weight deacy loss is 13.91, training accuracy is 0.000000, time 210.637 samples/sec
+epoch 0, total_step 7420, total loss is 43.09 , inference loss is 29.18, weight deacy loss is 13.91, training accuracy is 0.031250, time 241.864 samples/sec
+epoch 0, total_step 7440, total loss is 41.13 , inference loss is 27.22, weight deacy loss is 13.91, training accuracy is 0.031250, time 204.617 samples/sec
+epoch 0, total_step 7460, total loss is 37.48 , inference loss is 23.57, weight deacy loss is 13.91, training accuracy is 0.031250, time 199.403 samples/sec
+epoch 0, total_step 7480, total loss is 38.69 , inference loss is 24.79, weight deacy loss is 13.90, training accuracy is 0.062500, time 210.213 samples/sec
+epoch 0, total_step 7500, total loss is 40.64 , inference loss is 26.73, weight deacy loss is 13.90, training accuracy is 0.093750, time 15.042 samples/sec
+epoch 0, total_step 7520, total loss is 35.43 , inference loss is 21.53, weight deacy loss is 13.90, training accuracy is 0.218750, time 225.814 samples/sec
+epoch 0, total_step 7540, total loss is 37.59 , inference loss is 23.69, weight deacy loss is 13.90, training accuracy is 0.000000, time 209.282 samples/sec
+epoch 0, total_step 7560, total loss is 40.99 , inference loss is 27.09, weight deacy loss is 13.90, training accuracy is 0.031250, time 225.230 samples/sec
+epoch 0, total_step 7580, total loss is 35.87 , inference loss is 21.98, weight deacy loss is 13.90, training accuracy is 0.093750, time 205.609 samples/sec
+epoch 0, total_step 7600, total loss is 37.73 , inference loss is 23.83, weight deacy loss is 13.90, training accuracy is 0.062500, time 207.781 samples/sec
+epoch 0, total_step 7620, total loss is 39.77 , inference loss is 25.88, weight deacy loss is 13.89, training accuracy is 0.062500, time 230.291 samples/sec
+epoch 0, total_step 7640, total loss is 36.73 , inference loss is 22.84, weight deacy loss is 13.89, training accuracy is 0.062500, time 223.949 samples/sec
+epoch 0, total_step 7660, total loss is 39.34 , inference loss is 25.45, weight deacy loss is 13.89, training accuracy is 0.062500, time 209.006 samples/sec
+epoch 0, total_step 7680, total loss is 37.25 , inference loss is 23.36, weight deacy loss is 13.89, training accuracy is 0.093750, time 210.645 samples/sec
+epoch 0, total_step 7700, total loss is 32.33 , inference loss is 18.44, weight deacy loss is 13.89, training accuracy is 0.187500, time 218.225 samples/sec
+epoch 0, total_step 7720, total loss is 34.17 , inference loss is 20.28, weight deacy loss is 13.89, training accuracy is 0.062500, time 210.075 samples/sec
+epoch 0, total_step 7740, total loss is 37.77 , inference loss is 23.88, weight deacy loss is 13.88, training accuracy is 0.031250, time 236.036 samples/sec
+epoch 0, total_step 7760, total loss is 43.09 , inference loss is 29.21, weight deacy loss is 13.88, training accuracy is 0.031250, time 226.000 samples/sec
+epoch 0, total_step 7780, total loss is 36.18 , inference loss is 22.30, weight deacy loss is 13.88, training accuracy is 0.062500, time 194.750 samples/sec
+epoch 0, total_step 7800, total loss is 34.70 , inference loss is 20.82, weight deacy loss is 13.88, training accuracy is 0.093750, time 14.345 samples/sec
+epoch 0, total_step 7820, total loss is 39.51 , inference loss is 25.63, weight deacy loss is 13.88, training accuracy is 0.031250, time 234.641 samples/sec
+epoch 0, total_step 7840, total loss is 33.09 , inference loss is 19.21, weight deacy loss is 13.88, training accuracy is 0.093750, time 242.113 samples/sec
+epoch 0, total_step 7860, total loss is 38.45 , inference loss is 24.58, weight deacy loss is 13.88, training accuracy is 0.093750, time 215.093 samples/sec
+epoch 0, total_step 7880, total loss is 38.69 , inference loss is 24.82, weight deacy loss is 13.87, training accuracy is 0.062500, time 205.325 samples/sec
+epoch 0, total_step 7900, total loss is 38.39 , inference loss is 24.52, weight deacy loss is 13.87, training accuracy is 0.031250, time 215.247 samples/sec
+epoch 0, total_step 7920, total loss is 36.82 , inference loss is 22.95, weight deacy loss is 13.87, training accuracy is 0.031250, time 216.515 samples/sec
+epoch 0, total_step 7940, total loss is 39.11 , inference loss is 25.24, weight deacy loss is 13.87, training accuracy is 0.000000, time 229.758 samples/sec
+epoch 0, total_step 7960, total loss is 39.96 , inference loss is 26.10, weight deacy loss is 13.87, training accuracy is 0.000000, time 212.798 samples/sec
+epoch 0, total_step 7980, total loss is 46.40 , inference loss is 32.54, weight deacy loss is 13.87, training accuracy is 0.000000, time 222.117 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.873974000000004
+best_threshold_index 118 0.9468518518518518
+best_threshold_index 118 0.947037037037037
+best_threshold_index 118 0.9481481481481482
+best_threshold_index 118 0.945925925925926
+best_threshold_index 118 0.9503703703703704
+best_threshold_index 118 0.9457407407407408
+best_threshold_index 118 0.945925925925926
+best_threshold_index 118 0.9477777777777778
+best_threshold_index 118 0.9455555555555556
+best_threshold_index 118 0.945
+[lfw][8000]XNorm: 23.696459
+[lfw][8000]Accuracy-Flip: 0.94683+-0.01365
+test accuracy is: 0.9468333333333334
+epoch 0, total_step 8000, total loss is 42.39 , inference loss is 28.52, weight deacy loss is 13.87, training accuracy is 0.000000, time 17.550 samples/sec
+epoch 0, total_step 8020, total loss is 36.85 , inference loss is 22.99, weight deacy loss is 13.86, training accuracy is 0.156250, time 221.143 samples/sec
+epoch 0, total_step 8040, total loss is 32.61 , inference loss is 18.75, weight deacy loss is 13.86, training accuracy is 0.031250, time 201.256 samples/sec
+epoch 0, total_step 8060, total loss is 39.57 , inference loss is 25.71, weight deacy loss is 13.86, training accuracy is 0.031250, time 204.625 samples/sec
+epoch 0, total_step 8080, total loss is 36.38 , inference loss is 22.52, weight deacy loss is 13.86, training accuracy is 0.031250, time 218.427 samples/sec
+epoch 0, total_step 8100, total loss is 43.42 , inference loss is 29.56, weight deacy loss is 13.86, training accuracy is 0.000000, time 14.277 samples/sec
+epoch 0, total_step 8120, total loss is 41.32 , inference loss is 27.47, weight deacy loss is 13.86, training accuracy is 0.093750, time 199.239 samples/sec
+epoch 0, total_step 8140, total loss is 39.74 , inference loss is 25.88, weight deacy loss is 13.85, training accuracy is 0.156250, time 297.835 samples/sec
+epoch 0, total_step 8160, total loss is 38.28 , inference loss is 24.43, weight deacy loss is 13.85, training accuracy is 0.000000, time 227.528 samples/sec
+epoch 0, total_step 8180, total loss is 41.59 , inference loss is 27.74, weight deacy loss is 13.85, training accuracy is 0.093750, time 216.746 samples/sec
+epoch 0, total_step 8200, total loss is 40.40 , inference loss is 26.55, weight deacy loss is 13.85, training accuracy is 0.062500, time 219.845 samples/sec
+epoch 0, total_step 8220, total loss is 34.60 , inference loss is 20.75, weight deacy loss is 13.85, training accuracy is 0.062500, time 202.326 samples/sec
+epoch 0, total_step 8240, total loss is 40.05 , inference loss is 26.20, weight deacy loss is 13.85, training accuracy is 0.000000, time 205.917 samples/sec
+epoch 0, total_step 8260, total loss is 39.71 , inference loss is 25.87, weight deacy loss is 13.85, training accuracy is 0.062500, time 213.665 samples/sec
+epoch 0, total_step 8280, total loss is 37.95 , inference loss is 24.11, weight deacy loss is 13.84, training accuracy is 0.000000, time 239.015 samples/sec
+epoch 0, total_step 8300, total loss is 43.60 , inference loss is 29.76, weight deacy loss is 13.84, training accuracy is 0.000000, time 221.157 samples/sec
+epoch 0, total_step 8320, total loss is 42.61 , inference loss is 28.77, weight deacy loss is 13.84, training accuracy is 0.000000, time 224.814 samples/sec
+epoch 0, total_step 8340, total loss is 45.43 , inference loss is 31.59, weight deacy loss is 13.84, training accuracy is 0.031250, time 240.286 samples/sec
+epoch 0, total_step 8360, total loss is 37.55 , inference loss is 23.71, weight deacy loss is 13.84, training accuracy is 0.031250, time 210.923 samples/sec
+epoch 0, total_step 8380, total loss is 40.44 , inference loss is 26.61, weight deacy loss is 13.84, training accuracy is 0.031250, time 204.297 samples/sec
+epoch 0, total_step 8400, total loss is 34.55 , inference loss is 20.72, weight deacy loss is 13.84, training accuracy is 0.093750, time 14.586 samples/sec
+epoch 0, total_step 8420, total loss is 37.87 , inference loss is 24.04, weight deacy loss is 13.83, training accuracy is 0.062500, time 209.571 samples/sec
+epoch 0, total_step 8440, total loss is 39.62 , inference loss is 25.79, weight deacy loss is 13.83, training accuracy is 0.062500, time 216.791 samples/sec
+epoch 0, total_step 8460, total loss is 42.47 , inference loss is 28.64, weight deacy loss is 13.83, training accuracy is 0.031250, time 195.267 samples/sec
+epoch 0, total_step 8480, total loss is 36.91 , inference loss is 23.08, weight deacy loss is 13.83, training accuracy is 0.093750, time 226.194 samples/sec
+epoch 0, total_step 8500, total loss is 43.91 , inference loss is 30.08, weight deacy loss is 13.83, training accuracy is 0.000000, time 222.834 samples/sec
+epoch 0, total_step 8520, total loss is 37.08 , inference loss is 23.25, weight deacy loss is 13.83, training accuracy is 0.031250, time 218.287 samples/sec
+epoch 0, total_step 8540, total loss is 38.76 , inference loss is 24.93, weight deacy loss is 13.83, training accuracy is 0.093750, time 220.167 samples/sec
+epoch 0, total_step 8560, total loss is 34.81 , inference loss is 20.99, weight deacy loss is 13.82, training accuracy is 0.062500, time 220.903 samples/sec
+epoch 0, total_step 8580, total loss is 36.08 , inference loss is 22.26, weight deacy loss is 13.82, training accuracy is 0.000000, time 206.210 samples/sec
+epoch 0, total_step 8600, total loss is 34.32 , inference loss is 20.50, weight deacy loss is 13.82, training accuracy is 0.031250, time 226.966 samples/sec
+epoch 0, total_step 8620, total loss is 40.60 , inference loss is 26.78, weight deacy loss is 13.82, training accuracy is 0.093750, time 205.385 samples/sec
+epoch 0, total_step 8640, total loss is 35.78 , inference loss is 21.96, weight deacy loss is 13.82, training accuracy is 0.125000, time 212.787 samples/sec
+epoch 0, total_step 8660, total loss is 35.51 , inference loss is 21.69, weight deacy loss is 13.82, training accuracy is 0.093750, time 206.877 samples/sec
+epoch 0, total_step 8680, total loss is 39.30 , inference loss is 25.49, weight deacy loss is 13.81, training accuracy is 0.031250, time 212.629 samples/sec
+epoch 0, total_step 8700, total loss is 35.33 , inference loss is 21.52, weight deacy loss is 13.81, training accuracy is 0.031250, time 11.197 samples/sec
+epoch 0, total_step 8720, total loss is 37.98 , inference loss is 24.17, weight deacy loss is 13.81, training accuracy is 0.093750, time 216.084 samples/sec
+epoch 0, total_step 8740, total loss is 39.50 , inference loss is 25.69, weight deacy loss is 13.81, training accuracy is 0.000000, time 234.071 samples/sec
+epoch 0, total_step 8760, total loss is 42.16 , inference loss is 28.35, weight deacy loss is 13.81, training accuracy is 0.000000, time 222.676 samples/sec
+epoch 0, total_step 8780, total loss is 42.88 , inference loss is 29.07, weight deacy loss is 13.81, training accuracy is 0.000000, time 218.179 samples/sec
+epoch 0, total_step 8800, total loss is 36.88 , inference loss is 23.07, weight deacy loss is 13.81, training accuracy is 0.062500, time 216.198 samples/sec
+epoch 0, total_step 8820, total loss is 37.37 , inference loss is 23.57, weight deacy loss is 13.80, training accuracy is 0.062500, time 212.242 samples/sec
+epoch 0, total_step 8840, total loss is 35.60 , inference loss is 21.80, weight deacy loss is 13.80, training accuracy is 0.093750, time 220.005 samples/sec
+epoch 0, total_step 8860, total loss is 41.29 , inference loss is 27.49, weight deacy loss is 13.80, training accuracy is 0.093750, time 213.904 samples/sec
+epoch 0, total_step 8880, total loss is 40.78 , inference loss is 26.98, weight deacy loss is 13.80, training accuracy is 0.000000, time 216.909 samples/sec
+epoch 0, total_step 8900, total loss is 34.79 , inference loss is 20.99, weight deacy loss is 13.80, training accuracy is 0.062500, time 225.944 samples/sec
+epoch 0, total_step 8920, total loss is 36.73 , inference loss is 22.93, weight deacy loss is 13.80, training accuracy is 0.000000, time 297.888 samples/sec
+epoch 0, total_step 8940, total loss is 44.47 , inference loss is 30.67, weight deacy loss is 13.80, training accuracy is 0.031250, time 281.763 samples/sec
+epoch 0, total_step 8960, total loss is 34.84 , inference loss is 21.04, weight deacy loss is 13.79, training accuracy is 0.062500, time 225.300 samples/sec
+epoch 0, total_step 8980, total loss is 40.30 , inference loss is 26.51, weight deacy loss is 13.79, training accuracy is 0.000000, time 214.959 samples/sec
+epoch 0, total_step 9000, total loss is 38.29 , inference loss is 24.49, weight deacy loss is 13.79, training accuracy is 0.062500, time 11.738 samples/sec
+epoch 0, total_step 9020, total loss is 41.26 , inference loss is 27.47, weight deacy loss is 13.79, training accuracy is 0.000000, time 242.281 samples/sec
+epoch 0, total_step 9040, total loss is 39.46 , inference loss is 25.67, weight deacy loss is 13.79, training accuracy is 0.031250, time 226.454 samples/sec
+epoch 0, total_step 9060, total loss is 35.16 , inference loss is 21.37, weight deacy loss is 13.79, training accuracy is 0.031250, time 216.958 samples/sec
+epoch 0, total_step 9080, total loss is 34.47 , inference loss is 20.68, weight deacy loss is 13.79, training accuracy is 0.093750, time 224.427 samples/sec
+epoch 0, total_step 9100, total loss is 34.81 , inference loss is 21.02, weight deacy loss is 13.78, training accuracy is 0.062500, time 299.103 samples/sec
+epoch 0, total_step 9120, total loss is 42.70 , inference loss is 28.92, weight deacy loss is 13.78, training accuracy is 0.031250, time 211.688 samples/sec
+epoch 0, total_step 9140, total loss is 34.97 , inference loss is 21.19, weight deacy loss is 13.78, training accuracy is 0.031250, time 224.761 samples/sec
+epoch 0, total_step 9160, total loss is 37.53 , inference loss is 23.75, weight deacy loss is 13.78, training accuracy is 0.062500, time 211.513 samples/sec
+epoch 0, total_step 9180, total loss is 33.54 , inference loss is 19.76, weight deacy loss is 13.78, training accuracy is 0.062500, time 204.918 samples/sec
+epoch 0, total_step 9200, total loss is 37.12 , inference loss is 23.34, weight deacy loss is 13.78, training accuracy is 0.031250, time 209.608 samples/sec
+epoch 0, total_step 9220, total loss is 38.57 , inference loss is 24.79, weight deacy loss is 13.77, training accuracy is 0.062500, time 226.599 samples/sec2022-12-01 00:52:13.052448: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:13.053215: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:13.053284: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:13.053294: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:52:13.053299: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:52:14.511791: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:14.511874: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:14.511885: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:52:14.511890: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:52:14.528863: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:14.528937: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:14.528978: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:14.528987: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:52:14.528991: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:52:14.529087: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:14.529121: W tf_adapter/util/npu_attrs.cc:1077] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:14.529126: W tf_adapter/util/npu_attrs.cc:1207] The op_select_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:52:14.529130: W tf_adapter/util/npu_attrs.cc:1211] The optypelist_for_implmode option IS DEPRECATED. It will be removed in a future version. Please use op_precision_mode instead
+2022-12-01 00:52:14.529177: W tf_adapter/util/npu_attrs.cc:861] The enable_data_pre_proc option IS DEPRECATED. It will be removed in a future version.
+2022-12-01 00:52:15.411379: I tf_adapter/kernels/geop_npu.cc:854] The model has been compiled on the Ascend AI processor, current graph id is: 41
+
+epoch 0, total_step 9240, total loss is 42.11 , inference loss is 28.34, weight deacy loss is 13.77, training accuracy is 0.031250, time 207.626 samples/sec
+epoch 0, total_step 9260, total loss is 37.41 , inference loss is 23.64, weight deacy loss is 13.77, training accuracy is 0.000000, time 199.227 samples/sec
+epoch 0, total_step 9280, total loss is 35.69 , inference loss is 21.92, weight deacy loss is 13.77, training accuracy is 0.031250, time 219.659 samples/sec
+epoch 0, total_step 9300, total loss is 39.79 , inference loss is 26.02, weight deacy loss is 13.77, training accuracy is 0.031250, time 11.371 samples/sec
+epoch 0, total_step 9320, total loss is 42.09 , inference loss is 28.32, weight deacy loss is 13.77, training accuracy is 0.031250, time 212.577 samples/sec
+epoch 0, total_step 9340, total loss is 37.46 , inference loss is 23.69, weight deacy loss is 13.77, training accuracy is 0.031250, time 211.308 samples/sec
+epoch 0, total_step 9360, total loss is 41.62 , inference loss is 27.86, weight deacy loss is 13.76, training accuracy is 0.062500, time 299.256 samples/sec
+epoch 0, total_step 9380, total loss is 35.53 , inference loss is 21.77, weight deacy loss is 13.76, training accuracy is 0.031250, time 207.200 samples/sec
+epoch 0, total_step 9400, total loss is 37.20 , inference loss is 23.44, weight deacy loss is 13.76, training accuracy is 0.000000, time 238.390 samples/sec
+epoch 0, total_step 9420, total loss is 38.68 , inference loss is 24.92, weight deacy loss is 13.76, training accuracy is 0.093750, time 230.075 samples/sec
+epoch 0, total_step 9440, total loss is 34.62 , inference loss is 20.86, weight deacy loss is 13.76, training accuracy is 0.062500, time 229.152 samples/sec
+epoch 0, total_step 9460, total loss is 42.14 , inference loss is 28.39, weight deacy loss is 13.76, training accuracy is 0.000000, time 204.897 samples/sec
+epoch 0, total_step 9480, total loss is 37.48 , inference loss is 23.72, weight deacy loss is 13.76, training accuracy is 0.031250, time 226.051 samples/sec
+epoch 0, total_step 9500, total loss is 40.58 , inference loss is 26.82, weight deacy loss is 13.75, training accuracy is 0.062500, time 194.645 samples/sec
+epoch 0, total_step 9520, total loss is 42.56 , inference loss is 28.81, weight deacy loss is 13.75, training accuracy is 0.062500, time 225.028 samples/sec
+epoch 0, total_step 9540, total loss is 35.23 , inference loss is 21.48, weight deacy loss is 13.75, training accuracy is 0.062500, time 209.854 samples/sec
+epoch 0, total_step 9560, total loss is 37.77 , inference loss is 24.02, weight deacy loss is 13.75, training accuracy is 0.000000, time 206.543 samples/sec
+epoch 0, total_step 9580, total loss is 35.84 , inference loss is 22.10, weight deacy loss is 13.75, training accuracy is 0.031250, time 215.782 samples/sec
+epoch 0, total_step 9600, total loss is 41.05 , inference loss is 27.30, weight deacy loss is 13.75, training accuracy is 0.031250, time 14.040 samples/sec
+epoch 0, total_step 9620, total loss is 38.81 , inference loss is 25.07, weight deacy loss is 13.75, training accuracy is 0.031250, time 216.991 samples/sec
+epoch 0, total_step 9640, total loss is 37.39 , inference loss is 23.65, weight deacy loss is 13.74, training accuracy is 0.062500, time 299.394 samples/sec
+epoch 0, total_step 9660, total loss is 41.57 , inference loss is 27.83, weight deacy loss is 13.74, training accuracy is 0.031250, time 239.898 samples/sec
+epoch 0, total_step 9680, total loss is 42.13 , inference loss is 28.39, weight deacy loss is 13.74, training accuracy is 0.000000, time 224.964 samples/sec
+epoch 0, total_step 9700, total loss is 34.35 , inference loss is 20.61, weight deacy loss is 13.74, training accuracy is 0.062500, time 217.595 samples/sec
+epoch 0, total_step 9720, total loss is 37.81 , inference loss is 24.07, weight deacy loss is 13.74, training accuracy is 0.031250, time 215.046 samples/sec
+epoch 0, total_step 9740, total loss is 38.09 , inference loss is 24.36, weight deacy loss is 13.74, training accuracy is 0.125000, time 207.102 samples/sec
+epoch 0, total_step 9760, total loss is 36.76 , inference loss is 23.03, weight deacy loss is 13.73, training accuracy is 0.031250, time 195.709 samples/sec
+epoch 0, total_step 9780, total loss is 40.62 , inference loss is 26.89, weight deacy loss is 13.73, training accuracy is 0.031250, time 224.446 samples/sec
+epoch 0, total_step 9800, total loss is 40.01 , inference loss is 26.27, weight deacy loss is 13.73, training accuracy is 0.000000, time 207.848 samples/sec
+epoch 0, total_step 9820, total loss is 37.18 , inference loss is 23.45, weight deacy loss is 13.73, training accuracy is 0.000000, time 235.492 samples/sec
+epoch 0, total_step 9840, total loss is 38.27 , inference loss is 24.54, weight deacy loss is 13.73, training accuracy is 0.000000, time 212.637 samples/sec
+epoch 0, total_step 9860, total loss is 35.82 , inference loss is 22.09, weight deacy loss is 13.73, training accuracy is 0.062500, time 226.976 samples/sec
+epoch 0, total_step 9880, total loss is 38.34 , inference loss is 24.62, weight deacy loss is 13.73, training accuracy is 0.093750, time 215.915 samples/sec
+epoch 0, total_step 9900, total loss is 36.93 , inference loss is 23.21, weight deacy loss is 13.72, training accuracy is 0.093750, time 11.305 samples/sec
+epoch 0, total_step 9920, total loss is 39.27 , inference loss is 25.55, weight deacy loss is 13.72, training accuracy is 0.031250, time 227.270 samples/sec
+epoch 0, total_step 9940, total loss is 35.47 , inference loss is 21.75, weight deacy loss is 13.72, training accuracy is 0.031250, time 236.278 samples/sec
+epoch 0, total_step 9960, total loss is 42.21 , inference loss is 28.49, weight deacy loss is 13.72, training accuracy is 0.031250, time 220.243 samples/sec
+epoch 0, total_step 9980, total loss is 34.77 , inference loss is 21.05, weight deacy loss is 13.72, training accuracy is 0.125000, time 225.928 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.616218000000007
+best_threshold_index 102 0.9398148148148148
+best_threshold_index 103 0.9409259259259259
+best_threshold_index 102 0.942962962962963
+best_threshold_index 101 0.9414814814814815
+best_threshold_index 101 0.9425925925925925
+best_threshold_index 100 0.9383333333333334
+best_threshold_index 100 0.94
+best_threshold_index 102 0.9411111111111111
+best_threshold_index 103 0.9390740740740741
+best_threshold_index 102 0.9394444444444444
+[lfw][10000]XNorm: 24.284825
+[lfw][10000]Accuracy-Flip: 0.93883+-0.01263
+test accuracy is: 0.9388333333333332
+epoch 0, total_step 10000, total loss is 39.15 , inference loss is 25.43, weight deacy loss is 13.72, training accuracy is 0.093750, time 16.503 samples/sec
+epoch 0, total_step 10020, total loss is 31.71 , inference loss is 18.00, weight deacy loss is 13.72, training accuracy is 0.062500, time 213.531 samples/sec
+epoch 0, total_step 10040, total loss is 37.78 , inference loss is 24.07, weight deacy loss is 13.71, training accuracy is 0.000000, time 228.629 samples/sec
+epoch 0, total_step 10060, total loss is 38.33 , inference loss is 24.62, weight deacy loss is 13.71, training accuracy is 0.000000, time 218.386 samples/sec
+epoch 0, total_step 10080, total loss is 43.24 , inference loss is 29.53, weight deacy loss is 13.71, training accuracy is 0.031250, time 217.283 samples/sec
+epoch 0, total_step 10100, total loss is 37.62 , inference loss is 23.91, weight deacy loss is 13.71, training accuracy is 0.031250, time 244.365 samples/sec
+epoch 0, total_step 10120, total loss is 33.45 , inference loss is 19.74, weight deacy loss is 13.71, training accuracy is 0.062500, time 213.513 samples/sec
+epoch 0, total_step 10140, total loss is 37.50 , inference loss is 23.80, weight deacy loss is 13.71, training accuracy is 0.000000, time 294.478 samples/sec
+epoch 0, total_step 10160, total loss is 41.03 , inference loss is 27.32, weight deacy loss is 13.70, training accuracy is 0.000000, time 221.984 samples/sec
+epoch 0, total_step 10180, total loss is 36.06 , inference loss is 22.35, weight deacy loss is 13.70, training accuracy is 0.062500, time 238.556 samples/sec
+epoch 0, total_step 10200, total loss is 35.43 , inference loss is 21.73, weight deacy loss is 13.70, training accuracy is 0.031250, time 13.604 samples/sec
+epoch 0, total_step 10220, total loss is 33.77 , inference loss is 20.06, weight deacy loss is 13.70, training accuracy is 0.093750, time 237.188 samples/sec
+epoch 0, total_step 10240, total loss is 34.97 , inference loss is 21.27, weight deacy loss is 13.70, training accuracy is 0.093750, time 219.456 samples/sec
+epoch 0, total_step 10260, total loss is 36.18 , inference loss is 22.48, weight deacy loss is 13.70, training accuracy is 0.000000, time 210.050 samples/sec
+epoch 0, total_step 10280, total loss is 41.15 , inference loss is 27.46, weight deacy loss is 13.70, training accuracy is 0.031250, time 236.385 samples/sec
+epoch 0, total_step 10300, total loss is 39.04 , inference loss is 25.35, weight deacy loss is 13.69, training accuracy is 0.062500, time 225.125 samples/sec
+epoch 0, total_step 10320, total loss is 37.30 , inference loss is 23.61, weight deacy loss is 13.69, training accuracy is 0.031250, time 209.558 samples/sec
+epoch 0, total_step 10340, total loss is 37.65 , inference loss is 23.96, weight deacy loss is 13.69, training accuracy is 0.031250, time 224.924 samples/sec
+epoch 0, total_step 10360, total loss is 36.80 , inference loss is 23.11, weight deacy loss is 13.69, training accuracy is 0.031250, time 208.361 samples/sec
+epoch 0, total_step 10380, total loss is 41.40 , inference loss is 27.71, weight deacy loss is 13.69, training accuracy is 0.000000, time 234.784 samples/sec
+epoch 0, total_step 10400, total loss is 33.36 , inference loss is 19.68, weight deacy loss is 13.69, training accuracy is 0.093750, time 297.439 samples/sec
+epoch 0, total_step 10420, total loss is 37.79 , inference loss is 24.10, weight deacy loss is 13.69, training accuracy is 0.093750, time 216.656 samples/sec
+epoch 0, total_step 10440, total loss is 38.64 , inference loss is 24.96, weight deacy loss is 13.68, training accuracy is 0.000000, time 205.792 samples/sec
+epoch 0, total_step 10460, total loss is 34.04 , inference loss is 20.36, weight deacy loss is 13.68, training accuracy is 0.093750, time 218.365 samples/sec
+epoch 0, total_step 10480, total loss is 37.53 , inference loss is 23.85, weight deacy loss is 13.68, training accuracy is 0.031250, time 221.370 samples/sec
+epoch 0, total_step 10500, total loss is 36.63 , inference loss is 22.95, weight deacy loss is 13.68, training accuracy is 0.000000, time 12.344 samples/sec
+epoch 0, total_step 10520, total loss is 38.66 , inference loss is 24.98, weight deacy loss is 13.68, training accuracy is 0.031250, time 299.340 samples/sec
+epoch 0, total_step 10540, total loss is 30.89 , inference loss is 17.21, weight deacy loss is 13.68, training accuracy is 0.125000, time 297.041 samples/sec
+epoch 0, total_step 10560, total loss is 34.06 , inference loss is 20.38, weight deacy loss is 13.67, training accuracy is 0.125000, time 211.818 samples/sec
+epoch 0, total_step 10580, total loss is 35.84 , inference loss is 22.16, weight deacy loss is 13.67, training accuracy is 0.031250, time 207.694 samples/sec
+epoch 0, total_step 10600, total loss is 41.68 , inference loss is 28.01, weight deacy loss is 13.67, training accuracy is 0.062500, time 224.789 samples/sec
+epoch 0, total_step 10620, total loss is 33.83 , inference loss is 20.16, weight deacy loss is 13.67, training accuracy is 0.093750, time 194.790 samples/sec
+epoch 0, total_step 10640, total loss is 33.61 , inference loss is 19.94, weight deacy loss is 13.67, training accuracy is 0.093750, time 216.009 samples/sec
+epoch 0, total_step 10660, total loss is 42.60 , inference loss is 28.94, weight deacy loss is 13.67, training accuracy is 0.093750, time 215.061 samples/sec
+epoch 0, total_step 10680, total loss is 36.87 , inference loss is 23.20, weight deacy loss is 13.67, training accuracy is 0.031250, time 205.466 samples/sec
+epoch 0, total_step 10700, total loss is 37.79 , inference loss is 24.13, weight deacy loss is 13.66, training accuracy is 0.093750, time 218.989 samples/sec
+epoch 0, total_step 10720, total loss is 35.41 , inference loss is 21.74, weight deacy loss is 13.66, training accuracy is 0.031250, time 213.297 samples/sec
+epoch 0, total_step 10740, total loss is 37.41 , inference loss is 23.75, weight deacy loss is 13.66, training accuracy is 0.093750, time 204.586 samples/sec
+epoch 0, total_step 10760, total loss is 35.46 , inference loss is 21.80, weight deacy loss is 13.66, training accuracy is 0.062500, time 212.932 samples/sec
+epoch 0, total_step 10780, total loss is 39.79 , inference loss is 26.13, weight deacy loss is 13.66, training accuracy is 0.062500, time 224.739 samples/sec
+epoch 0, total_step 10800, total loss is 42.03 , inference loss is 28.37, weight deacy loss is 13.66, training accuracy is 0.031250, time 13.269 samples/sec
+epoch 0, total_step 10820, total loss is 35.47 , inference loss is 21.81, weight deacy loss is 13.66, training accuracy is 0.093750, time 204.054 samples/sec
+epoch 0, total_step 10840, total loss is 39.20 , inference loss is 25.55, weight deacy loss is 13.66, training accuracy is 0.031250, time 221.419 samples/sec
+epoch 0, total_step 10860, total loss is 35.59 , inference loss is 21.93, weight deacy loss is 13.65, training accuracy is 0.031250, time 220.813 samples/sec
+epoch 0, total_step 10880, total loss is 40.16 , inference loss is 26.51, weight deacy loss is 13.65, training accuracy is 0.031250, time 210.773 samples/sec
+epoch 0, total_step 10900, total loss is 37.47 , inference loss is 23.81, weight deacy loss is 13.65, training accuracy is 0.125000, time 210.543 samples/sec
+epoch 0, total_step 10920, total loss is 37.50 , inference loss is 23.85, weight deacy loss is 13.65, training accuracy is 0.031250, time 221.759 samples/sec
+epoch 0, total_step 10940, total loss is 35.48 , inference loss is 21.84, weight deacy loss is 13.65, training accuracy is 0.062500, time 234.893 samples/sec
+epoch 0, total_step 10960, total loss is 37.98 , inference loss is 24.34, weight deacy loss is 13.65, training accuracy is 0.031250, time 198.103 samples/sec
+epoch 0, total_step 10980, total loss is 38.70 , inference loss is 25.05, weight deacy loss is 13.64, training accuracy is 0.062500, time 215.292 samples/sec
+epoch 0, total_step 11000, total loss is 37.54 , inference loss is 23.89, weight deacy loss is 13.64, training accuracy is 0.031250, time 200.639 samples/sec
+epoch 0, total_step 11020, total loss is 36.94 , inference loss is 23.30, weight deacy loss is 13.64, training accuracy is 0.031250, time 211.616 samples/sec
+epoch 0, total_step 11040, total loss is 36.18 , inference loss is 22.54, weight deacy loss is 13.64, training accuracy is 0.062500, time 219.291 samples/sec
+epoch 0, total_step 11060, total loss is 37.76 , inference loss is 24.12, weight deacy loss is 13.64, training accuracy is 0.093750, time 213.990 samples/sec
+epoch 0, total_step 11080, total loss is 38.11 , inference loss is 24.47, weight deacy loss is 13.64, training accuracy is 0.000000, time 223.095 samples/sec
+epoch 0, total_step 11100, total loss is 37.83 , inference loss is 24.20, weight deacy loss is 13.64, training accuracy is 0.000000, time 14.066 samples/sec
+epoch 0, total_step 11120, total loss is 36.44 , inference loss is 22.80, weight deacy loss is 13.63, training accuracy is 0.031250, time 221.715 samples/sec
+epoch 0, total_step 11140, total loss is 35.46 , inference loss is 21.83, weight deacy loss is 13.63, training accuracy is 0.031250, time 209.111 samples/sec
+epoch 0, total_step 11160, total loss is 37.77 , inference loss is 24.14, weight deacy loss is 13.63, training accuracy is 0.125000, time 218.671 samples/sec
+epoch 0, total_step 11180, total loss is 38.87 , inference loss is 25.24, weight deacy loss is 13.63, training accuracy is 0.000000, time 209.045 samples/sec
+epoch 0, total_step 11200, total loss is 38.26 , inference loss is 24.63, weight deacy loss is 13.63, training accuracy is 0.125000, time 212.387 samples/sec
+epoch 0, total_step 11220, total loss is 39.02 , inference loss is 25.40, weight deacy loss is 13.63, training accuracy is 0.062500, time 209.909 samples/sec
+epoch 0, total_step 11240, total loss is 31.34 , inference loss is 17.72, weight deacy loss is 13.63, training accuracy is 0.062500, time 241.065 samples/sec
+epoch 0, total_step 11260, total loss is 39.94 , inference loss is 26.32, weight deacy loss is 13.62, training accuracy is 0.031250, time 225.848 samples/sec
+epoch 0, total_step 11280, total loss is 37.76 , inference loss is 24.13, weight deacy loss is 13.62, training accuracy is 0.062500, time 226.225 samples/sec
+epoch 0, total_step 11300, total loss is 32.42 , inference loss is 18.80, weight deacy loss is 13.62, training accuracy is 0.062500, time 223.987 samples/sec
+epoch 0, total_step 11320, total loss is 38.34 , inference loss is 24.72, weight deacy loss is 13.62, training accuracy is 0.000000, time 201.631 samples/sec
+epoch 0, total_step 11340, total loss is 39.23 , inference loss is 25.61, weight deacy loss is 13.62, training accuracy is 0.031250, time 221.521 samples/sec
+epoch 0, total_step 11360, total loss is 42.65 , inference loss is 29.03, weight deacy loss is 13.62, training accuracy is 0.000000, time 211.564 samples/sec
+epoch 0, total_step 11380, total loss is 36.60 , inference loss is 22.98, weight deacy loss is 13.62, training accuracy is 0.000000, time 298.038 samples/sec
+epoch 0, total_step 11400, total loss is 37.76 , inference loss is 24.14, weight deacy loss is 13.61, training accuracy is 0.062500, time 12.876 samples/sec
+epoch 0, total_step 11420, total loss is 37.51 , inference loss is 23.89, weight deacy loss is 13.61, training accuracy is 0.000000, time 221.214 samples/sec
+epoch 0, total_step 11440, total loss is 36.00 , inference loss is 22.39, weight deacy loss is 13.61, training accuracy is 0.000000, time 202.944 samples/sec
+epoch 0, total_step 11460, total loss is 42.85 , inference loss is 29.24, weight deacy loss is 13.61, training accuracy is 0.031250, time 220.847 samples/sec
+epoch 0, total_step 11480, total loss is 40.95 , inference loss is 27.35, weight deacy loss is 13.61, training accuracy is 0.000000, time 235.209 samples/sec
+epoch 0, total_step 11500, total loss is 40.88 , inference loss is 27.27, weight deacy loss is 13.61, training accuracy is 0.000000, time 219.859 samples/sec
+epoch 0, total_step 11520, total loss is 37.54 , inference loss is 23.93, weight deacy loss is 13.61, training accuracy is 0.093750, time 231.044 samples/sec
+epoch 0, total_step 11540, total loss is 41.14 , inference loss is 27.54, weight deacy loss is 13.60, training accuracy is 0.093750, time 204.428 samples/sec
+epoch 0, total_step 11560, total loss is 37.62 , inference loss is 24.02, weight deacy loss is 13.60, training accuracy is 0.000000, time 208.425 samples/sec
+epoch 0, total_step 11580, total loss is 39.66 , inference loss is 26.06, weight deacy loss is 13.60, training accuracy is 0.000000, time 223.426 samples/sec
+epoch 0, total_step 11600, total loss is 40.00 , inference loss is 26.40, weight deacy loss is 13.60, training accuracy is 0.000000, time 248.482 samples/sec
+epoch 0, total_step 11620, total loss is 37.26 , inference loss is 23.66, weight deacy loss is 13.60, training accuracy is 0.031250, time 208.148 samples/sec
+epoch 0, total_step 11640, total loss is 43.82 , inference loss is 30.23, weight deacy loss is 13.60, training accuracy is 0.031250, time 211.388 samples/sec
+epoch 0, total_step 11660, total loss is 37.73 , inference loss is 24.14, weight deacy loss is 13.60, training accuracy is 0.000000, time 299.042 samples/sec
+epoch 0, total_step 11680, total loss is 39.08 , inference loss is 25.49, weight deacy loss is 13.59, training accuracy is 0.031250, time 205.445 samples/sec
+epoch 0, total_step 11700, total loss is 38.30 , inference loss is 24.71, weight deacy loss is 13.59, training accuracy is 0.093750, time 11.894 samples/sec
+epoch 0, total_step 11720, total loss is 37.84 , inference loss is 24.25, weight deacy loss is 13.59, training accuracy is 0.031250, time 223.323 samples/sec
+epoch 0, total_step 11740, total loss is 36.05 , inference loss is 22.46, weight deacy loss is 13.59, training accuracy is 0.000000, time 228.791 samples/sec
+epoch 0, total_step 11760, total loss is 37.19 , inference loss is 23.60, weight deacy loss is 13.59, training accuracy is 0.062500, time 222.793 samples/sec
+epoch 0, total_step 11780, total loss is 36.80 , inference loss is 23.21, weight deacy loss is 13.59, training accuracy is 0.093750, time 224.475 samples/sec
+epoch 0, total_step 11800, total loss is 35.95 , inference loss is 22.36, weight deacy loss is 13.59, training accuracy is 0.093750, time 232.493 samples/sec
+epoch 0, total_step 11820, total loss is 38.66 , inference loss is 25.07, weight deacy loss is 13.59, training accuracy is 0.000000, time 216.316 samples/sec
+epoch 0, total_step 11840, total loss is 37.59 , inference loss is 24.01, weight deacy loss is 13.58, training accuracy is 0.062500, time 223.586 samples/sec
+epoch 0, total_step 11860, total loss is 34.22 , inference loss is 20.64, weight deacy loss is 13.58, training accuracy is 0.125000, time 196.631 samples/sec
+epoch 0, total_step 11880, total loss is 38.27 , inference loss is 24.69, weight deacy loss is 13.58, training accuracy is 0.031250, time 215.438 samples/sec
+epoch 0, total_step 11900, total loss is 37.59 , inference loss is 24.01, weight deacy loss is 13.58, training accuracy is 0.062500, time 219.288 samples/sec
+epoch 0, total_step 11920, total loss is 38.89 , inference loss is 25.32, weight deacy loss is 13.58, training accuracy is 0.000000, time 296.830 samples/sec
+epoch 0, total_step 11940, total loss is 42.45 , inference loss is 28.88, weight deacy loss is 13.58, training accuracy is 0.031250, time 220.086 samples/sec
+epoch 0, total_step 11960, total loss is 41.92 , inference loss is 28.35, weight deacy loss is 13.58, training accuracy is 0.000000, time 214.338 samples/sec
+epoch 0, total_step 11980, total loss is 37.53 , inference loss is 23.96, weight deacy loss is 13.57, training accuracy is 0.000000, time 193.839 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.997808999999997
+best_threshold_index 95 0.9512962962962963
+best_threshold_index 96 0.9512962962962963
+best_threshold_index 95 0.952037037037037
+best_threshold_index 95 0.952962962962963
+best_threshold_index 96 0.9531481481481482
+best_threshold_index 95 0.9503703703703704
+best_threshold_index 95 0.9512962962962963
+best_threshold_index 95 0.9514814814814815
+best_threshold_index 95 0.9483333333333334
+best_threshold_index 100 0.9503703703703704
+[lfw][12000]XNorm: 24.727908
+[lfw][12000]Accuracy-Flip: 0.94967+-0.01190
+test accuracy is: 0.9496666666666668
+epoch 0, total_step 12000, total loss is 39.99 , inference loss is 26.42, weight deacy loss is 13.57, training accuracy is 0.031250, time 17.938 samples/sec
+epoch 0, total_step 12020, total loss is 38.73 , inference loss is 25.16, weight deacy loss is 13.57, training accuracy is 0.062500, time 195.087 samples/sec
+epoch 0, total_step 12040, total loss is 37.61 , inference loss is 24.04, weight deacy loss is 13.57, training accuracy is 0.031250, time 241.873 samples/sec
+epoch 0, total_step 12060, total loss is 35.60 , inference loss is 22.03, weight deacy loss is 13.57, training accuracy is 0.062500, time 216.549 samples/sec
+epoch 0, total_step 12080, total loss is 37.07 , inference loss is 23.51, weight deacy loss is 13.57, training accuracy is 0.000000, time 212.745 samples/sec
+epoch 0, total_step 12100, total loss is 38.31 , inference loss is 24.74, weight deacy loss is 13.57, training accuracy is 0.000000, time 225.995 samples/sec
+epoch 0, total_step 12120, total loss is 40.09 , inference loss is 26.52, weight deacy loss is 13.56, training accuracy is 0.000000, time 224.176 samples/sec
+epoch 0, total_step 12140, total loss is 37.19 , inference loss is 23.63, weight deacy loss is 13.56, training accuracy is 0.062500, time 210.239 samples/sec
+epoch 0, total_step 12160, total loss is 36.03 , inference loss is 22.47, weight deacy loss is 13.56, training accuracy is 0.062500, time 220.922 samples/sec
+epoch 0, total_step 12180, total loss is 35.77 , inference loss is 22.21, weight deacy loss is 13.56, training accuracy is 0.062500, time 214.781 samples/sec
+epoch 0, total_step 12200, total loss is 37.18 , inference loss is 23.62, weight deacy loss is 13.56, training accuracy is 0.062500, time 297.216 samples/sec
+epoch 0, total_step 12220, total loss is 39.60 , inference loss is 26.04, weight deacy loss is 13.56, training accuracy is 0.031250, time 212.977 samples/sec
+epoch 0, total_step 12240, total loss is 38.13 , inference loss is 24.57, weight deacy loss is 13.56, training accuracy is 0.093750, time 220.416 samples/sec
+epoch 0, total_step 12260, total loss is 36.56 , inference loss is 23.00, weight deacy loss is 13.56, training accuracy is 0.062500, time 217.870 samples/sec
+epoch 0, total_step 12280, total loss is 39.13 , inference loss is 25.57, weight deacy loss is 13.55, training accuracy is 0.000000, time 214.480 samples/sec
+epoch 0, total_step 12300, total loss is 40.95 , inference loss is 27.40, weight deacy loss is 13.55, training accuracy is 0.000000, time 14.479 samples/sec
+epoch 0, total_step 12320, total loss is 41.26 , inference loss is 27.71, weight deacy loss is 13.55, training accuracy is 0.031250, time 298.140 samples/sec
+epoch 0, total_step 12340, total loss is 38.98 , inference loss is 25.43, weight deacy loss is 13.55, training accuracy is 0.000000, time 184.501 samples/sec
+epoch 0, total_step 12360, total loss is 36.35 , inference loss is 22.80, weight deacy loss is 13.55, training accuracy is 0.093750, time 227.407 samples/sec
+epoch 0, total_step 12380, total loss is 34.97 , inference loss is 21.43, weight deacy loss is 13.55, training accuracy is 0.031250, time 242.054 samples/sec
+epoch 0, total_step 12400, total loss is 36.75 , inference loss is 23.20, weight deacy loss is 13.55, training accuracy is 0.000000, time 227.707 samples/sec
+epoch 0, total_step 12420, total loss is 38.68 , inference loss is 25.14, weight deacy loss is 13.54, training accuracy is 0.000000, time 225.047 samples/sec
+epoch 0, total_step 12440, total loss is 37.81 , inference loss is 24.27, weight deacy loss is 13.54, training accuracy is 0.031250, time 299.968 samples/sec
+epoch 0, total_step 12460, total loss is 36.44 , inference loss is 22.90, weight deacy loss is 13.54, training accuracy is 0.062500, time 220.937 samples/sec
+epoch 0, total_step 12480, total loss is 36.20 , inference loss is 22.66, weight deacy loss is 13.54, training accuracy is 0.031250, time 221.390 samples/sec
+epoch 0, total_step 12500, total loss is 38.21 , inference loss is 24.67, weight deacy loss is 13.54, training accuracy is 0.031250, time 204.303 samples/sec
+epoch 0, total_step 12520, total loss is 37.63 , inference loss is 24.09, weight deacy loss is 13.54, training accuracy is 0.000000, time 218.937 samples/sec
+epoch 0, total_step 12540, total loss is 41.22 , inference loss is 27.68, weight deacy loss is 13.54, training accuracy is 0.000000, time 205.895 samples/sec
+epoch 0, total_step 12560, total loss is 36.79 , inference loss is 23.25, weight deacy loss is 13.53, training accuracy is 0.031250, time 213.180 samples/sec
+epoch 0, total_step 12580, total loss is 39.26 , inference loss is 25.73, weight deacy loss is 13.53, training accuracy is 0.000000, time 214.350 samples/sec
+epoch 0, total_step 12600, total loss is 31.46 , inference loss is 17.92, weight deacy loss is 13.53, training accuracy is 0.031250, time 14.604 samples/sec
+epoch 0, total_step 12620, total loss is 38.04 , inference loss is 24.51, weight deacy loss is 13.53, training accuracy is 0.000000, time 214.998 samples/sec
+epoch 0, total_step 12640, total loss is 41.43 , inference loss is 27.90, weight deacy loss is 13.53, training accuracy is 0.031250, time 204.065 samples/sec
+epoch 0, total_step 12660, total loss is 31.72 , inference loss is 18.19, weight deacy loss is 13.53, training accuracy is 0.062500, time 215.844 samples/sec
+epoch 0, total_step 12680, total loss is 40.08 , inference loss is 26.56, weight deacy loss is 13.53, training accuracy is 0.093750, time 206.267 samples/sec
+epoch 0, total_step 12700, total loss is 41.64 , inference loss is 28.12, weight deacy loss is 13.52, training accuracy is 0.031250, time 297.176 samples/sec
+epoch 0, total_step 12720, total loss is 39.78 , inference loss is 26.26, weight deacy loss is 13.52, training accuracy is 0.000000, time 204.020 samples/sec
+epoch 0, total_step 12740, total loss is 33.69 , inference loss is 20.17, weight deacy loss is 13.52, training accuracy is 0.000000, time 200.741 samples/sec
+epoch 0, total_step 12760, total loss is 34.32 , inference loss is 20.80, weight deacy loss is 13.52, training accuracy is 0.000000, time 218.919 samples/sec
+epoch 0, total_step 12780, total loss is 37.79 , inference loss is 24.27, weight deacy loss is 13.52, training accuracy is 0.000000, time 225.545 samples/sec
+epoch 0, total_step 12800, total loss is 34.23 , inference loss is 20.71, weight deacy loss is 13.52, training accuracy is 0.156250, time 226.172 samples/sec
+epoch 0, total_step 12820, total loss is 38.38 , inference loss is 24.87, weight deacy loss is 13.52, training accuracy is 0.062500, time 228.134 samples/sec
+epoch 0, total_step 12840, total loss is 43.66 , inference loss is 30.15, weight deacy loss is 13.52, training accuracy is 0.031250, time 210.118 samples/sec
+epoch 0, total_step 12860, total loss is 38.14 , inference loss is 24.62, weight deacy loss is 13.51, training accuracy is 0.031250, time 223.364 samples/sec
+epoch 0, total_step 12880, total loss is 36.81 , inference loss is 23.30, weight deacy loss is 13.51, training accuracy is 0.031250, time 208.410 samples/sec
+epoch 0, total_step 12900, total loss is 38.30 , inference loss is 24.79, weight deacy loss is 13.51, training accuracy is 0.031250, time 15.440 samples/sec
+epoch 0, total_step 12920, total loss is 40.88 , inference loss is 27.37, weight deacy loss is 13.51, training accuracy is 0.000000, time 207.877 samples/sec
+epoch 0, total_step 12940, total loss is 40.37 , inference loss is 26.86, weight deacy loss is 13.51, training accuracy is 0.031250, time 213.916 samples/sec
+epoch 0, total_step 12960, total loss is 41.41 , inference loss is 27.90, weight deacy loss is 13.51, training accuracy is 0.000000, time 230.018 samples/sec
+epoch 0, total_step 12980, total loss is 38.91 , inference loss is 25.40, weight deacy loss is 13.51, training accuracy is 0.031250, time 300.302 samples/sec
+epoch 0, total_step 13000, total loss is 38.76 , inference loss is 25.26, weight deacy loss is 13.50, training accuracy is 0.031250, time 211.906 samples/sec
+epoch 0, total_step 13020, total loss is 41.81 , inference loss is 28.31, weight deacy loss is 13.50, training accuracy is 0.000000, time 215.845 samples/sec
+epoch 0, total_step 13040, total loss is 42.65 , inference loss is 29.15, weight deacy loss is 13.50, training accuracy is 0.000000, time 229.857 samples/sec
+epoch 0, total_step 13060, total loss is 34.86 , inference loss is 21.36, weight deacy loss is 13.50, training accuracy is 0.062500, time 230.896 samples/sec
+epoch 0, total_step 13080, total loss is 39.20 , inference loss is 25.71, weight deacy loss is 13.50, training accuracy is 0.000000, time 214.365 samples/sec
+epoch 0, total_step 13100, total loss is 36.24 , inference loss is 22.75, weight deacy loss is 13.50, training accuracy is 0.000000, time 225.950 samples/sec
+epoch 0, total_step 13120, total loss is 40.56 , inference loss is 27.06, weight deacy loss is 13.50, training accuracy is 0.031250, time 203.559 samples/sec
+epoch 0, total_step 13140, total loss is 39.36 , inference loss is 25.86, weight deacy loss is 13.49, training accuracy is 0.093750, time 220.586 samples/sec
+epoch 0, total_step 13160, total loss is 40.36 , inference loss is 26.87, weight deacy loss is 13.49, training accuracy is 0.062500, time 225.186 samples/sec
+epoch 0, total_step 13180, total loss is 32.95 , inference loss is 19.46, weight deacy loss is 13.49, training accuracy is 0.062500, time 226.345 samples/sec
+epoch 0, total_step 13200, total loss is 39.38 , inference loss is 25.88, weight deacy loss is 13.49, training accuracy is 0.000000, time 12.092 samples/sec
+epoch 0, total_step 13220, total loss is 40.58 , inference loss is 27.09, weight deacy loss is 13.49, training accuracy is 0.031250, time 216.023 samples/sec
+epoch 0, total_step 13240, total loss is 46.04 , inference loss is 32.55, weight deacy loss is 13.49, training accuracy is 0.000000, time 300.312 samples/sec
+epoch 0, total_step 13260, total loss is 39.09 , inference loss is 25.60, weight deacy loss is 13.49, training accuracy is 0.031250, time 224.483 samples/sec
+epoch 0, total_step 13280, total loss is 40.34 , inference loss is 26.86, weight deacy loss is 13.49, training accuracy is 0.093750, time 218.065 samples/sec
+epoch 0, total_step 13300, total loss is 39.67 , inference loss is 26.18, weight deacy loss is 13.48, training accuracy is 0.000000, time 222.821 samples/sec
+epoch 0, total_step 13320, total loss is 35.13 , inference loss is 21.64, weight deacy loss is 13.48, training accuracy is 0.062500, time 225.784 samples/sec
+epoch 0, total_step 13340, total loss is 36.40 , inference loss is 22.92, weight deacy loss is 13.48, training accuracy is 0.000000, time 210.670 samples/sec
+epoch 0, total_step 13360, total loss is 39.08 , inference loss is 25.60, weight deacy loss is 13.48, training accuracy is 0.000000, time 243.307 samples/sec
+epoch 0, total_step 13380, total loss is 41.30 , inference loss is 27.82, weight deacy loss is 13.48, training accuracy is 0.000000, time 209.024 samples/sec
+epoch 0, total_step 13400, total loss is 39.38 , inference loss is 25.90, weight deacy loss is 13.48, training accuracy is 0.062500, time 212.324 samples/sec
+epoch 0, total_step 13420, total loss is 43.33 , inference loss is 29.86, weight deacy loss is 13.48, training accuracy is 0.000000, time 214.658 samples/sec
+epoch 0, total_step 13440, total loss is 34.61 , inference loss is 21.13, weight deacy loss is 13.47, training accuracy is 0.031250, time 203.046 samples/sec
+epoch 0, total_step 13460, total loss is 40.37 , inference loss is 26.90, weight deacy loss is 13.47, training accuracy is 0.031250, time 211.516 samples/sec
+epoch 0, total_step 13480, total loss is 40.28 , inference loss is 26.81, weight deacy loss is 13.47, training accuracy is 0.000000, time 208.244 samples/sec
+epoch 0, total_step 13500, total loss is 35.66 , inference loss is 22.19, weight deacy loss is 13.47, training accuracy is 0.093750, time 14.059 samples/sec
+epoch 0, total_step 13520, total loss is 36.24 , inference loss is 22.77, weight deacy loss is 13.47, training accuracy is 0.000000, time 220.341 samples/sec
+epoch 0, total_step 13540, total loss is 39.94 , inference loss is 26.47, weight deacy loss is 13.47, training accuracy is 0.031250, time 228.827 samples/sec
+epoch 0, total_step 13560, total loss is 39.49 , inference loss is 26.02, weight deacy loss is 13.47, training accuracy is 0.000000, time 221.117 samples/sec
+epoch 0, total_step 13580, total loss is 33.54 , inference loss is 20.07, weight deacy loss is 13.46, training accuracy is 0.062500, time 213.486 samples/sec
+epoch 0, total_step 13600, total loss is 38.02 , inference loss is 24.56, weight deacy loss is 13.46, training accuracy is 0.000000, time 218.919 samples/sec
+epoch 0, total_step 13620, total loss is 42.46 , inference loss is 29.00, weight deacy loss is 13.46, training accuracy is 0.031250, time 230.773 samples/sec
+epoch 0, total_step 13640, total loss is 37.28 , inference loss is 23.82, weight deacy loss is 13.46, training accuracy is 0.062500, time 219.762 samples/sec
+epoch 0, total_step 13660, total loss is 36.39 , inference loss is 22.93, weight deacy loss is 13.46, training accuracy is 0.000000, time 228.099 samples/sec
+epoch 0, total_step 13680, total loss is 44.51 , inference loss is 31.06, weight deacy loss is 13.46, training accuracy is 0.000000, time 183.957 samples/sec
+epoch 0, total_step 13700, total loss is 36.37 , inference loss is 22.92, weight deacy loss is 13.46, training accuracy is 0.062500, time 225.160 samples/sec
+epoch 0, total_step 13720, total loss is 38.12 , inference loss is 24.67, weight deacy loss is 13.45, training accuracy is 0.031250, time 213.079 samples/sec
+epoch 0, total_step 13740, total loss is 41.80 , inference loss is 28.34, weight deacy loss is 13.45, training accuracy is 0.000000, time 226.404 samples/sec
+epoch 0, total_step 13760, total loss is 36.45 , inference loss is 23.00, weight deacy loss is 13.45, training accuracy is 0.000000, time 212.294 samples/sec
+epoch 0, total_step 13780, total loss is 37.18 , inference loss is 23.73, weight deacy loss is 13.45, training accuracy is 0.000000, time 206.092 samples/sec
+epoch 0, total_step 13800, total loss is 34.79 , inference loss is 21.34, weight deacy loss is 13.45, training accuracy is 0.062500, time 13.525 samples/sec
+epoch 0, total_step 13820, total loss is 38.95 , inference loss is 25.51, weight deacy loss is 13.45, training accuracy is 0.000000, time 216.952 samples/sec
+epoch 0, total_step 13840, total loss is 39.95 , inference loss is 26.51, weight deacy loss is 13.45, training accuracy is 0.000000, time 239.524 samples/sec
+epoch 0, total_step 13860, total loss is 35.47 , inference loss is 22.02, weight deacy loss is 13.45, training accuracy is 0.031250, time 210.739 samples/sec
+epoch 0, total_step 13880, total loss is 43.88 , inference loss is 30.43, weight deacy loss is 13.44, training accuracy is 0.031250, time 222.922 samples/sec
+epoch 0, total_step 13900, total loss is 37.69 , inference loss is 24.25, weight deacy loss is 13.44, training accuracy is 0.031250, time 201.660 samples/sec
+epoch 0, total_step 13920, total loss is 35.11 , inference loss is 21.67, weight deacy loss is 13.44, training accuracy is 0.125000, time 219.178 samples/sec
+epoch 0, total_step 13940, total loss is 37.75 , inference loss is 24.31, weight deacy loss is 13.44, training accuracy is 0.062500, time 223.775 samples/sec
+epoch 0, total_step 13960, total loss is 36.75 , inference loss is 23.31, weight deacy loss is 13.44, training accuracy is 0.093750, time 223.028 samples/sec
+epoch 0, total_step 13980, total loss is 39.05 , inference loss is 25.62, weight deacy loss is 13.44, training accuracy is 0.031250, time 223.692 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.117879000000013
+best_threshold_index 93 0.9553703703703704
+best_threshold_index 90 0.9561111111111111
+best_threshold_index 96 0.9561111111111111
+best_threshold_index 90 0.957962962962963
+best_threshold_index 90 0.9572222222222222
+best_threshold_index 90 0.9538888888888889
+best_threshold_index 93 0.9555555555555556
+best_threshold_index 90 0.9553703703703704
+best_threshold_index 90 0.9538888888888889
+best_threshold_index 90 0.9542592592592593
+[lfw][14000]XNorm: 26.058693
+[lfw][14000]Accuracy-Flip: 0.95333+-0.01213
+test accuracy is: 0.9533333333333334
+epoch 0, total_step 14000, total loss is 38.55 , inference loss is 25.11, weight deacy loss is 13.44, training accuracy is 0.000000, time 18.171 samples/sec
+epoch 0, total_step 14020, total loss is 37.11 , inference loss is 23.68, weight deacy loss is 13.43, training accuracy is 0.031250, time 225.992 samples/sec
+epoch 0, total_step 14040, total loss is 35.85 , inference loss is 22.41, weight deacy loss is 13.43, training accuracy is 0.031250, time 213.327 samples/sec
+epoch 0, total_step 14060, total loss is 40.02 , inference loss is 26.59, weight deacy loss is 13.43, training accuracy is 0.031250, time 209.927 samples/sec
+epoch 0, total_step 14080, total loss is 38.06 , inference loss is 24.63, weight deacy loss is 13.43, training accuracy is 0.000000, time 212.122 samples/sec
+epoch 0, total_step 14100, total loss is 38.63 , inference loss is 25.20, weight deacy loss is 13.43, training accuracy is 0.000000, time 15.665 samples/sec
+epoch 0, total_step 14120, total loss is 35.79 , inference loss is 22.36, weight deacy loss is 13.43, training accuracy is 0.062500, time 301.091 samples/sec
+epoch 0, total_step 14140, total loss is 33.67 , inference loss is 20.25, weight deacy loss is 13.43, training accuracy is 0.062500, time 222.884 samples/sec
+epoch 0, total_step 14160, total loss is 41.01 , inference loss is 27.58, weight deacy loss is 13.42, training accuracy is 0.062500, time 225.926 samples/sec
+epoch 0, total_step 14180, total loss is 41.56 , inference loss is 28.14, weight deacy loss is 13.42, training accuracy is 0.031250, time 218.788 samples/sec
+epoch 0, total_step 14200, total loss is 36.00 , inference loss is 22.58, weight deacy loss is 13.42, training accuracy is 0.031250, time 221.069 samples/sec
+epoch 0, total_step 14220, total loss is 36.76 , inference loss is 23.34, weight deacy loss is 13.42, training accuracy is 0.031250, time 223.021 samples/sec
+epoch 0, total_step 14240, total loss is 38.36 , inference loss is 24.94, weight deacy loss is 13.42, training accuracy is 0.000000, time 215.100 samples/sec
+epoch 0, total_step 14260, total loss is 36.05 , inference loss is 22.63, weight deacy loss is 13.42, training accuracy is 0.031250, time 229.308 samples/sec
+epoch 0, total_step 14280, total loss is 37.89 , inference loss is 24.47, weight deacy loss is 13.42, training accuracy is 0.000000, time 210.691 samples/sec
+epoch 0, total_step 14300, total loss is 37.80 , inference loss is 24.38, weight deacy loss is 13.42, training accuracy is 0.000000, time 220.940 samples/sec
+epoch 0, total_step 14320, total loss is 36.62 , inference loss is 23.20, weight deacy loss is 13.41, training accuracy is 0.031250, time 217.100 samples/sec
+epoch 0, total_step 14340, total loss is 37.45 , inference loss is 24.04, weight deacy loss is 13.41, training accuracy is 0.031250, time 195.522 samples/sec
+epoch 0, total_step 14360, total loss is 39.50 , inference loss is 26.09, weight deacy loss is 13.41, training accuracy is 0.000000, time 215.555 samples/sec
+epoch 0, total_step 14380, total loss is 37.36 , inference loss is 23.95, weight deacy loss is 13.41, training accuracy is 0.000000, time 232.753 samples/sec
+epoch 0, total_step 14400, total loss is 41.96 , inference loss is 28.55, weight deacy loss is 13.41, training accuracy is 0.031250, time 14.629 samples/sec
+epoch 0, total_step 14420, total loss is 38.48 , inference loss is 25.07, weight deacy loss is 13.41, training accuracy is 0.031250, time 229.254 samples/sec
+epoch 0, total_step 14440, total loss is 38.89 , inference loss is 25.48, weight deacy loss is 13.41, training accuracy is 0.000000, time 207.990 samples/sec
+epoch 0, total_step 14460, total loss is 42.58 , inference loss is 29.18, weight deacy loss is 13.40, training accuracy is 0.093750, time 202.294 samples/sec
+epoch 0, total_step 14480, total loss is 40.25 , inference loss is 26.84, weight deacy loss is 13.40, training accuracy is 0.000000, time 194.393 samples/sec
+epoch 0, total_step 14500, total loss is 40.70 , inference loss is 27.30, weight deacy loss is 13.40, training accuracy is 0.000000, time 224.711 samples/sec
+epoch 0, total_step 14520, total loss is 36.80 , inference loss is 23.40, weight deacy loss is 13.40, training accuracy is 0.000000, time 219.406 samples/sec
+epoch 0, total_step 14540, total loss is 40.07 , inference loss is 26.67, weight deacy loss is 13.40, training accuracy is 0.000000, time 220.591 samples/sec
+epoch 0, total_step 14560, total loss is 39.68 , inference loss is 26.28, weight deacy loss is 13.40, training accuracy is 0.031250, time 214.990 samples/sec
+epoch 0, total_step 14580, total loss is 37.08 , inference loss is 23.68, weight deacy loss is 13.40, training accuracy is 0.000000, time 222.882 samples/sec
+epoch 0, total_step 14600, total loss is 34.32 , inference loss is 20.92, weight deacy loss is 13.40, training accuracy is 0.031250, time 298.009 samples/sec
+epoch 0, total_step 14620, total loss is 38.68 , inference loss is 25.28, weight deacy loss is 13.39, training accuracy is 0.031250, time 209.670 samples/sec
+epoch 0, total_step 14640, total loss is 38.11 , inference loss is 24.71, weight deacy loss is 13.39, training accuracy is 0.000000, time 221.591 samples/sec
+epoch 0, total_step 14660, total loss is 42.21 , inference loss is 28.82, weight deacy loss is 13.39, training accuracy is 0.000000, time 211.584 samples/sec
+epoch 0, total_step 14680, total loss is 36.76 , inference loss is 23.37, weight deacy loss is 13.39, training accuracy is 0.000000, time 213.023 samples/sec
+epoch 0, total_step 14700, total loss is 40.19 , inference loss is 26.81, weight deacy loss is 13.39, training accuracy is 0.000000, time 15.393 samples/sec
+epoch 0, total_step 14720, total loss is 36.58 , inference loss is 23.19, weight deacy loss is 13.39, training accuracy is 0.031250, time 214.713 samples/sec
+epoch 0, total_step 14740, total loss is 41.93 , inference loss is 28.54, weight deacy loss is 13.39, training accuracy is 0.000000, time 224.580 samples/sec
+epoch 0, total_step 14760, total loss is 40.00 , inference loss is 26.61, weight deacy loss is 13.38, training accuracy is 0.031250, time 217.963 samples/sec
+epoch 0, total_step 14780, total loss is 42.65 , inference loss is 29.26, weight deacy loss is 13.38, training accuracy is 0.000000, time 212.725 samples/sec
+epoch 0, total_step 14800, total loss is 39.91 , inference loss is 26.53, weight deacy loss is 13.38, training accuracy is 0.031250, time 212.115 samples/sec
+epoch 0, total_step 14820, total loss is 40.92 , inference loss is 27.54, weight deacy loss is 13.38, training accuracy is 0.000000, time 227.484 samples/sec
+epoch 0, total_step 14840, total loss is 40.84 , inference loss is 27.46, weight deacy loss is 13.38, training accuracy is 0.000000, time 224.444 samples/sec
+epoch 0, total_step 14860, total loss is 43.56 , inference loss is 30.19, weight deacy loss is 13.38, training accuracy is 0.000000, time 297.697 samples/sec
+epoch 0, total_step 14880, total loss is 43.70 , inference loss is 30.32, weight deacy loss is 13.38, training accuracy is 0.031250, time 226.424 samples/sec
+epoch 0, total_step 14900, total loss is 38.88 , inference loss is 25.51, weight deacy loss is 13.38, training accuracy is 0.000000, time 224.531 samples/sec
+epoch 0, total_step 14920, total loss is 36.92 , inference loss is 23.54, weight deacy loss is 13.37, training accuracy is 0.031250, time 206.854 samples/sec
+epoch 0, total_step 14940, total loss is 45.98 , inference loss is 32.61, weight deacy loss is 13.37, training accuracy is 0.031250, time 221.752 samples/sec
+epoch 0, total_step 14960, total loss is 38.66 , inference loss is 25.29, weight deacy loss is 13.37, training accuracy is 0.000000, time 219.503 samples/sec
+epoch 0, total_step 14980, total loss is 43.98 , inference loss is 30.61, weight deacy loss is 13.37, training accuracy is 0.000000, time 209.145 samples/sec
+epoch 0, total_step 15000, total loss is 37.50 , inference loss is 24.13, weight deacy loss is 13.37, training accuracy is 0.000000, time 11.566 samples/sec
+epoch 0, total_step 15020, total loss is 38.17 , inference loss is 24.80, weight deacy loss is 13.37, training accuracy is 0.000000, time 197.171 samples/sec
+epoch 0, total_step 15040, total loss is 38.61 , inference loss is 25.24, weight deacy loss is 13.37, training accuracy is 0.000000, time 182.634 samples/sec
+epoch 0, total_step 15060, total loss is 36.63 , inference loss is 23.27, weight deacy loss is 13.37, training accuracy is 0.000000, time 204.886 samples/sec
+epoch 0, total_step 15080, total loss is 35.00 , inference loss is 21.64, weight deacy loss is 13.36, training accuracy is 0.031250, time 193.870 samples/sec
+epoch 0, total_step 15100, total loss is 32.96 , inference loss is 19.59, weight deacy loss is 13.36, training accuracy is 0.031250, time 177.908 samples/sec
+epoch 0, total_step 15120, total loss is 36.54 , inference loss is 23.18, weight deacy loss is 13.36, training accuracy is 0.062500, time 201.945 samples/sec
+epoch 0, total_step 15140, total loss is 35.45 , inference loss is 22.10, weight deacy loss is 13.36, training accuracy is 0.031250, time 298.447 samples/sec
+epoch 0, total_step 15160, total loss is 32.68 , inference loss is 19.32, weight deacy loss is 13.36, training accuracy is 0.000000, time 194.501 samples/sec
+epoch 0, total_step 15180, total loss is 32.67 , inference loss is 19.31, weight deacy loss is 13.36, training accuracy is 0.031250, time 175.926 samples/sec
+epoch 0, total_step 15200, total loss is 33.19 , inference loss is 19.84, weight deacy loss is 13.35, training accuracy is 0.031250, time 190.640 samples/sec
+epoch 0, total_step 15220, total loss is 39.04 , inference loss is 25.69, weight deacy loss is 13.35, training accuracy is 0.000000, time 216.734 samples/sec
+epoch 0, total_step 15240, total loss is 36.01 , inference loss is 22.66, weight deacy loss is 13.35, training accuracy is 0.000000, time 225.411 samples/sec
+epoch 0, total_step 15260, total loss is 35.77 , inference loss is 22.42, weight deacy loss is 13.35, training accuracy is 0.031250, time 221.000 samples/sec
+epoch 0, total_step 15280, total loss is 38.24 , inference loss is 24.89, weight deacy loss is 13.35, training accuracy is 0.031250, time 198.375 samples/sec2022-12-01 01:13:44.953921: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7248 of 10000
+2022-12-01 01:13:48.269929: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 0, total_step 15300, total loss is 30.14 , inference loss is 16.80, weight deacy loss is 13.34, training accuracy is 0.000000, time 14.457 samples/sec
+epoch 0, total_step 15320, total loss is 29.36 , inference loss is 16.01, weight deacy loss is 13.34, training accuracy is 0.125000, time 197.092 samples/sec
+End of epoch 0
+epoch 1, total_step 15340, total loss is 62.58 , inference loss is 49.24, weight deacy loss is 13.34, training accuracy is 0.000000, time 213.152 samples/sec
+epoch 1, total_step 15360, total loss is 46.78 , inference loss is 33.44, weight deacy loss is 13.34, training accuracy is 0.000000, time 224.316 samples/sec
+epoch 1, total_step 15380, total loss is 40.23 , inference loss is 26.89, weight deacy loss is 13.34, training accuracy is 0.031250, time 223.842 samples/sec
+epoch 1, total_step 15400, total loss is 42.22 , inference loss is 28.88, weight deacy loss is 13.34, training accuracy is 0.093750, time 211.782 samples/sec
+epoch 1, total_step 15420, total loss is 28.99 , inference loss is 15.65, weight deacy loss is 13.33, training accuracy is 0.218750, time 210.252 samples/sec
+epoch 1, total_step 15440, total loss is 31.34 , inference loss is 18.01, weight deacy loss is 13.33, training accuracy is 0.093750, time 207.684 samples/sec
+epoch 1, total_step 15460, total loss is 31.77 , inference loss is 18.44, weight deacy loss is 13.33, training accuracy is 0.125000, time 239.540 samples/sec
+epoch 1, total_step 15480, total loss is 34.84 , inference loss is 21.51, weight deacy loss is 13.33, training accuracy is 0.062500, time 217.104 samples/sec
+epoch 1, total_step 15500, total loss is 35.14 , inference loss is 21.82, weight deacy loss is 13.33, training accuracy is 0.187500, time 209.224 samples/sec
+epoch 1, total_step 15520, total loss is 30.15 , inference loss is 16.82, weight deacy loss is 13.32, training accuracy is 0.250000, time 215.481 samples/sec
+epoch 1, total_step 15540, total loss is 36.36 , inference loss is 23.04, weight deacy loss is 13.32, training accuracy is 0.062500, time 224.359 samples/sec
+epoch 1, total_step 15560, total loss is 33.69 , inference loss is 20.37, weight deacy loss is 13.32, training accuracy is 0.093750, time 212.863 samples/sec
+epoch 1, total_step 15580, total loss is 34.40 , inference loss is 21.09, weight deacy loss is 13.32, training accuracy is 0.093750, time 213.870 samples/sec
+epoch 1, total_step 15600, total loss is 33.89 , inference loss is 20.58, weight deacy loss is 13.32, training accuracy is 0.093750, time 14.050 samples/sec
+epoch 1, total_step 15620, total loss is 31.92 , inference loss is 18.60, weight deacy loss is 13.31, training accuracy is 0.093750, time 293.997 samples/sec
+epoch 1, total_step 15640, total loss is 33.92 , inference loss is 20.60, weight deacy loss is 13.31, training accuracy is 0.093750, time 218.951 samples/sec
+epoch 1, total_step 15660, total loss is 34.29 , inference loss is 20.98, weight deacy loss is 13.31, training accuracy is 0.125000, time 225.414 samples/sec
+epoch 1, total_step 15680, total loss is 35.21 , inference loss is 21.90, weight deacy loss is 13.31, training accuracy is 0.125000, time 199.353 samples/sec
+epoch 1, total_step 15700, total loss is 35.42 , inference loss is 22.12, weight deacy loss is 13.31, training accuracy is 0.156250, time 207.541 samples/sec
+epoch 1, total_step 15720, total loss is 34.96 , inference loss is 21.66, weight deacy loss is 13.30, training accuracy is 0.156250, time 222.549 samples/sec
+epoch 1, total_step 15740, total loss is 33.62 , inference loss is 20.32, weight deacy loss is 13.30, training accuracy is 0.156250, time 223.562 samples/sec
+epoch 1, total_step 15760, total loss is 33.05 , inference loss is 19.75, weight deacy loss is 13.30, training accuracy is 0.062500, time 229.408 samples/sec
+epoch 1, total_step 15780, total loss is 33.42 , inference loss is 20.12, weight deacy loss is 13.30, training accuracy is 0.093750, time 216.465 samples/sec
+epoch 1, total_step 15800, total loss is 31.97 , inference loss is 18.67, weight deacy loss is 13.30, training accuracy is 0.093750, time 222.496 samples/sec
+epoch 1, total_step 15820, total loss is 35.58 , inference loss is 22.28, weight deacy loss is 13.29, training accuracy is 0.156250, time 221.208 samples/sec
+epoch 1, total_step 15840, total loss is 30.42 , inference loss is 17.12, weight deacy loss is 13.29, training accuracy is 0.093750, time 200.500 samples/sec
+epoch 1, total_step 15860, total loss is 30.09 , inference loss is 16.80, weight deacy loss is 13.29, training accuracy is 0.093750, time 221.385 samples/sec
+epoch 1, total_step 15880, total loss is 25.18 , inference loss is 11.89, weight deacy loss is 13.29, training accuracy is 0.250000, time 205.777 samples/sec
+epoch 1, total_step 15900, total loss is 31.43 , inference loss is 18.14, weight deacy loss is 13.29, training accuracy is 0.187500, time 14.210 samples/sec
+epoch 1, total_step 15920, total loss is 34.94 , inference loss is 21.66, weight deacy loss is 13.29, training accuracy is 0.093750, time 213.430 samples/sec
+epoch 1, total_step 15940, total loss is 33.74 , inference loss is 20.46, weight deacy loss is 13.28, training accuracy is 0.093750, time 215.778 samples/sec
+epoch 1, total_step 15960, total loss is 29.70 , inference loss is 16.41, weight deacy loss is 13.28, training accuracy is 0.125000, time 226.504 samples/sec
+epoch 1, total_step 15980, total loss is 32.01 , inference loss is 18.73, weight deacy loss is 13.28, training accuracy is 0.093750, time 212.397 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.228116000000014
+best_threshold_index 101 0.9598148148148148
+best_threshold_index 101 0.9601851851851851
+best_threshold_index 101 0.9598148148148148
+best_threshold_index 101 0.9611111111111111
+best_threshold_index 101 0.960925925925926
+best_threshold_index 101 0.9575925925925926
+best_threshold_index 101 0.9598148148148148
+best_threshold_index 101 0.9594444444444444
+best_threshold_index 101 0.9575925925925926
+best_threshold_index 101 0.9587037037037037
+[lfw][16000]XNorm: 24.356007
+[lfw][16000]Accuracy-Flip: 0.95950+-0.01038
+test accuracy is: 0.9595
+epoch 1, total_step 16000, total loss is 36.09 , inference loss is 22.81, weight deacy loss is 13.28, training accuracy is 0.093750, time 18.785 samples/sec
+epoch 1, total_step 16020, total loss is 32.17 , inference loss is 18.89, weight deacy loss is 13.28, training accuracy is 0.031250, time 222.024 samples/sec
+epoch 1, total_step 16040, total loss is 35.15 , inference loss is 21.88, weight deacy loss is 13.27, training accuracy is 0.093750, time 228.617 samples/sec
+epoch 1, total_step 16060, total loss is 31.57 , inference loss is 18.30, weight deacy loss is 13.27, training accuracy is 0.125000, time 203.901 samples/sec
+epoch 1, total_step 16080, total loss is 34.37 , inference loss is 21.10, weight deacy loss is 13.27, training accuracy is 0.031250, time 203.162 samples/sec
+epoch 1, total_step 16100, total loss is 34.20 , inference loss is 20.93, weight deacy loss is 13.27, training accuracy is 0.062500, time 233.088 samples/sec
+epoch 1, total_step 16120, total loss is 34.34 , inference loss is 21.08, weight deacy loss is 13.27, training accuracy is 0.093750, time 220.455 samples/sec
+epoch 1, total_step 16140, total loss is 32.46 , inference loss is 19.20, weight deacy loss is 13.26, training accuracy is 0.125000, time 223.146 samples/sec
+epoch 1, total_step 16160, total loss is 33.76 , inference loss is 20.50, weight deacy loss is 13.26, training accuracy is 0.000000, time 221.914 samples/sec
+epoch 1, total_step 16180, total loss is 32.45 , inference loss is 19.19, weight deacy loss is 13.26, training accuracy is 0.093750, time 299.403 samples/sec
+epoch 1, total_step 16200, total loss is 32.12 , inference loss is 18.86, weight deacy loss is 13.26, training accuracy is 0.093750, time 13.536 samples/sec
+epoch 1, total_step 16220, total loss is 35.79 , inference loss is 22.53, weight deacy loss is 13.26, training accuracy is 0.062500, time 223.359 samples/sec
+epoch 1, total_step 16240, total loss is 33.70 , inference loss is 20.45, weight deacy loss is 13.25, training accuracy is 0.031250, time 204.027 samples/sec
+epoch 1, total_step 16260, total loss is 33.38 , inference loss is 20.13, weight deacy loss is 13.25, training accuracy is 0.062500, time 212.010 samples/sec
+epoch 1, total_step 16280, total loss is 31.84 , inference loss is 18.59, weight deacy loss is 13.25, training accuracy is 0.125000, time 202.617 samples/sec
+epoch 1, total_step 16300, total loss is 32.79 , inference loss is 19.54, weight deacy loss is 13.25, training accuracy is 0.093750, time 223.857 samples/sec
+epoch 1, total_step 16320, total loss is 31.88 , inference loss is 18.63, weight deacy loss is 13.25, training accuracy is 0.125000, time 206.736 samples/sec
+epoch 1, total_step 16340, total loss is 33.32 , inference loss is 20.08, weight deacy loss is 13.24, training accuracy is 0.093750, time 237.420 samples/sec
+epoch 1, total_step 16360, total loss is 30.28 , inference loss is 17.04, weight deacy loss is 13.24, training accuracy is 0.093750, time 208.940 samples/sec
+epoch 1, total_step 16380, total loss is 35.16 , inference loss is 21.91, weight deacy loss is 13.24, training accuracy is 0.031250, time 220.732 samples/sec
+epoch 1, total_step 16400, total loss is 33.21 , inference loss is 19.97, weight deacy loss is 13.24, training accuracy is 0.250000, time 213.212 samples/sec
+epoch 1, total_step 16420, total loss is 35.46 , inference loss is 22.22, weight deacy loss is 13.24, training accuracy is 0.062500, time 220.465 samples/sec
+epoch 1, total_step 16440, total loss is 31.23 , inference loss is 17.99, weight deacy loss is 13.23, training accuracy is 0.156250, time 298.531 samples/sec
+epoch 1, total_step 16460, total loss is 35.31 , inference loss is 22.08, weight deacy loss is 13.23, training accuracy is 0.031250, time 211.192 samples/sec
+epoch 1, total_step 16480, total loss is 29.40 , inference loss is 16.17, weight deacy loss is 13.23, training accuracy is 0.156250, time 219.453 samples/sec
+epoch 1, total_step 16500, total loss is 35.02 , inference loss is 21.79, weight deacy loss is 13.23, training accuracy is 0.093750, time 11.307 samples/sec
+epoch 1, total_step 16520, total loss is 37.65 , inference loss is 24.43, weight deacy loss is 13.23, training accuracy is 0.062500, time 205.055 samples/sec
+epoch 1, total_step 16540, total loss is 38.35 , inference loss is 25.12, weight deacy loss is 13.23, training accuracy is 0.062500, time 225.190 samples/sec
+epoch 1, total_step 16560, total loss is 34.92 , inference loss is 21.70, weight deacy loss is 13.22, training accuracy is 0.093750, time 235.775 samples/sec
+epoch 1, total_step 16580, total loss is 39.13 , inference loss is 25.91, weight deacy loss is 13.22, training accuracy is 0.031250, time 203.792 samples/sec
+epoch 1, total_step 16600, total loss is 37.60 , inference loss is 24.38, weight deacy loss is 13.22, training accuracy is 0.031250, time 225.527 samples/sec
+epoch 1, total_step 16620, total loss is 35.46 , inference loss is 22.24, weight deacy loss is 13.22, training accuracy is 0.062500, time 217.761 samples/sec
+epoch 1, total_step 16640, total loss is 32.49 , inference loss is 19.27, weight deacy loss is 13.22, training accuracy is 0.187500, time 221.147 samples/sec
+epoch 1, total_step 16660, total loss is 33.01 , inference loss is 19.80, weight deacy loss is 13.21, training accuracy is 0.093750, time 212.396 samples/sec
+epoch 1, total_step 16680, total loss is 33.49 , inference loss is 20.28, weight deacy loss is 13.21, training accuracy is 0.093750, time 208.973 samples/sec
+epoch 1, total_step 16700, total loss is 35.45 , inference loss is 22.24, weight deacy loss is 13.21, training accuracy is 0.031250, time 294.394 samples/sec
+epoch 1, total_step 16720, total loss is 34.95 , inference loss is 21.74, weight deacy loss is 13.21, training accuracy is 0.062500, time 212.562 samples/sec
+epoch 1, total_step 16740, total loss is 31.88 , inference loss is 18.67, weight deacy loss is 13.21, training accuracy is 0.125000, time 221.841 samples/sec
+epoch 1, total_step 16760, total loss is 32.16 , inference loss is 18.95, weight deacy loss is 13.20, training accuracy is 0.031250, time 207.726 samples/sec
+epoch 1, total_step 16780, total loss is 35.66 , inference loss is 22.46, weight deacy loss is 13.20, training accuracy is 0.125000, time 225.191 samples/sec
+epoch 1, total_step 16800, total loss is 35.99 , inference loss is 22.79, weight deacy loss is 13.20, training accuracy is 0.093750, time 14.663 samples/sec
+epoch 1, total_step 16820, total loss is 35.45 , inference loss is 22.26, weight deacy loss is 13.20, training accuracy is 0.062500, time 198.323 samples/sec
+epoch 1, total_step 16840, total loss is 35.61 , inference loss is 22.41, weight deacy loss is 13.20, training accuracy is 0.031250, time 214.046 samples/sec
+epoch 1, total_step 16860, total loss is 33.85 , inference loss is 20.66, weight deacy loss is 13.19, training accuracy is 0.093750, time 218.621 samples/sec
+epoch 1, total_step 16880, total loss is 33.73 , inference loss is 20.53, weight deacy loss is 13.19, training accuracy is 0.125000, time 218.671 samples/sec
+epoch 1, total_step 16900, total loss is 34.25 , inference loss is 21.05, weight deacy loss is 13.19, training accuracy is 0.093750, time 211.460 samples/sec
+epoch 1, total_step 16920, total loss is 35.51 , inference loss is 22.32, weight deacy loss is 13.19, training accuracy is 0.031250, time 198.232 samples/sec
+epoch 1, total_step 16940, total loss is 35.54 , inference loss is 22.35, weight deacy loss is 13.19, training accuracy is 0.062500, time 233.410 samples/sec
+epoch 1, total_step 16960, total loss is 33.20 , inference loss is 20.02, weight deacy loss is 13.19, training accuracy is 0.125000, time 299.086 samples/sec
+epoch 1, total_step 16980, total loss is 29.16 , inference loss is 15.98, weight deacy loss is 13.18, training accuracy is 0.125000, time 224.728 samples/sec
+epoch 1, total_step 17000, total loss is 37.28 , inference loss is 24.10, weight deacy loss is 13.18, training accuracy is 0.093750, time 227.878 samples/sec
+epoch 1, total_step 17020, total loss is 35.28 , inference loss is 22.10, weight deacy loss is 13.18, training accuracy is 0.156250, time 209.201 samples/sec
+epoch 1, total_step 17040, total loss is 31.90 , inference loss is 18.73, weight deacy loss is 13.18, training accuracy is 0.125000, time 211.868 samples/sec
+epoch 1, total_step 17060, total loss is 32.68 , inference loss is 19.50, weight deacy loss is 13.18, training accuracy is 0.062500, time 220.948 samples/sec
+epoch 1, total_step 17080, total loss is 36.34 , inference loss is 23.17, weight deacy loss is 13.17, training accuracy is 0.031250, time 214.921 samples/sec
+epoch 1, total_step 17100, total loss is 34.69 , inference loss is 21.52, weight deacy loss is 13.17, training accuracy is 0.093750, time 13.721 samples/sec
+epoch 1, total_step 17120, total loss is 32.80 , inference loss is 19.63, weight deacy loss is 13.17, training accuracy is 0.187500, time 211.019 samples/sec
+epoch 1, total_step 17140, total loss is 29.38 , inference loss is 16.21, weight deacy loss is 13.17, training accuracy is 0.187500, time 197.484 samples/sec
+epoch 1, total_step 17160, total loss is 32.73 , inference loss is 19.57, weight deacy loss is 13.17, training accuracy is 0.125000, time 209.988 samples/sec
+epoch 1, total_step 17180, total loss is 32.66 , inference loss is 19.50, weight deacy loss is 13.16, training accuracy is 0.156250, time 206.336 samples/sec
+epoch 1, total_step 17200, total loss is 31.69 , inference loss is 18.53, weight deacy loss is 13.16, training accuracy is 0.187500, time 220.791 samples/sec
+epoch 1, total_step 17220, total loss is 29.41 , inference loss is 16.25, weight deacy loss is 13.16, training accuracy is 0.093750, time 207.516 samples/sec
+epoch 1, total_step 17240, total loss is 37.28 , inference loss is 24.12, weight deacy loss is 13.16, training accuracy is 0.031250, time 296.542 samples/sec
+epoch 1, total_step 17260, total loss is 30.75 , inference loss is 17.59, weight deacy loss is 13.16, training accuracy is 0.125000, time 203.746 samples/sec
+epoch 1, total_step 17280, total loss is 29.63 , inference loss is 16.48, weight deacy loss is 13.15, training accuracy is 0.218750, time 238.575 samples/sec
+epoch 1, total_step 17300, total loss is 30.02 , inference loss is 16.87, weight deacy loss is 13.15, training accuracy is 0.125000, time 218.635 samples/sec
+epoch 1, total_step 17320, total loss is 30.65 , inference loss is 17.50, weight deacy loss is 13.15, training accuracy is 0.187500, time 215.702 samples/sec
+epoch 1, total_step 17340, total loss is 31.90 , inference loss is 18.75, weight deacy loss is 13.15, training accuracy is 0.125000, time 210.024 samples/sec
+epoch 1, total_step 17360, total loss is 27.89 , inference loss is 14.75, weight deacy loss is 13.15, training accuracy is 0.187500, time 205.561 samples/sec
+epoch 1, total_step 17380, total loss is 32.06 , inference loss is 18.91, weight deacy loss is 13.14, training accuracy is 0.062500, time 212.056 samples/sec
+epoch 1, total_step 17400, total loss is 31.36 , inference loss is 18.22, weight deacy loss is 13.14, training accuracy is 0.156250, time 13.271 samples/sec
+epoch 1, total_step 17420, total loss is 32.64 , inference loss is 19.50, weight deacy loss is 13.14, training accuracy is 0.156250, time 220.461 samples/sec
+epoch 1, total_step 17440, total loss is 33.91 , inference loss is 20.77, weight deacy loss is 13.14, training accuracy is 0.031250, time 224.261 samples/sec
+epoch 1, total_step 17460, total loss is 32.67 , inference loss is 19.54, weight deacy loss is 13.14, training accuracy is 0.062500, time 222.509 samples/sec
+epoch 1, total_step 17480, total loss is 29.85 , inference loss is 16.71, weight deacy loss is 13.14, training accuracy is 0.156250, time 231.293 samples/sec
+epoch 1, total_step 17500, total loss is 28.92 , inference loss is 15.79, weight deacy loss is 13.13, training accuracy is 0.125000, time 301.447 samples/sec
+epoch 1, total_step 17520, total loss is 24.43 , inference loss is 11.30, weight deacy loss is 13.13, training accuracy is 0.250000, time 222.794 samples/sec
+epoch 1, total_step 17540, total loss is 30.56 , inference loss is 17.43, weight deacy loss is 13.13, training accuracy is 0.250000, time 227.334 samples/sec
+epoch 1, total_step 17560, total loss is 30.23 , inference loss is 17.11, weight deacy loss is 13.13, training accuracy is 0.187500, time 215.644 samples/sec
+epoch 1, total_step 17580, total loss is 29.56 , inference loss is 16.44, weight deacy loss is 13.13, training accuracy is 0.187500, time 222.487 samples/sec
+epoch 1, total_step 17600, total loss is 27.36 , inference loss is 14.23, weight deacy loss is 13.12, training accuracy is 0.093750, time 210.745 samples/sec
+epoch 1, total_step 17620, total loss is 26.87 , inference loss is 13.75, weight deacy loss is 13.12, training accuracy is 0.187500, time 226.844 samples/sec
+epoch 1, total_step 17640, total loss is 34.07 , inference loss is 20.95, weight deacy loss is 13.12, training accuracy is 0.093750, time 223.053 samples/sec
+epoch 1, total_step 17660, total loss is 29.00 , inference loss is 15.89, weight deacy loss is 13.12, training accuracy is 0.156250, time 189.138 samples/sec
+epoch 1, total_step 17680, total loss is 30.94 , inference loss is 17.82, weight deacy loss is 13.12, training accuracy is 0.093750, time 210.925 samples/sec
+epoch 1, total_step 17700, total loss is 33.43 , inference loss is 20.32, weight deacy loss is 13.11, training accuracy is 0.031250, time 14.776 samples/sec
+epoch 1, total_step 17720, total loss is 29.24 , inference loss is 16.13, weight deacy loss is 13.11, training accuracy is 0.187500, time 224.713 samples/sec
+epoch 1, total_step 17740, total loss is 33.22 , inference loss is 20.11, weight deacy loss is 13.11, training accuracy is 0.187500, time 212.695 samples/sec
+epoch 1, total_step 17760, total loss is 25.89 , inference loss is 12.78, weight deacy loss is 13.11, training accuracy is 0.250000, time 218.631 samples/sec
+epoch 1, total_step 17780, total loss is 32.26 , inference loss is 19.15, weight deacy loss is 13.11, training accuracy is 0.125000, time 215.033 samples/sec
+epoch 1, total_step 17800, total loss is 30.09 , inference loss is 16.98, weight deacy loss is 13.10, training accuracy is 0.062500, time 224.624 samples/sec
+epoch 1, total_step 17820, total loss is 30.38 , inference loss is 17.28, weight deacy loss is 13.10, training accuracy is 0.093750, time 296.523 samples/sec
+epoch 1, total_step 17840, total loss is 31.83 , inference loss is 18.73, weight deacy loss is 13.10, training accuracy is 0.093750, time 215.483 samples/sec
+epoch 1, total_step 17860, total loss is 27.53 , inference loss is 14.43, weight deacy loss is 13.10, training accuracy is 0.125000, time 223.307 samples/sec
+epoch 1, total_step 17880, total loss is 30.08 , inference loss is 16.98, weight deacy loss is 13.10, training accuracy is 0.125000, time 204.382 samples/sec
+epoch 1, total_step 17900, total loss is 30.03 , inference loss is 16.94, weight deacy loss is 13.09, training accuracy is 0.093750, time 211.301 samples/sec
+epoch 1, total_step 17920, total loss is 29.45 , inference loss is 16.36, weight deacy loss is 13.09, training accuracy is 0.062500, time 216.942 samples/sec
+epoch 1, total_step 17940, total loss is 26.55 , inference loss is 13.45, weight deacy loss is 13.09, training accuracy is 0.281250, time 298.411 samples/sec
+epoch 1, total_step 17960, total loss is 28.76 , inference loss is 15.67, weight deacy loss is 13.09, training accuracy is 0.250000, time 218.965 samples/sec
+epoch 1, total_step 17980, total loss is 25.48 , inference loss is 12.39, weight deacy loss is 13.09, training accuracy is 0.218750, time 217.743 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.763921999999994
+best_threshold_index 105 0.9587037037037037
+best_threshold_index 101 0.9598148148148148
+best_threshold_index 105 0.9598148148148148
+best_threshold_index 102 0.9598148148148148
+best_threshold_index 101 0.960925925925926
+best_threshold_index 104 0.9594444444444444
+best_threshold_index 105 0.9585185185185185
+best_threshold_index 102 0.9587037037037037
+best_threshold_index 101 0.9568518518518518
+best_threshold_index 105 0.9581481481481482
+[lfw][18000]XNorm: 24.336557
+[lfw][18000]Accuracy-Flip: 0.95717+-0.00997
+test accuracy is: 0.9571666666666667
+epoch 1, total_step 18000, total loss is 28.34 , inference loss is 15.26, weight deacy loss is 13.09, training accuracy is 0.187500, time 17.224 samples/sec
+epoch 1, total_step 18020, total loss is 27.79 , inference loss is 14.70, weight deacy loss is 13.08, training accuracy is 0.250000, time 296.239 samples/sec
+epoch 1, total_step 18040, total loss is 32.74 , inference loss is 19.66, weight deacy loss is 13.08, training accuracy is 0.156250, time 211.569 samples/sec
+epoch 1, total_step 18060, total loss is 34.46 , inference loss is 21.38, weight deacy loss is 13.08, training accuracy is 0.062500, time 228.442 samples/sec
+epoch 1, total_step 18080, total loss is 31.07 , inference loss is 17.99, weight deacy loss is 13.08, training accuracy is 0.093750, time 218.223 samples/sec
+epoch 1, total_step 18100, total loss is 29.94 , inference loss is 16.87, weight deacy loss is 13.08, training accuracy is 0.062500, time 237.720 samples/sec
+epoch 1, total_step 18120, total loss is 32.48 , inference loss is 19.41, weight deacy loss is 13.07, training accuracy is 0.156250, time 223.268 samples/sec
+epoch 1, total_step 18140, total loss is 29.82 , inference loss is 16.75, weight deacy loss is 13.07, training accuracy is 0.093750, time 227.359 samples/sec
+epoch 1, total_step 18160, total loss is 28.18 , inference loss is 15.11, weight deacy loss is 13.07, training accuracy is 0.187500, time 206.859 samples/sec
+epoch 1, total_step 18180, total loss is 31.24 , inference loss is 18.17, weight deacy loss is 13.07, training accuracy is 0.062500, time 217.803 samples/sec
+epoch 1, total_step 18200, total loss is 31.60 , inference loss is 18.53, weight deacy loss is 13.07, training accuracy is 0.125000, time 233.869 samples/sec
+epoch 1, total_step 18220, total loss is 28.84 , inference loss is 15.77, weight deacy loss is 13.06, training accuracy is 0.156250, time 215.923 samples/sec
+epoch 1, total_step 18240, total loss is 30.81 , inference loss is 17.75, weight deacy loss is 13.06, training accuracy is 0.156250, time 224.742 samples/sec
+epoch 1, total_step 18260, total loss is 26.89 , inference loss is 13.83, weight deacy loss is 13.06, training accuracy is 0.093750, time 211.681 samples/sec
+epoch 1, total_step 18280, total loss is 27.71 , inference loss is 14.65, weight deacy loss is 13.06, training accuracy is 0.156250, time 210.667 samples/sec
+epoch 1, total_step 18300, total loss is 35.05 , inference loss is 22.00, weight deacy loss is 13.06, training accuracy is 0.031250, time 13.570 samples/sec
+epoch 1, total_step 18320, total loss is 28.91 , inference loss is 15.85, weight deacy loss is 13.05, training accuracy is 0.156250, time 210.129 samples/sec
+epoch 1, total_step 18340, total loss is 30.29 , inference loss is 17.24, weight deacy loss is 13.05, training accuracy is 0.156250, time 232.431 samples/sec
+epoch 1, total_step 18360, total loss is 28.78 , inference loss is 15.73, weight deacy loss is 13.05, training accuracy is 0.281250, time 225.573 samples/sec
+epoch 1, total_step 18380, total loss is 32.22 , inference loss is 19.17, weight deacy loss is 13.05, training accuracy is 0.125000, time 225.556 samples/sec
+epoch 1, total_step 18400, total loss is 28.50 , inference loss is 15.45, weight deacy loss is 13.05, training accuracy is 0.187500, time 216.435 samples/sec
+epoch 1, total_step 18420, total loss is 30.25 , inference loss is 17.20, weight deacy loss is 13.05, training accuracy is 0.125000, time 220.570 samples/sec
+epoch 1, total_step 18440, total loss is 32.62 , inference loss is 19.58, weight deacy loss is 13.04, training accuracy is 0.125000, time 220.941 samples/sec
+epoch 1, total_step 18460, total loss is 30.40 , inference loss is 17.36, weight deacy loss is 13.04, training accuracy is 0.093750, time 220.224 samples/sec
+epoch 1, total_step 18480, total loss is 32.68 , inference loss is 19.64, weight deacy loss is 13.04, training accuracy is 0.093750, time 224.134 samples/sec
+epoch 1, total_step 18500, total loss is 31.48 , inference loss is 18.45, weight deacy loss is 13.04, training accuracy is 0.062500, time 213.447 samples/sec
+epoch 1, total_step 18520, total loss is 34.77 , inference loss is 21.73, weight deacy loss is 13.04, training accuracy is 0.125000, time 198.117 samples/sec
+epoch 1, total_step 18540, total loss is 29.55 , inference loss is 16.51, weight deacy loss is 13.03, training accuracy is 0.218750, time 221.376 samples/sec
+epoch 1, total_step 18560, total loss is 32.65 , inference loss is 19.62, weight deacy loss is 13.03, training accuracy is 0.062500, time 227.957 samples/sec
+epoch 1, total_step 18580, total loss is 32.53 , inference loss is 19.50, weight deacy loss is 13.03, training accuracy is 0.062500, time 210.570 samples/sec
+epoch 1, total_step 18600, total loss is 26.40 , inference loss is 13.37, weight deacy loss is 13.03, training accuracy is 0.156250, time 14.457 samples/sec
+epoch 1, total_step 18620, total loss is 32.87 , inference loss is 19.85, weight deacy loss is 13.03, training accuracy is 0.062500, time 229.425 samples/sec
+epoch 1, total_step 18640, total loss is 39.04 , inference loss is 26.01, weight deacy loss is 13.03, training accuracy is 0.000000, time 214.300 samples/sec
+epoch 1, total_step 18660, total loss is 28.47 , inference loss is 15.45, weight deacy loss is 13.02, training accuracy is 0.093750, time 217.408 samples/sec
+epoch 1, total_step 18680, total loss is 36.01 , inference loss is 22.99, weight deacy loss is 13.02, training accuracy is 0.125000, time 218.031 samples/sec
+epoch 1, total_step 18700, total loss is 32.27 , inference loss is 19.25, weight deacy loss is 13.02, training accuracy is 0.031250, time 211.112 samples/sec
+epoch 1, total_step 18720, total loss is 30.21 , inference loss is 17.19, weight deacy loss is 13.02, training accuracy is 0.125000, time 222.070 samples/sec
+epoch 1, total_step 18740, total loss is 31.15 , inference loss is 18.13, weight deacy loss is 13.02, training accuracy is 0.125000, time 212.419 samples/sec
+epoch 1, total_step 18760, total loss is 32.72 , inference loss is 19.71, weight deacy loss is 13.01, training accuracy is 0.156250, time 208.841 samples/sec
+epoch 1, total_step 18780, total loss is 29.88 , inference loss is 16.87, weight deacy loss is 13.01, training accuracy is 0.250000, time 223.243 samples/sec
+epoch 1, total_step 18800, total loss is 30.82 , inference loss is 17.81, weight deacy loss is 13.01, training accuracy is 0.062500, time 209.438 samples/sec
+epoch 1, total_step 18820, total loss is 30.44 , inference loss is 17.44, weight deacy loss is 13.01, training accuracy is 0.031250, time 231.078 samples/sec
+epoch 1, total_step 18840, total loss is 35.23 , inference loss is 22.22, weight deacy loss is 13.01, training accuracy is 0.031250, time 237.892 samples/sec
+epoch 1, total_step 18860, total loss is 30.81 , inference loss is 17.80, weight deacy loss is 13.00, training accuracy is 0.156250, time 298.036 samples/sec
+epoch 1, total_step 18880, total loss is 35.75 , inference loss is 22.75, weight deacy loss is 13.00, training accuracy is 0.062500, time 207.048 samples/sec
+epoch 1, total_step 18900, total loss is 29.60 , inference loss is 16.60, weight deacy loss is 13.00, training accuracy is 0.125000, time 14.066 samples/sec
+epoch 1, total_step 18920, total loss is 34.73 , inference loss is 21.73, weight deacy loss is 13.00, training accuracy is 0.187500, time 211.912 samples/sec
+epoch 1, total_step 18940, total loss is 32.05 , inference loss is 19.05, weight deacy loss is 13.00, training accuracy is 0.125000, time 232.660 samples/sec
+epoch 1, total_step 18960, total loss is 32.09 , inference loss is 19.10, weight deacy loss is 13.00, training accuracy is 0.093750, time 210.970 samples/sec
+epoch 1, total_step 18980, total loss is 33.04 , inference loss is 20.05, weight deacy loss is 12.99, training accuracy is 0.031250, time 235.915 samples/sec
+epoch 1, total_step 19000, total loss is 32.55 , inference loss is 19.56, weight deacy loss is 12.99, training accuracy is 0.125000, time 187.352 samples/sec
+epoch 1, total_step 19020, total loss is 31.56 , inference loss is 18.57, weight deacy loss is 12.99, training accuracy is 0.093750, time 216.003 samples/sec
+epoch 1, total_step 19040, total loss is 32.00 , inference loss is 19.01, weight deacy loss is 12.99, training accuracy is 0.093750, time 209.734 samples/sec
+epoch 1, total_step 19060, total loss is 33.37 , inference loss is 20.38, weight deacy loss is 12.99, training accuracy is 0.125000, time 210.135 samples/sec
+epoch 1, total_step 19080, total loss is 34.86 , inference loss is 21.87, weight deacy loss is 12.98, training accuracy is 0.062500, time 224.997 samples/sec
+epoch 1, total_step 19100, total loss is 28.88 , inference loss is 15.89, weight deacy loss is 12.98, training accuracy is 0.187500, time 223.602 samples/sec
+epoch 1, total_step 19120, total loss is 30.08 , inference loss is 17.10, weight deacy loss is 12.98, training accuracy is 0.218750, time 184.819 samples/sec
+epoch 1, total_step 19140, total loss is 30.94 , inference loss is 17.96, weight deacy loss is 12.98, training accuracy is 0.125000, time 221.149 samples/sec
+epoch 1, total_step 19160, total loss is 35.46 , inference loss is 22.48, weight deacy loss is 12.98, training accuracy is 0.062500, time 207.775 samples/sec
+epoch 1, total_step 19180, total loss is 30.88 , inference loss is 17.90, weight deacy loss is 12.98, training accuracy is 0.093750, time 206.669 samples/sec
+epoch 1, total_step 19200, total loss is 30.88 , inference loss is 17.91, weight deacy loss is 12.97, training accuracy is 0.156250, time 14.155 samples/sec
+epoch 1, total_step 19220, total loss is 29.83 , inference loss is 16.86, weight deacy loss is 12.97, training accuracy is 0.156250, time 228.017 samples/sec
+epoch 1, total_step 19240, total loss is 34.78 , inference loss is 21.81, weight deacy loss is 12.97, training accuracy is 0.000000, time 225.171 samples/sec
+epoch 1, total_step 19260, total loss is 29.85 , inference loss is 16.88, weight deacy loss is 12.97, training accuracy is 0.031250, time 210.890 samples/sec
+epoch 1, total_step 19280, total loss is 34.69 , inference loss is 21.72, weight deacy loss is 12.97, training accuracy is 0.031250, time 213.268 samples/sec
+epoch 1, total_step 19300, total loss is 35.89 , inference loss is 22.92, weight deacy loss is 12.96, training accuracy is 0.031250, time 218.455 samples/sec
+epoch 1, total_step 19320, total loss is 30.55 , inference loss is 17.59, weight deacy loss is 12.96, training accuracy is 0.062500, time 216.764 samples/sec
+epoch 1, total_step 19340, total loss is 35.02 , inference loss is 22.06, weight deacy loss is 12.96, training accuracy is 0.031250, time 219.702 samples/sec
+epoch 1, total_step 19360, total loss is 31.46 , inference loss is 18.50, weight deacy loss is 12.96, training accuracy is 0.093750, time 205.254 samples/sec
+epoch 1, total_step 19380, total loss is 33.27 , inference loss is 20.31, weight deacy loss is 12.96, training accuracy is 0.125000, time 217.134 samples/sec
+epoch 1, total_step 19400, total loss is 34.06 , inference loss is 21.11, weight deacy loss is 12.96, training accuracy is 0.093750, time 299.018 samples/sec
+epoch 1, total_step 19420, total loss is 32.56 , inference loss is 19.61, weight deacy loss is 12.95, training accuracy is 0.156250, time 217.798 samples/sec
+epoch 1, total_step 19440, total loss is 36.66 , inference loss is 23.70, weight deacy loss is 12.95, training accuracy is 0.062500, time 219.391 samples/sec
+epoch 1, total_step 19460, total loss is 35.71 , inference loss is 22.76, weight deacy loss is 12.95, training accuracy is 0.062500, time 233.203 samples/sec
+epoch 1, total_step 19480, total loss is 32.95 , inference loss is 20.00, weight deacy loss is 12.95, training accuracy is 0.156250, time 213.045 samples/sec
+epoch 1, total_step 19500, total loss is 33.67 , inference loss is 20.73, weight deacy loss is 12.95, training accuracy is 0.062500, time 14.050 samples/sec
+epoch 1, total_step 19520, total loss is 28.81 , inference loss is 15.86, weight deacy loss is 12.94, training accuracy is 0.125000, time 213.838 samples/sec
+epoch 1, total_step 19540, total loss is 33.78 , inference loss is 20.83, weight deacy loss is 12.94, training accuracy is 0.062500, time 218.659 samples/sec
+epoch 1, total_step 19560, total loss is 33.69 , inference loss is 20.75, weight deacy loss is 12.94, training accuracy is 0.031250, time 216.405 samples/sec
+epoch 1, total_step 19580, total loss is 29.75 , inference loss is 16.81, weight deacy loss is 12.94, training accuracy is 0.093750, time 224.354 samples/sec
+epoch 1, total_step 19600, total loss is 28.46 , inference loss is 15.52, weight deacy loss is 12.94, training accuracy is 0.062500, time 296.301 samples/sec
+epoch 1, total_step 19620, total loss is 35.87 , inference loss is 22.94, weight deacy loss is 12.94, training accuracy is 0.093750, time 298.608 samples/sec
+epoch 1, total_step 19640, total loss is 27.02 , inference loss is 14.08, weight deacy loss is 12.93, training accuracy is 0.093750, time 222.568 samples/sec
+epoch 1, total_step 19660, total loss is 35.00 , inference loss is 22.07, weight deacy loss is 12.93, training accuracy is 0.031250, time 236.438 samples/sec
+epoch 1, total_step 19680, total loss is 33.34 , inference loss is 20.41, weight deacy loss is 12.93, training accuracy is 0.062500, time 224.253 samples/sec
+epoch 1, total_step 19700, total loss is 30.07 , inference loss is 17.14, weight deacy loss is 12.93, training accuracy is 0.125000, time 216.101 samples/sec
+epoch 1, total_step 19720, total loss is 31.30 , inference loss is 18.37, weight deacy loss is 12.93, training accuracy is 0.062500, time 229.968 samples/sec
+epoch 1, total_step 19740, total loss is 32.63 , inference loss is 19.70, weight deacy loss is 12.92, training accuracy is 0.125000, time 238.177 samples/sec
+epoch 1, total_step 19760, total loss is 29.44 , inference loss is 16.52, weight deacy loss is 12.92, training accuracy is 0.125000, time 219.899 samples/sec
+epoch 1, total_step 19780, total loss is 35.32 , inference loss is 22.39, weight deacy loss is 12.92, training accuracy is 0.062500, time 226.464 samples/sec
+epoch 1, total_step 19800, total loss is 29.61 , inference loss is 16.69, weight deacy loss is 12.92, training accuracy is 0.187500, time 15.402 samples/sec
+epoch 1, total_step 19820, total loss is 30.52 , inference loss is 17.61, weight deacy loss is 12.92, training accuracy is 0.093750, time 227.272 samples/sec
+epoch 1, total_step 19840, total loss is 29.05 , inference loss is 16.14, weight deacy loss is 12.92, training accuracy is 0.031250, time 214.284 samples/sec
+epoch 1, total_step 19860, total loss is 36.18 , inference loss is 23.26, weight deacy loss is 12.91, training accuracy is 0.125000, time 226.063 samples/sec
+epoch 1, total_step 19880, total loss is 31.78 , inference loss is 18.87, weight deacy loss is 12.91, training accuracy is 0.093750, time 223.226 samples/sec
+epoch 1, total_step 19900, total loss is 36.52 , inference loss is 23.61, weight deacy loss is 12.91, training accuracy is 0.031250, time 222.652 samples/sec
+epoch 1, total_step 19920, total loss is 27.97 , inference loss is 15.06, weight deacy loss is 12.91, training accuracy is 0.093750, time 208.627 samples/sec
+epoch 1, total_step 19940, total loss is 28.82 , inference loss is 15.92, weight deacy loss is 12.91, training accuracy is 0.062500, time 216.568 samples/sec
+epoch 1, total_step 19960, total loss is 32.37 , inference loss is 19.46, weight deacy loss is 12.91, training accuracy is 0.062500, time 217.711 samples/sec
+epoch 1, total_step 19980, total loss is 31.14 , inference loss is 18.24, weight deacy loss is 12.90, training accuracy is 0.156250, time 219.775 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.618454000000025
+best_threshold_index 117 0.9679629629629629
+best_threshold_index 116 0.9655555555555555
+best_threshold_index 116 0.9675925925925926
+best_threshold_index 111 0.9664814814814815
+best_threshold_index 111 0.9696296296296296
+best_threshold_index 116 0.9664814814814815
+best_threshold_index 117 0.9668518518518519
+best_threshold_index 116 0.9666666666666667
+best_threshold_index 116 0.9644444444444444
+best_threshold_index 117 0.9664814814814815
+[lfw][20000]XNorm: 22.208202
+[lfw][20000]Accuracy-Flip: 0.96467+-0.01256
+test accuracy is: 0.9646666666666667
+epoch 1, total_step 20000, total loss is 31.35 , inference loss is 18.45, weight deacy loss is 12.90, training accuracy is 0.156250, time 14.092 samples/sec
+epoch 1, total_step 20020, total loss is 30.23 , inference loss is 17.33, weight deacy loss is 12.90, training accuracy is 0.062500, time 209.742 samples/sec
+epoch 1, total_step 20040, total loss is 29.92 , inference loss is 17.02, weight deacy loss is 12.90, training accuracy is 0.156250, time 209.293 samples/sec
+epoch 1, total_step 20060, total loss is 34.09 , inference loss is 21.19, weight deacy loss is 12.90, training accuracy is 0.000000, time 228.334 samples/sec
+epoch 1, total_step 20080, total loss is 29.39 , inference loss is 16.49, weight deacy loss is 12.89, training accuracy is 0.093750, time 204.760 samples/sec
+epoch 1, total_step 20100, total loss is 31.47 , inference loss is 18.58, weight deacy loss is 12.89, training accuracy is 0.125000, time 14.494 samples/sec
+epoch 1, total_step 20120, total loss is 35.26 , inference loss is 22.37, weight deacy loss is 12.89, training accuracy is 0.031250, time 188.754 samples/sec
+epoch 1, total_step 20140, total loss is 31.39 , inference loss is 18.50, weight deacy loss is 12.89, training accuracy is 0.062500, time 223.872 samples/sec
+epoch 1, total_step 20160, total loss is 32.30 , inference loss is 19.41, weight deacy loss is 12.89, training accuracy is 0.125000, time 225.467 samples/sec
+epoch 1, total_step 20180, total loss is 29.49 , inference loss is 16.60, weight deacy loss is 12.89, training accuracy is 0.125000, time 224.502 samples/sec
+epoch 1, total_step 20200, total loss is 35.37 , inference loss is 22.49, weight deacy loss is 12.88, training accuracy is 0.093750, time 225.746 samples/sec
+epoch 1, total_step 20220, total loss is 27.02 , inference loss is 14.13, weight deacy loss is 12.88, training accuracy is 0.093750, time 206.378 samples/sec
+epoch 1, total_step 20240, total loss is 31.43 , inference loss is 18.55, weight deacy loss is 12.88, training accuracy is 0.125000, time 204.315 samples/sec
+epoch 1, total_step 20260, total loss is 34.86 , inference loss is 21.99, weight deacy loss is 12.88, training accuracy is 0.093750, time 228.710 samples/sec
+epoch 1, total_step 20280, total loss is 31.21 , inference loss is 18.34, weight deacy loss is 12.88, training accuracy is 0.156250, time 213.366 samples/sec
+epoch 1, total_step 20300, total loss is 33.38 , inference loss is 20.51, weight deacy loss is 12.87, training accuracy is 0.156250, time 215.997 samples/sec
+epoch 1, total_step 20320, total loss is 33.57 , inference loss is 20.70, weight deacy loss is 12.87, training accuracy is 0.031250, time 222.719 samples/sec
+epoch 1, total_step 20340, total loss is 31.52 , inference loss is 18.65, weight deacy loss is 12.87, training accuracy is 0.093750, time 236.699 samples/sec
+epoch 1, total_step 20360, total loss is 34.27 , inference loss is 21.40, weight deacy loss is 12.87, training accuracy is 0.125000, time 207.172 samples/sec
+epoch 1, total_step 20380, total loss is 34.53 , inference loss is 21.66, weight deacy loss is 12.87, training accuracy is 0.125000, time 217.164 samples/sec
+epoch 1, total_step 20400, total loss is 27.31 , inference loss is 14.45, weight deacy loss is 12.87, training accuracy is 0.125000, time 13.478 samples/sec
+epoch 1, total_step 20420, total loss is 33.02 , inference loss is 20.15, weight deacy loss is 12.86, training accuracy is 0.125000, time 228.500 samples/sec
+epoch 1, total_step 20440, total loss is 31.33 , inference loss is 18.47, weight deacy loss is 12.86, training accuracy is 0.093750, time 209.241 samples/sec
+epoch 1, total_step 20460, total loss is 33.88 , inference loss is 21.02, weight deacy loss is 12.86, training accuracy is 0.062500, time 217.279 samples/sec
+epoch 1, total_step 20480, total loss is 32.95 , inference loss is 20.09, weight deacy loss is 12.86, training accuracy is 0.062500, time 216.913 samples/sec
+epoch 1, total_step 20500, total loss is 32.72 , inference loss is 19.86, weight deacy loss is 12.86, training accuracy is 0.093750, time 207.609 samples/sec
+epoch 1, total_step 20520, total loss is 29.08 , inference loss is 16.23, weight deacy loss is 12.85, training accuracy is 0.031250, time 217.860 samples/sec
+epoch 1, total_step 20540, total loss is 35.33 , inference loss is 22.48, weight deacy loss is 12.85, training accuracy is 0.156250, time 204.769 samples/sec
+epoch 1, total_step 20560, total loss is 34.18 , inference loss is 21.33, weight deacy loss is 12.85, training accuracy is 0.125000, time 243.519 samples/sec
+epoch 1, total_step 20580, total loss is 28.85 , inference loss is 16.00, weight deacy loss is 12.85, training accuracy is 0.156250, time 215.914 samples/sec
+epoch 1, total_step 20600, total loss is 29.20 , inference loss is 16.35, weight deacy loss is 12.85, training accuracy is 0.093750, time 201.183 samples/sec
+epoch 1, total_step 20620, total loss is 31.84 , inference loss is 18.99, weight deacy loss is 12.85, training accuracy is 0.125000, time 214.726 samples/sec
+epoch 1, total_step 20640, total loss is 28.12 , inference loss is 15.28, weight deacy loss is 12.84, training accuracy is 0.093750, time 220.088 samples/sec
+epoch 1, total_step 20660, total loss is 31.99 , inference loss is 19.15, weight deacy loss is 12.84, training accuracy is 0.062500, time 214.374 samples/sec
+epoch 1, total_step 20680, total loss is 30.30 , inference loss is 17.46, weight deacy loss is 12.84, training accuracy is 0.125000, time 221.438 samples/sec
+epoch 1, total_step 20700, total loss is 29.83 , inference loss is 16.99, weight deacy loss is 12.84, training accuracy is 0.125000, time 13.858 samples/sec
+epoch 1, total_step 20720, total loss is 31.19 , inference loss is 18.35, weight deacy loss is 12.84, training accuracy is 0.031250, time 202.437 samples/sec
+epoch 1, total_step 20740, total loss is 32.01 , inference loss is 19.18, weight deacy loss is 12.84, training accuracy is 0.093750, time 215.824 samples/sec
+epoch 1, total_step 20760, total loss is 30.92 , inference loss is 18.08, weight deacy loss is 12.83, training accuracy is 0.093750, time 238.162 samples/sec
+epoch 1, total_step 20780, total loss is 26.92 , inference loss is 14.09, weight deacy loss is 12.83, training accuracy is 0.156250, time 223.475 samples/sec
+epoch 1, total_step 20800, total loss is 34.21 , inference loss is 21.38, weight deacy loss is 12.83, training accuracy is 0.093750, time 221.015 samples/sec
+epoch 1, total_step 20820, total loss is 29.73 , inference loss is 16.90, weight deacy loss is 12.83, training accuracy is 0.156250, time 224.509 samples/sec
+epoch 1, total_step 20840, total loss is 26.80 , inference loss is 13.98, weight deacy loss is 12.83, training accuracy is 0.281250, time 221.946 samples/sec
+epoch 1, total_step 20860, total loss is 31.74 , inference loss is 18.91, weight deacy loss is 12.82, training accuracy is 0.125000, time 210.463 samples/sec
+epoch 1, total_step 20880, total loss is 32.21 , inference loss is 19.39, weight deacy loss is 12.82, training accuracy is 0.125000, time 210.411 samples/sec
+epoch 1, total_step 20900, total loss is 31.47 , inference loss is 18.65, weight deacy loss is 12.82, training accuracy is 0.062500, time 207.192 samples/sec
+epoch 1, total_step 20920, total loss is 28.86 , inference loss is 16.04, weight deacy loss is 12.82, training accuracy is 0.218750, time 238.808 samples/sec
+epoch 1, total_step 20940, total loss is 32.20 , inference loss is 19.38, weight deacy loss is 12.82, training accuracy is 0.125000, time 225.214 samples/sec
+epoch 1, total_step 20960, total loss is 31.97 , inference loss is 19.16, weight deacy loss is 12.82, training accuracy is 0.125000, time 300.414 samples/sec
+epoch 1, total_step 20980, total loss is 29.80 , inference loss is 16.98, weight deacy loss is 12.81, training accuracy is 0.187500, time 205.002 samples/sec
+epoch 1, total_step 21000, total loss is 30.78 , inference loss is 17.97, weight deacy loss is 12.81, training accuracy is 0.000000, time 13.607 samples/sec
+epoch 1, total_step 21020, total loss is 31.24 , inference loss is 18.43, weight deacy loss is 12.81, training accuracy is 0.093750, time 247.174 samples/sec
+epoch 1, total_step 21040, total loss is 32.19 , inference loss is 19.39, weight deacy loss is 12.81, training accuracy is 0.000000, time 214.227 samples/sec
+epoch 1, total_step 21060, total loss is 31.78 , inference loss is 18.97, weight deacy loss is 12.81, training accuracy is 0.156250, time 231.754 samples/sec
+epoch 1, total_step 21080, total loss is 30.86 , inference loss is 18.06, weight deacy loss is 12.80, training accuracy is 0.093750, time 205.342 samples/sec
+epoch 1, total_step 21100, total loss is 33.34 , inference loss is 20.54, weight deacy loss is 12.80, training accuracy is 0.125000, time 212.852 samples/sec
+epoch 1, total_step 21120, total loss is 31.65 , inference loss is 18.85, weight deacy loss is 12.80, training accuracy is 0.156250, time 204.503 samples/sec
+epoch 1, total_step 21140, total loss is 24.63 , inference loss is 11.83, weight deacy loss is 12.80, training accuracy is 0.250000, time 224.235 samples/sec
+epoch 1, total_step 21160, total loss is 31.61 , inference loss is 18.81, weight deacy loss is 12.80, training accuracy is 0.312500, time 217.376 samples/sec
+epoch 1, total_step 21180, total loss is 28.86 , inference loss is 16.07, weight deacy loss is 12.80, training accuracy is 0.156250, time 223.078 samples/sec
+epoch 1, total_step 21200, total loss is 24.49 , inference loss is 11.70, weight deacy loss is 12.79, training accuracy is 0.218750, time 209.783 samples/sec
+epoch 1, total_step 21220, total loss is 32.29 , inference loss is 19.50, weight deacy loss is 12.79, training accuracy is 0.125000, time 299.276 samples/sec
+epoch 1, total_step 21240, total loss is 33.44 , inference loss is 20.65, weight deacy loss is 12.79, training accuracy is 0.000000, time 297.917 samples/sec
+epoch 1, total_step 21260, total loss is 31.42 , inference loss is 18.63, weight deacy loss is 12.79, training accuracy is 0.062500, time 298.479 samples/sec
+epoch 1, total_step 21280, total loss is 33.36 , inference loss is 20.58, weight deacy loss is 12.79, training accuracy is 0.125000, time 222.381 samples/sec
+epoch 1, total_step 21300, total loss is 31.73 , inference loss is 18.95, weight deacy loss is 12.79, training accuracy is 0.093750, time 14.766 samples/sec
+epoch 1, total_step 21320, total loss is 35.70 , inference loss is 22.92, weight deacy loss is 12.78, training accuracy is 0.093750, time 220.705 samples/sec
+epoch 1, total_step 21340, total loss is 32.44 , inference loss is 19.66, weight deacy loss is 12.78, training accuracy is 0.093750, time 208.690 samples/sec
+epoch 1, total_step 21360, total loss is 33.98 , inference loss is 21.20, weight deacy loss is 12.78, training accuracy is 0.093750, time 299.426 samples/sec
+epoch 1, total_step 21380, total loss is 29.78 , inference loss is 17.00, weight deacy loss is 12.78, training accuracy is 0.156250, time 294.654 samples/sec
+epoch 1, total_step 21400, total loss is 31.44 , inference loss is 18.66, weight deacy loss is 12.78, training accuracy is 0.031250, time 213.832 samples/sec
+epoch 1, total_step 21420, total loss is 32.73 , inference loss is 19.96, weight deacy loss is 12.77, training accuracy is 0.093750, time 224.738 samples/sec
+epoch 1, total_step 21440, total loss is 27.99 , inference loss is 15.22, weight deacy loss is 12.77, training accuracy is 0.156250, time 197.109 samples/sec
+epoch 1, total_step 21460, total loss is 29.96 , inference loss is 17.19, weight deacy loss is 12.77, training accuracy is 0.062500, time 225.300 samples/sec
+epoch 1, total_step 21480, total loss is 30.41 , inference loss is 17.64, weight deacy loss is 12.77, training accuracy is 0.187500, time 207.265 samples/sec
+epoch 1, total_step 21500, total loss is 28.96 , inference loss is 16.19, weight deacy loss is 12.77, training accuracy is 0.156250, time 211.530 samples/sec
+epoch 1, total_step 21520, total loss is 28.88 , inference loss is 16.11, weight deacy loss is 12.77, training accuracy is 0.062500, time 238.813 samples/sec
+epoch 1, total_step 21540, total loss is 27.23 , inference loss is 14.46, weight deacy loss is 12.76, training accuracy is 0.093750, time 221.376 samples/sec
+epoch 1, total_step 21560, total loss is 28.84 , inference loss is 16.08, weight deacy loss is 12.76, training accuracy is 0.125000, time 226.009 samples/sec
+epoch 1, total_step 21580, total loss is 31.45 , inference loss is 18.69, weight deacy loss is 12.76, training accuracy is 0.031250, time 212.425 samples/sec
+epoch 1, total_step 21600, total loss is 27.77 , inference loss is 15.01, weight deacy loss is 12.76, training accuracy is 0.187500, time 13.807 samples/sec
+epoch 1, total_step 21620, total loss is 30.46 , inference loss is 17.71, weight deacy loss is 12.76, training accuracy is 0.093750, time 223.689 samples/sec
+epoch 1, total_step 21640, total loss is 31.54 , inference loss is 18.79, weight deacy loss is 12.75, training accuracy is 0.125000, time 223.679 samples/sec
+epoch 1, total_step 21660, total loss is 32.66 , inference loss is 19.91, weight deacy loss is 12.75, training accuracy is 0.062500, time 298.741 samples/sec
+epoch 1, total_step 21680, total loss is 34.81 , inference loss is 22.06, weight deacy loss is 12.75, training accuracy is 0.062500, time 210.030 samples/sec
+epoch 1, total_step 21700, total loss is 32.00 , inference loss is 19.25, weight deacy loss is 12.75, training accuracy is 0.125000, time 239.585 samples/sec
+epoch 1, total_step 21720, total loss is 32.07 , inference loss is 19.32, weight deacy loss is 12.75, training accuracy is 0.031250, time 219.171 samples/sec
+epoch 1, total_step 21740, total loss is 31.10 , inference loss is 18.35, weight deacy loss is 12.75, training accuracy is 0.125000, time 213.514 samples/sec
+epoch 1, total_step 21760, total loss is 35.00 , inference loss is 22.26, weight deacy loss is 12.74, training accuracy is 0.000000, time 214.066 samples/sec
+epoch 1, total_step 21780, total loss is 32.46 , inference loss is 19.72, weight deacy loss is 12.74, training accuracy is 0.093750, time 237.591 samples/sec
+epoch 1, total_step 21800, total loss is 30.45 , inference loss is 17.71, weight deacy loss is 12.74, training accuracy is 0.062500, time 208.450 samples/sec
+epoch 1, total_step 21820, total loss is 32.40 , inference loss is 19.66, weight deacy loss is 12.74, training accuracy is 0.000000, time 213.723 samples/sec
+epoch 1, total_step 21840, total loss is 28.02 , inference loss is 15.29, weight deacy loss is 12.74, training accuracy is 0.156250, time 226.250 samples/sec
+epoch 1, total_step 21860, total loss is 31.95 , inference loss is 19.21, weight deacy loss is 12.74, training accuracy is 0.125000, time 216.433 samples/sec
+epoch 1, total_step 21880, total loss is 33.63 , inference loss is 20.90, weight deacy loss is 12.73, training accuracy is 0.062500, time 199.916 samples/sec
+epoch 1, total_step 21900, total loss is 31.72 , inference loss is 18.99, weight deacy loss is 12.73, training accuracy is 0.093750, time 13.368 samples/sec
+epoch 1, total_step 21920, total loss is 33.69 , inference loss is 20.96, weight deacy loss is 12.73, training accuracy is 0.062500, time 226.394 samples/sec
+epoch 1, total_step 21940, total loss is 35.04 , inference loss is 22.31, weight deacy loss is 12.73, training accuracy is 0.093750, time 210.859 samples/sec
+epoch 1, total_step 21960, total loss is 33.63 , inference loss is 20.90, weight deacy loss is 12.73, training accuracy is 0.062500, time 213.739 samples/sec
+epoch 1, total_step 21980, total loss is 33.76 , inference loss is 21.03, weight deacy loss is 12.72, training accuracy is 0.031250, time 210.071 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.185083000000006
+best_threshold_index 105 0.9612962962962963
+best_threshold_index 105 0.962037037037037
+best_threshold_index 104 0.9631481481481482
+best_threshold_index 105 0.9638888888888889
+best_threshold_index 105 0.9624074074074074
+best_threshold_index 105 0.9601851851851851
+best_threshold_index 105 0.9616666666666667
+best_threshold_index 105 0.9614814814814815
+best_threshold_index 105 0.9596296296296296
+best_threshold_index 105 0.960925925925926
+[lfw][22000]XNorm: 24.974489
+[lfw][22000]Accuracy-Flip: 0.96117+-0.01167
+test accuracy is: 0.9611666666666668
+epoch 1, total_step 22000, total loss is 29.99 , inference loss is 17.27, weight deacy loss is 12.72, training accuracy is 0.062500, time 18.754 samples/sec
+epoch 1, total_step 22020, total loss is 28.00 , inference loss is 15.28, weight deacy loss is 12.72, training accuracy is 0.062500, time 296.069 samples/sec
+epoch 1, total_step 22040, total loss is 28.80 , inference loss is 16.08, weight deacy loss is 12.72, training accuracy is 0.125000, time 195.153 samples/sec
+epoch 1, total_step 22060, total loss is 31.05 , inference loss is 18.33, weight deacy loss is 12.72, training accuracy is 0.125000, time 214.061 samples/sec
+epoch 1, total_step 22080, total loss is 25.12 , inference loss is 12.41, weight deacy loss is 12.72, training accuracy is 0.125000, time 222.472 samples/sec
+epoch 1, total_step 22100, total loss is 32.12 , inference loss is 19.40, weight deacy loss is 12.71, training accuracy is 0.062500, time 218.466 samples/sec
+epoch 1, total_step 22120, total loss is 27.73 , inference loss is 15.02, weight deacy loss is 12.71, training accuracy is 0.156250, time 207.987 samples/sec
+epoch 1, total_step 22140, total loss is 30.52 , inference loss is 17.81, weight deacy loss is 12.71, training accuracy is 0.125000, time 227.219 samples/sec
+epoch 1, total_step 22160, total loss is 30.00 , inference loss is 17.29, weight deacy loss is 12.71, training accuracy is 0.125000, time 208.489 samples/sec
+epoch 1, total_step 22180, total loss is 32.30 , inference loss is 19.59, weight deacy loss is 12.71, training accuracy is 0.062500, time 232.394 samples/sec
+epoch 1, total_step 22200, total loss is 30.57 , inference loss is 17.86, weight deacy loss is 12.71, training accuracy is 0.093750, time 13.621 samples/sec
+epoch 1, total_step 22220, total loss is 31.95 , inference loss is 19.25, weight deacy loss is 12.70, training accuracy is 0.062500, time 230.863 samples/sec
+epoch 1, total_step 22240, total loss is 29.54 , inference loss is 16.84, weight deacy loss is 12.70, training accuracy is 0.062500, time 224.003 samples/sec
+epoch 1, total_step 22260, total loss is 29.91 , inference loss is 17.21, weight deacy loss is 12.70, training accuracy is 0.093750, time 214.713 samples/sec
+epoch 1, total_step 22280, total loss is 28.96 , inference loss is 16.26, weight deacy loss is 12.70, training accuracy is 0.125000, time 196.522 samples/sec
+epoch 1, total_step 22300, total loss is 33.62 , inference loss is 20.92, weight deacy loss is 12.70, training accuracy is 0.125000, time 208.971 samples/sec
+epoch 1, total_step 22320, total loss is 28.30 , inference loss is 15.60, weight deacy loss is 12.69, training accuracy is 0.156250, time 224.502 samples/sec
+epoch 1, total_step 22340, total loss is 29.87 , inference loss is 17.18, weight deacy loss is 12.69, training accuracy is 0.062500, time 239.715 samples/sec
+epoch 1, total_step 22360, total loss is 31.18 , inference loss is 18.49, weight deacy loss is 12.69, training accuracy is 0.062500, time 211.918 samples/sec
+epoch 1, total_step 22380, total loss is 34.76 , inference loss is 22.07, weight deacy loss is 12.69, training accuracy is 0.093750, time 225.373 samples/sec
+epoch 1, total_step 22400, total loss is 30.77 , inference loss is 18.08, weight deacy loss is 12.69, training accuracy is 0.125000, time 234.989 samples/sec
+epoch 1, total_step 22420, total loss is 31.16 , inference loss is 18.48, weight deacy loss is 12.69, training accuracy is 0.125000, time 221.738 samples/sec
+epoch 1, total_step 22440, total loss is 38.54 , inference loss is 25.86, weight deacy loss is 12.68, training accuracy is 0.000000, time 225.479 samples/sec
+epoch 1, total_step 22460, total loss is 26.05 , inference loss is 13.37, weight deacy loss is 12.68, training accuracy is 0.093750, time 222.726 samples/sec
+epoch 1, total_step 22480, total loss is 32.95 , inference loss is 20.27, weight deacy loss is 12.68, training accuracy is 0.062500, time 207.645 samples/sec
+epoch 1, total_step 22500, total loss is 29.95 , inference loss is 17.27, weight deacy loss is 12.68, training accuracy is 0.000000, time 11.844 samples/sec
+epoch 1, total_step 22520, total loss is 36.81 , inference loss is 24.14, weight deacy loss is 12.68, training accuracy is 0.031250, time 207.611 samples/sec
+epoch 1, total_step 22540, total loss is 27.44 , inference loss is 14.76, weight deacy loss is 12.68, training accuracy is 0.218750, time 296.163 samples/sec
+epoch 1, total_step 22560, total loss is 30.91 , inference loss is 18.23, weight deacy loss is 12.67, training accuracy is 0.187500, time 218.583 samples/sec
+epoch 1, total_step 22580, total loss is 29.47 , inference loss is 16.80, weight deacy loss is 12.67, training accuracy is 0.187500, time 220.273 samples/sec
+epoch 1, total_step 22600, total loss is 30.22 , inference loss is 17.55, weight deacy loss is 12.67, training accuracy is 0.125000, time 216.539 samples/sec
+epoch 1, total_step 22620, total loss is 26.62 , inference loss is 13.95, weight deacy loss is 12.67, training accuracy is 0.125000, time 240.341 samples/sec
+epoch 1, total_step 22640, total loss is 31.40 , inference loss is 18.73, weight deacy loss is 12.67, training accuracy is 0.093750, time 208.453 samples/sec
+epoch 1, total_step 22660, total loss is 30.54 , inference loss is 17.88, weight deacy loss is 12.67, training accuracy is 0.062500, time 240.427 samples/sec
+epoch 1, total_step 22680, total loss is 35.64 , inference loss is 22.98, weight deacy loss is 12.66, training accuracy is 0.000000, time 222.015 samples/sec
+epoch 1, total_step 22700, total loss is 29.48 , inference loss is 16.81, weight deacy loss is 12.66, training accuracy is 0.062500, time 204.756 samples/sec
+epoch 1, total_step 22720, total loss is 27.72 , inference loss is 15.06, weight deacy loss is 12.66, training accuracy is 0.093750, time 207.144 samples/sec
+epoch 1, total_step 22740, total loss is 30.31 , inference loss is 17.65, weight deacy loss is 12.66, training accuracy is 0.156250, time 228.411 samples/sec
+epoch 1, total_step 22760, total loss is 29.20 , inference loss is 16.55, weight deacy loss is 12.66, training accuracy is 0.093750, time 216.109 samples/sec
+epoch 1, total_step 22780, total loss is 28.46 , inference loss is 15.81, weight deacy loss is 12.65, training accuracy is 0.093750, time 215.066 samples/sec
+epoch 1, total_step 22800, total loss is 27.91 , inference loss is 15.26, weight deacy loss is 12.65, training accuracy is 0.187500, time 13.224 samples/sec
+epoch 1, total_step 22820, total loss is 27.36 , inference loss is 14.70, weight deacy loss is 12.65, training accuracy is 0.125000, time 222.511 samples/sec
+epoch 1, total_step 22840, total loss is 31.67 , inference loss is 19.02, weight deacy loss is 12.65, training accuracy is 0.093750, time 225.919 samples/sec
+epoch 1, total_step 22860, total loss is 33.42 , inference loss is 20.77, weight deacy loss is 12.65, training accuracy is 0.093750, time 213.002 samples/sec
+epoch 1, total_step 22880, total loss is 29.49 , inference loss is 16.84, weight deacy loss is 12.65, training accuracy is 0.062500, time 223.929 samples/sec
+epoch 1, total_step 22900, total loss is 29.52 , inference loss is 16.88, weight deacy loss is 12.64, training accuracy is 0.156250, time 221.780 samples/sec
+epoch 1, total_step 22920, total loss is 26.13 , inference loss is 13.49, weight deacy loss is 12.64, training accuracy is 0.187500, time 201.228 samples/sec
+epoch 1, total_step 22940, total loss is 30.05 , inference loss is 17.41, weight deacy loss is 12.64, training accuracy is 0.187500, time 238.549 samples/sec
+epoch 1, total_step 22960, total loss is 33.33 , inference loss is 20.69, weight deacy loss is 12.64, training accuracy is 0.156250, time 211.607 samples/sec
+epoch 1, total_step 22980, total loss is 27.24 , inference loss is 14.60, weight deacy loss is 12.64, training accuracy is 0.093750, time 212.755 samples/sec
+epoch 1, total_step 23000, total loss is 25.95 , inference loss is 13.31, weight deacy loss is 12.64, training accuracy is 0.125000, time 221.698 samples/sec
+epoch 1, total_step 23020, total loss is 29.73 , inference loss is 17.10, weight deacy loss is 12.63, training accuracy is 0.093750, time 228.021 samples/sec
+epoch 1, total_step 23040, total loss is 29.44 , inference loss is 16.81, weight deacy loss is 12.63, training accuracy is 0.125000, time 208.933 samples/sec
+epoch 1, total_step 23060, total loss is 26.82 , inference loss is 14.19, weight deacy loss is 12.63, training accuracy is 0.218750, time 217.753 samples/sec
+epoch 1, total_step 23080, total loss is 32.19 , inference loss is 19.57, weight deacy loss is 12.63, training accuracy is 0.093750, time 218.503 samples/sec
+epoch 1, total_step 23100, total loss is 27.02 , inference loss is 14.39, weight deacy loss is 12.63, training accuracy is 0.187500, time 14.331 samples/sec
+epoch 1, total_step 23120, total loss is 29.78 , inference loss is 17.16, weight deacy loss is 12.62, training accuracy is 0.156250, time 299.211 samples/sec
+epoch 1, total_step 23140, total loss is 29.56 , inference loss is 16.94, weight deacy loss is 12.62, training accuracy is 0.187500, time 296.122 samples/sec
+epoch 1, total_step 23160, total loss is 32.68 , inference loss is 20.06, weight deacy loss is 12.62, training accuracy is 0.187500, time 207.778 samples/sec
+epoch 1, total_step 23180, total loss is 30.36 , inference loss is 17.74, weight deacy loss is 12.62, training accuracy is 0.156250, time 218.275 samples/sec
+epoch 1, total_step 23200, total loss is 34.79 , inference loss is 22.17, weight deacy loss is 12.62, training accuracy is 0.031250, time 210.483 samples/sec
+epoch 1, total_step 23220, total loss is 29.40 , inference loss is 16.78, weight deacy loss is 12.62, training accuracy is 0.187500, time 235.802 samples/sec
+epoch 1, total_step 23240, total loss is 31.51 , inference loss is 18.90, weight deacy loss is 12.61, training accuracy is 0.062500, time 206.154 samples/sec
+epoch 1, total_step 23260, total loss is 31.76 , inference loss is 19.15, weight deacy loss is 12.61, training accuracy is 0.093750, time 297.434 samples/sec
+epoch 1, total_step 23280, total loss is 27.82 , inference loss is 15.21, weight deacy loss is 12.61, training accuracy is 0.218750, time 206.028 samples/sec
+epoch 1, total_step 23300, total loss is 27.39 , inference loss is 14.79, weight deacy loss is 12.61, training accuracy is 0.156250, time 202.531 samples/sec
+epoch 1, total_step 23320, total loss is 31.77 , inference loss is 19.16, weight deacy loss is 12.61, training accuracy is 0.093750, time 230.253 samples/sec
+epoch 1, total_step 23340, total loss is 33.41 , inference loss is 20.80, weight deacy loss is 12.61, training accuracy is 0.062500, time 221.920 samples/sec
+epoch 1, total_step 23360, total loss is 25.68 , inference loss is 13.08, weight deacy loss is 12.60, training accuracy is 0.093750, time 219.069 samples/sec
+epoch 1, total_step 23380, total loss is 30.68 , inference loss is 18.08, weight deacy loss is 12.60, training accuracy is 0.187500, time 219.731 samples/sec
+epoch 1, total_step 23400, total loss is 30.76 , inference loss is 18.16, weight deacy loss is 12.60, training accuracy is 0.156250, time 13.425 samples/sec
+epoch 1, total_step 23420, total loss is 33.11 , inference loss is 20.51, weight deacy loss is 12.60, training accuracy is 0.062500, time 221.027 samples/sec
+epoch 1, total_step 23440, total loss is 29.04 , inference loss is 16.44, weight deacy loss is 12.60, training accuracy is 0.125000, time 193.328 samples/sec
+epoch 1, total_step 23460, total loss is 28.85 , inference loss is 16.25, weight deacy loss is 12.59, training accuracy is 0.125000, time 212.003 samples/sec
+epoch 1, total_step 23480, total loss is 32.74 , inference loss is 20.14, weight deacy loss is 12.59, training accuracy is 0.125000, time 222.713 samples/sec
+epoch 1, total_step 23500, total loss is 28.54 , inference loss is 15.95, weight deacy loss is 12.59, training accuracy is 0.156250, time 221.890 samples/sec
+epoch 1, total_step 23520, total loss is 31.45 , inference loss is 18.86, weight deacy loss is 12.59, training accuracy is 0.093750, time 294.763 samples/sec
+epoch 1, total_step 23540, total loss is 33.98 , inference loss is 21.39, weight deacy loss is 12.59, training accuracy is 0.062500, time 212.063 samples/sec
+epoch 1, total_step 23560, total loss is 33.93 , inference loss is 21.34, weight deacy loss is 12.59, training accuracy is 0.062500, time 217.808 samples/sec
+epoch 1, total_step 23580, total loss is 29.96 , inference loss is 17.37, weight deacy loss is 12.58, training accuracy is 0.156250, time 213.071 samples/sec
+epoch 1, total_step 23600, total loss is 31.83 , inference loss is 19.25, weight deacy loss is 12.58, training accuracy is 0.062500, time 214.320 samples/sec
+epoch 1, total_step 23620, total loss is 31.56 , inference loss is 18.98, weight deacy loss is 12.58, training accuracy is 0.093750, time 215.816 samples/sec
+epoch 1, total_step 23640, total loss is 28.35 , inference loss is 15.77, weight deacy loss is 12.58, training accuracy is 0.218750, time 217.760 samples/sec
+epoch 1, total_step 23660, total loss is 31.24 , inference loss is 18.66, weight deacy loss is 12.58, training accuracy is 0.093750, time 208.904 samples/sec
+epoch 1, total_step 23680, total loss is 31.58 , inference loss is 19.00, weight deacy loss is 12.58, training accuracy is 0.031250, time 215.575 samples/sec
+epoch 1, total_step 23700, total loss is 28.26 , inference loss is 15.69, weight deacy loss is 12.57, training accuracy is 0.156250, time 13.431 samples/sec
+epoch 1, total_step 23720, total loss is 36.90 , inference loss is 24.33, weight deacy loss is 12.57, training accuracy is 0.093750, time 205.549 samples/sec
+epoch 1, total_step 23740, total loss is 32.09 , inference loss is 19.52, weight deacy loss is 12.57, training accuracy is 0.156250, time 220.374 samples/sec
+epoch 1, total_step 23760, total loss is 30.59 , inference loss is 18.03, weight deacy loss is 12.57, training accuracy is 0.125000, time 227.424 samples/sec
+epoch 1, total_step 23780, total loss is 35.93 , inference loss is 23.37, weight deacy loss is 12.57, training accuracy is 0.031250, time 299.464 samples/sec
+epoch 1, total_step 23800, total loss is 28.32 , inference loss is 15.75, weight deacy loss is 12.57, training accuracy is 0.093750, time 212.224 samples/sec
+epoch 1, total_step 23820, total loss is 33.11 , inference loss is 20.54, weight deacy loss is 12.56, training accuracy is 0.125000, time 226.031 samples/sec
+epoch 1, total_step 23840, total loss is 33.89 , inference loss is 21.33, weight deacy loss is 12.56, training accuracy is 0.062500, time 238.076 samples/sec
+epoch 1, total_step 23860, total loss is 31.33 , inference loss is 18.77, weight deacy loss is 12.56, training accuracy is 0.125000, time 216.058 samples/sec
+epoch 1, total_step 23880, total loss is 30.39 , inference loss is 17.83, weight deacy loss is 12.56, training accuracy is 0.093750, time 217.098 samples/sec
+epoch 1, total_step 23900, total loss is 34.13 , inference loss is 21.58, weight deacy loss is 12.56, training accuracy is 0.031250, time 216.325 samples/sec
+epoch 1, total_step 23920, total loss is 31.63 , inference loss is 19.08, weight deacy loss is 12.55, training accuracy is 0.062500, time 204.985 samples/sec
+epoch 1, total_step 23940, total loss is 25.92 , inference loss is 13.37, weight deacy loss is 12.55, training accuracy is 0.156250, time 211.961 samples/sec
+epoch 1, total_step 23960, total loss is 29.69 , inference loss is 17.14, weight deacy loss is 12.55, training accuracy is 0.062500, time 235.051 samples/sec
+epoch 1, total_step 23980, total loss is 32.37 , inference loss is 19.82, weight deacy loss is 12.55, training accuracy is 0.187500, time 206.935 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.48513200000002
+best_threshold_index 110 0.9711111111111111
+best_threshold_index 110 0.9718518518518519
+best_threshold_index 110 0.9716666666666667
+best_threshold_index 108 0.9714814814814815
+best_threshold_index 110 0.9718518518518519
+best_threshold_index 110 0.9698148148148148
+best_threshold_index 110 0.9705555555555555
+best_threshold_index 107 0.9705555555555555
+best_threshold_index 110 0.9688888888888889
+best_threshold_index 108 0.9694444444444444
+[lfw][24000]XNorm: 24.657568
+[lfw][24000]Accuracy-Flip: 0.96933+-0.00898
+test accuracy is: 0.9693333333333334
+epoch 1, total_step 24000, total loss is 31.20 , inference loss is 18.65, weight deacy loss is 12.55, training accuracy is 0.125000, time 17.094 samples/sec
+epoch 1, total_step 24020, total loss is 34.47 , inference loss is 21.92, weight deacy loss is 12.55, training accuracy is 0.031250, time 247.498 samples/sec
+epoch 1, total_step 24040, total loss is 32.44 , inference loss is 19.90, weight deacy loss is 12.54, training accuracy is 0.093750, time 234.916 samples/sec
+epoch 1, total_step 24060, total loss is 32.15 , inference loss is 19.60, weight deacy loss is 12.54, training accuracy is 0.031250, time 211.436 samples/sec
+epoch 1, total_step 24080, total loss is 30.73 , inference loss is 18.19, weight deacy loss is 12.54, training accuracy is 0.031250, time 207.138 samples/sec
+epoch 1, total_step 24100, total loss is 30.52 , inference loss is 17.98, weight deacy loss is 12.54, training accuracy is 0.093750, time 219.368 samples/sec
+epoch 1, total_step 24120, total loss is 27.73 , inference loss is 15.20, weight deacy loss is 12.54, training accuracy is 0.062500, time 221.377 samples/sec
+epoch 1, total_step 24140, total loss is 31.12 , inference loss is 18.59, weight deacy loss is 12.54, training accuracy is 0.125000, time 209.273 samples/sec
+epoch 1, total_step 24160, total loss is 30.82 , inference loss is 18.28, weight deacy loss is 12.53, training accuracy is 0.062500, time 298.947 samples/sec
+epoch 1, total_step 24180, total loss is 31.73 , inference loss is 19.19, weight deacy loss is 12.53, training accuracy is 0.062500, time 225.784 samples/sec
+epoch 1, total_step 24200, total loss is 33.53 , inference loss is 21.00, weight deacy loss is 12.53, training accuracy is 0.062500, time 206.721 samples/sec
+epoch 1, total_step 24220, total loss is 31.15 , inference loss is 18.62, weight deacy loss is 12.53, training accuracy is 0.031250, time 212.184 samples/sec
+epoch 1, total_step 24240, total loss is 34.52 , inference loss is 22.00, weight deacy loss is 12.53, training accuracy is 0.000000, time 218.944 samples/sec
+epoch 1, total_step 24260, total loss is 28.86 , inference loss is 16.34, weight deacy loss is 12.53, training accuracy is 0.125000, time 209.932 samples/sec
+epoch 1, total_step 24280, total loss is 30.40 , inference loss is 17.88, weight deacy loss is 12.52, training accuracy is 0.062500, time 219.687 samples/sec
+epoch 1, total_step 24300, total loss is 32.20 , inference loss is 19.68, weight deacy loss is 12.52, training accuracy is 0.000000, time 14.234 samples/sec
+epoch 1, total_step 24320, total loss is 28.52 , inference loss is 16.00, weight deacy loss is 12.52, training accuracy is 0.187500, time 225.836 samples/sec
+epoch 1, total_step 24340, total loss is 30.18 , inference loss is 17.66, weight deacy loss is 12.52, training accuracy is 0.187500, time 213.204 samples/sec
+epoch 1, total_step 24360, total loss is 30.04 , inference loss is 17.53, weight deacy loss is 12.52, training accuracy is 0.125000, time 196.317 samples/sec
+epoch 1, total_step 24380, total loss is 30.59 , inference loss is 18.07, weight deacy loss is 12.52, training accuracy is 0.125000, time 225.382 samples/sec
+epoch 1, total_step 24400, total loss is 28.43 , inference loss is 15.92, weight deacy loss is 12.51, training accuracy is 0.156250, time 225.355 samples/sec
+epoch 1, total_step 24420, total loss is 28.74 , inference loss is 16.23, weight deacy loss is 12.51, training accuracy is 0.031250, time 298.954 samples/sec
+epoch 1, total_step 24440, total loss is 33.55 , inference loss is 21.04, weight deacy loss is 12.51, training accuracy is 0.062500, time 208.531 samples/sec
+epoch 1, total_step 24460, total loss is 29.57 , inference loss is 17.06, weight deacy loss is 12.51, training accuracy is 0.062500, time 223.753 samples/sec
+epoch 1, total_step 24480, total loss is 27.07 , inference loss is 14.57, weight deacy loss is 12.51, training accuracy is 0.125000, time 235.223 samples/sec
+epoch 1, total_step 24500, total loss is 33.23 , inference loss is 20.72, weight deacy loss is 12.51, training accuracy is 0.062500, time 216.106 samples/sec
+epoch 1, total_step 24520, total loss is 30.99 , inference loss is 18.49, weight deacy loss is 12.50, training accuracy is 0.187500, time 223.697 samples/sec
+epoch 1, total_step 24540, total loss is 29.44 , inference loss is 16.94, weight deacy loss is 12.50, training accuracy is 0.093750, time 204.414 samples/sec
+epoch 1, total_step 24560, total loss is 30.07 , inference loss is 17.57, weight deacy loss is 12.50, training accuracy is 0.031250, time 220.545 samples/sec
+epoch 1, total_step 24580, total loss is 30.69 , inference loss is 18.19, weight deacy loss is 12.50, training accuracy is 0.093750, time 199.797 samples/sec
+epoch 1, total_step 24600, total loss is 29.69 , inference loss is 17.19, weight deacy loss is 12.50, training accuracy is 0.156250, time 12.036 samples/sec
+epoch 1, total_step 24620, total loss is 34.44 , inference loss is 21.95, weight deacy loss is 12.50, training accuracy is 0.093750, time 231.500 samples/sec
+epoch 1, total_step 24640, total loss is 30.72 , inference loss is 18.23, weight deacy loss is 12.49, training accuracy is 0.062500, time 224.380 samples/sec
+epoch 1, total_step 24660, total loss is 33.08 , inference loss is 20.59, weight deacy loss is 12.49, training accuracy is 0.093750, time 225.870 samples/sec
+epoch 1, total_step 24680, total loss is 30.06 , inference loss is 17.57, weight deacy loss is 12.49, training accuracy is 0.187500, time 296.934 samples/sec
+epoch 1, total_step 24700, total loss is 28.68 , inference loss is 16.19, weight deacy loss is 12.49, training accuracy is 0.218750, time 200.032 samples/sec
+epoch 1, total_step 24720, total loss is 29.67 , inference loss is 17.19, weight deacy loss is 12.49, training accuracy is 0.031250, time 210.909 samples/sec
+epoch 1, total_step 24740, total loss is 28.96 , inference loss is 16.47, weight deacy loss is 12.49, training accuracy is 0.187500, time 209.103 samples/sec
+epoch 1, total_step 24760, total loss is 34.67 , inference loss is 22.19, weight deacy loss is 12.48, training accuracy is 0.000000, time 213.956 samples/sec
+epoch 1, total_step 24780, total loss is 29.20 , inference loss is 16.72, weight deacy loss is 12.48, training accuracy is 0.125000, time 218.405 samples/sec
+epoch 1, total_step 24800, total loss is 27.12 , inference loss is 14.64, weight deacy loss is 12.48, training accuracy is 0.156250, time 215.061 samples/sec
+epoch 1, total_step 24820, total loss is 29.30 , inference loss is 16.83, weight deacy loss is 12.48, training accuracy is 0.125000, time 212.095 samples/sec
+epoch 1, total_step 24840, total loss is 28.50 , inference loss is 16.03, weight deacy loss is 12.48, training accuracy is 0.156250, time 218.509 samples/sec
+epoch 1, total_step 24860, total loss is 31.66 , inference loss is 19.18, weight deacy loss is 12.47, training accuracy is 0.125000, time 216.842 samples/sec
+epoch 1, total_step 24880, total loss is 30.23 , inference loss is 17.75, weight deacy loss is 12.47, training accuracy is 0.000000, time 215.028 samples/sec
+epoch 1, total_step 24900, total loss is 34.74 , inference loss is 22.27, weight deacy loss is 12.47, training accuracy is 0.125000, time 14.397 samples/sec
+epoch 1, total_step 24920, total loss is 35.94 , inference loss is 23.47, weight deacy loss is 12.47, training accuracy is 0.093750, time 226.119 samples/sec
+epoch 1, total_step 24940, total loss is 31.35 , inference loss is 18.88, weight deacy loss is 12.47, training accuracy is 0.062500, time 207.616 samples/sec
+epoch 1, total_step 24960, total loss is 28.73 , inference loss is 16.26, weight deacy loss is 12.47, training accuracy is 0.156250, time 225.095 samples/sec
+epoch 1, total_step 24980, total loss is 31.46 , inference loss is 19.00, weight deacy loss is 12.46, training accuracy is 0.125000, time 215.060 samples/sec
+epoch 1, total_step 25000, total loss is 33.02 , inference loss is 20.55, weight deacy loss is 12.46, training accuracy is 0.156250, time 207.731 samples/sec
+epoch 1, total_step 25020, total loss is 30.26 , inference loss is 17.79, weight deacy loss is 12.46, training accuracy is 0.093750, time 226.699 samples/sec
+epoch 1, total_step 25040, total loss is 31.63 , inference loss is 19.17, weight deacy loss is 12.46, training accuracy is 0.093750, time 226.105 samples/sec
+epoch 1, total_step 25060, total loss is 31.98 , inference loss is 19.52, weight deacy loss is 12.46, training accuracy is 0.000000, time 232.924 samples/sec
+epoch 1, total_step 25080, total loss is 28.99 , inference loss is 16.53, weight deacy loss is 12.46, training accuracy is 0.093750, time 223.507 samples/sec
+epoch 1, total_step 25100, total loss is 31.72 , inference loss is 19.27, weight deacy loss is 12.45, training accuracy is 0.125000, time 222.211 samples/sec
+epoch 1, total_step 25120, total loss is 30.38 , inference loss is 17.93, weight deacy loss is 12.45, training accuracy is 0.250000, time 297.728 samples/sec
+epoch 1, total_step 25140, total loss is 31.66 , inference loss is 19.21, weight deacy loss is 12.45, training accuracy is 0.125000, time 213.817 samples/sec
+epoch 1, total_step 25160, total loss is 32.31 , inference loss is 19.86, weight deacy loss is 12.45, training accuracy is 0.062500, time 227.366 samples/sec
+epoch 1, total_step 25180, total loss is 28.56 , inference loss is 16.11, weight deacy loss is 12.45, training accuracy is 0.093750, time 216.374 samples/sec
+epoch 1, total_step 25200, total loss is 30.65 , inference loss is 18.20, weight deacy loss is 12.45, training accuracy is 0.031250, time 11.092 samples/sec
+epoch 1, total_step 25220, total loss is 32.67 , inference loss is 20.23, weight deacy loss is 12.44, training accuracy is 0.000000, time 210.335 samples/sec
+epoch 1, total_step 25240, total loss is 26.00 , inference loss is 13.56, weight deacy loss is 12.44, training accuracy is 0.125000, time 231.242 samples/sec
+epoch 1, total_step 25260, total loss is 31.58 , inference loss is 19.14, weight deacy loss is 12.44, training accuracy is 0.062500, time 211.147 samples/sec
+epoch 1, total_step 25280, total loss is 28.26 , inference loss is 15.82, weight deacy loss is 12.44, training accuracy is 0.156250, time 219.946 samples/sec
+epoch 1, total_step 25300, total loss is 29.22 , inference loss is 16.79, weight deacy loss is 12.44, training accuracy is 0.093750, time 200.013 samples/sec
+epoch 1, total_step 25320, total loss is 26.91 , inference loss is 14.48, weight deacy loss is 12.44, training accuracy is 0.093750, time 212.641 samples/sec
+epoch 1, total_step 25340, total loss is 32.79 , inference loss is 20.36, weight deacy loss is 12.43, training accuracy is 0.031250, time 213.859 samples/sec
+epoch 1, total_step 25360, total loss is 28.19 , inference loss is 15.75, weight deacy loss is 12.43, training accuracy is 0.125000, time 206.756 samples/sec
+epoch 1, total_step 25380, total loss is 30.35 , inference loss is 17.92, weight deacy loss is 12.43, training accuracy is 0.031250, time 298.329 samples/sec
+epoch 1, total_step 25400, total loss is 26.74 , inference loss is 14.31, weight deacy loss is 12.43, training accuracy is 0.187500, time 207.515 samples/sec
+epoch 1, total_step 25420, total loss is 25.07 , inference loss is 12.64, weight deacy loss is 12.43, training accuracy is 0.156250, time 224.654 samples/sec
+epoch 1, total_step 25440, total loss is 29.08 , inference loss is 16.65, weight deacy loss is 12.43, training accuracy is 0.156250, time 214.897 samples/sec
+epoch 1, total_step 25460, total loss is 30.37 , inference loss is 17.95, weight deacy loss is 12.42, training accuracy is 0.125000, time 201.235 samples/sec
+epoch 1, total_step 25480, total loss is 27.74 , inference loss is 15.31, weight deacy loss is 12.42, training accuracy is 0.062500, time 229.249 samples/sec
+epoch 1, total_step 25500, total loss is 30.62 , inference loss is 18.20, weight deacy loss is 12.42, training accuracy is 0.031250, time 14.135 samples/sec
+epoch 1, total_step 25520, total loss is 30.32 , inference loss is 17.90, weight deacy loss is 12.42, training accuracy is 0.125000, time 195.882 samples/sec
+epoch 1, total_step 25540, total loss is 29.37 , inference loss is 16.95, weight deacy loss is 12.42, training accuracy is 0.000000, time 210.963 samples/sec
+epoch 1, total_step 25560, total loss is 27.25 , inference loss is 14.83, weight deacy loss is 12.42, training accuracy is 0.156250, time 211.839 samples/sec
+epoch 1, total_step 25580, total loss is 32.64 , inference loss is 20.23, weight deacy loss is 12.41, training accuracy is 0.156250, time 236.988 samples/sec
+epoch 1, total_step 25600, total loss is 30.79 , inference loss is 18.38, weight deacy loss is 12.41, training accuracy is 0.062500, time 218.533 samples/sec
+epoch 1, total_step 25620, total loss is 28.38 , inference loss is 15.97, weight deacy loss is 12.41, training accuracy is 0.187500, time 217.833 samples/sec
+epoch 1, total_step 25640, total loss is 30.46 , inference loss is 18.06, weight deacy loss is 12.41, training accuracy is 0.093750, time 184.454 samples/sec
+epoch 1, total_step 25660, total loss is 28.51 , inference loss is 16.10, weight deacy loss is 12.41, training accuracy is 0.093750, time 189.019 samples/sec
+epoch 1, total_step 25680, total loss is 30.89 , inference loss is 18.49, weight deacy loss is 12.41, training accuracy is 0.125000, time 211.403 samples/sec
+epoch 1, total_step 25700, total loss is 30.39 , inference loss is 17.99, weight deacy loss is 12.40, training accuracy is 0.156250, time 219.824 samples/sec
+epoch 1, total_step 25720, total loss is 32.79 , inference loss is 20.39, weight deacy loss is 12.40, training accuracy is 0.125000, time 216.686 samples/sec
+epoch 1, total_step 25740, total loss is 30.92 , inference loss is 18.52, weight deacy loss is 12.40, training accuracy is 0.000000, time 231.571 samples/sec
+epoch 1, total_step 25760, total loss is 28.41 , inference loss is 16.01, weight deacy loss is 12.40, training accuracy is 0.093750, time 221.440 samples/sec
+epoch 1, total_step 25780, total loss is 28.74 , inference loss is 16.34, weight deacy loss is 12.40, training accuracy is 0.156250, time 196.371 samples/sec
+epoch 1, total_step 25800, total loss is 28.15 , inference loss is 15.75, weight deacy loss is 12.40, training accuracy is 0.218750, time 14.214 samples/sec
+epoch 1, total_step 25820, total loss is 27.19 , inference loss is 14.79, weight deacy loss is 12.39, training accuracy is 0.093750, time 225.421 samples/sec
+epoch 1, total_step 25840, total loss is 34.20 , inference loss is 21.81, weight deacy loss is 12.39, training accuracy is 0.093750, time 225.196 samples/sec
+epoch 1, total_step 25860, total loss is 30.66 , inference loss is 18.27, weight deacy loss is 12.39, training accuracy is 0.156250, time 215.422 samples/sec
+epoch 1, total_step 25880, total loss is 27.59 , inference loss is 15.20, weight deacy loss is 12.39, training accuracy is 0.062500, time 219.026 samples/sec
+epoch 1, total_step 25900, total loss is 28.02 , inference loss is 15.63, weight deacy loss is 12.39, training accuracy is 0.093750, time 222.216 samples/sec
+epoch 1, total_step 25920, total loss is 34.13 , inference loss is 21.74, weight deacy loss is 12.39, training accuracy is 0.093750, time 298.301 samples/sec
+epoch 1, total_step 25940, total loss is 29.56 , inference loss is 17.18, weight deacy loss is 12.38, training accuracy is 0.062500, time 216.512 samples/sec
+epoch 1, total_step 25960, total loss is 33.59 , inference loss is 21.21, weight deacy loss is 12.38, training accuracy is 0.031250, time 213.439 samples/sec
+epoch 1, total_step 25980, total loss is 27.47 , inference loss is 15.09, weight deacy loss is 12.38, training accuracy is 0.125000, time 221.700 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.37247
+best_threshold_index 109 0.9681481481481482
+best_threshold_index 109 0.9674074074074074
+best_threshold_index 103 0.9692592592592593
+best_threshold_index 104 0.9681481481481482
+best_threshold_index 109 0.9679629629629629
+best_threshold_index 103 0.9661111111111111
+best_threshold_index 109 0.9685185185185186
+best_threshold_index 103 0.9672222222222222
+best_threshold_index 103 0.9661111111111111
+best_threshold_index 101 0.9672222222222222
+[lfw][26000]XNorm: 23.189070
+[lfw][26000]Accuracy-Flip: 0.96483+-0.00835
+test accuracy is: 0.9648333333333333
+epoch 1, total_step 26000, total loss is 28.70 , inference loss is 16.32, weight deacy loss is 12.38, training accuracy is 0.125000, time 17.263 samples/sec
+epoch 1, total_step 26020, total loss is 35.73 , inference loss is 23.36, weight deacy loss is 12.38, training accuracy is 0.000000, time 212.148 samples/sec
+epoch 1, total_step 26040, total loss is 28.18 , inference loss is 15.80, weight deacy loss is 12.38, training accuracy is 0.093750, time 236.096 samples/sec
+epoch 1, total_step 26060, total loss is 26.40 , inference loss is 14.03, weight deacy loss is 12.37, training accuracy is 0.156250, time 209.151 samples/sec
+epoch 1, total_step 26080, total loss is 28.50 , inference loss is 16.13, weight deacy loss is 12.37, training accuracy is 0.187500, time 208.292 samples/sec
+epoch 1, total_step 26100, total loss is 30.39 , inference loss is 18.02, weight deacy loss is 12.37, training accuracy is 0.062500, time 14.383 samples/sec
+epoch 1, total_step 26120, total loss is 33.44 , inference loss is 21.07, weight deacy loss is 12.37, training accuracy is 0.125000, time 213.331 samples/sec
+epoch 1, total_step 26140, total loss is 26.30 , inference loss is 13.93, weight deacy loss is 12.37, training accuracy is 0.125000, time 209.629 samples/sec
+epoch 1, total_step 26160, total loss is 28.89 , inference loss is 16.52, weight deacy loss is 12.37, training accuracy is 0.062500, time 236.743 samples/sec
+epoch 1, total_step 26180, total loss is 30.74 , inference loss is 18.38, weight deacy loss is 12.36, training accuracy is 0.125000, time 220.904 samples/sec
+epoch 1, total_step 26200, total loss is 31.26 , inference loss is 18.90, weight deacy loss is 12.36, training accuracy is 0.093750, time 213.303 samples/sec
+epoch 1, total_step 26220, total loss is 29.41 , inference loss is 17.05, weight deacy loss is 12.36, training accuracy is 0.093750, time 208.102 samples/sec
+epoch 1, total_step 26240, total loss is 30.08 , inference loss is 17.72, weight deacy loss is 12.36, training accuracy is 0.031250, time 216.873 samples/sec
+epoch 1, total_step 26260, total loss is 29.54 , inference loss is 17.18, weight deacy loss is 12.36, training accuracy is 0.093750, time 207.293 samples/sec
+epoch 1, total_step 26280, total loss is 29.34 , inference loss is 16.98, weight deacy loss is 12.36, training accuracy is 0.187500, time 300.743 samples/sec
+epoch 1, total_step 26300, total loss is 35.05 , inference loss is 22.69, weight deacy loss is 12.35, training accuracy is 0.062500, time 223.039 samples/sec
+epoch 1, total_step 26320, total loss is 34.29 , inference loss is 21.94, weight deacy loss is 12.35, training accuracy is 0.062500, time 232.602 samples/sec
+epoch 1, total_step 26340, total loss is 26.97 , inference loss is 14.62, weight deacy loss is 12.35, training accuracy is 0.062500, time 204.610 samples/sec
+epoch 1, total_step 26360, total loss is 29.00 , inference loss is 16.65, weight deacy loss is 12.35, training accuracy is 0.187500, time 212.686 samples/sec
+epoch 1, total_step 26380, total loss is 31.75 , inference loss is 19.41, weight deacy loss is 12.35, training accuracy is 0.125000, time 246.929 samples/sec
+epoch 1, total_step 26400, total loss is 28.16 , inference loss is 15.82, weight deacy loss is 12.35, training accuracy is 0.218750, time 13.670 samples/sec
+epoch 1, total_step 26420, total loss is 27.79 , inference loss is 15.44, weight deacy loss is 12.34, training accuracy is 0.125000, time 218.628 samples/sec
+epoch 1, total_step 26440, total loss is 25.96 , inference loss is 13.62, weight deacy loss is 12.34, training accuracy is 0.218750, time 224.817 samples/sec
+epoch 1, total_step 26460, total loss is 29.32 , inference loss is 16.98, weight deacy loss is 12.34, training accuracy is 0.093750, time 227.245 samples/sec
+epoch 1, total_step 26480, total loss is 29.69 , inference loss is 17.35, weight deacy loss is 12.34, training accuracy is 0.125000, time 201.601 samples/sec
+epoch 1, total_step 26500, total loss is 31.09 , inference loss is 18.75, weight deacy loss is 12.34, training accuracy is 0.312500, time 200.634 samples/sec
+epoch 1, total_step 26520, total loss is 33.50 , inference loss is 21.17, weight deacy loss is 12.34, training accuracy is 0.125000, time 213.920 samples/sec
+epoch 1, total_step 26540, total loss is 32.25 , inference loss is 19.91, weight deacy loss is 12.33, training accuracy is 0.062500, time 297.370 samples/sec
+epoch 1, total_step 26560, total loss is 28.79 , inference loss is 16.46, weight deacy loss is 12.33, training accuracy is 0.062500, time 212.291 samples/sec
+epoch 1, total_step 26580, total loss is 29.35 , inference loss is 17.02, weight deacy loss is 12.33, training accuracy is 0.125000, time 213.248 samples/sec
+epoch 1, total_step 26600, total loss is 29.78 , inference loss is 17.45, weight deacy loss is 12.33, training accuracy is 0.062500, time 216.912 samples/sec
+epoch 1, total_step 26620, total loss is 30.99 , inference loss is 18.66, weight deacy loss is 12.33, training accuracy is 0.000000, time 225.667 samples/sec
+epoch 1, total_step 26640, total loss is 28.25 , inference loss is 15.92, weight deacy loss is 12.33, training accuracy is 0.125000, time 210.765 samples/sec
+epoch 1, total_step 26660, total loss is 30.63 , inference loss is 18.31, weight deacy loss is 12.32, training accuracy is 0.062500, time 217.315 samples/sec
+epoch 1, total_step 26680, total loss is 29.20 , inference loss is 16.87, weight deacy loss is 12.32, training accuracy is 0.125000, time 217.214 samples/sec
+epoch 1, total_step 26700, total loss is 36.25 , inference loss is 23.93, weight deacy loss is 12.32, training accuracy is 0.125000, time 14.622 samples/sec
+epoch 1, total_step 26720, total loss is 27.71 , inference loss is 15.39, weight deacy loss is 12.32, training accuracy is 0.125000, time 224.635 samples/sec
+epoch 1, total_step 26740, total loss is 30.75 , inference loss is 18.43, weight deacy loss is 12.32, training accuracy is 0.156250, time 207.230 samples/sec
+epoch 1, total_step 26760, total loss is 30.77 , inference loss is 18.46, weight deacy loss is 12.32, training accuracy is 0.156250, time 211.350 samples/sec
+epoch 1, total_step 26780, total loss is 28.94 , inference loss is 16.63, weight deacy loss is 12.31, training accuracy is 0.156250, time 202.739 samples/sec
+epoch 1, total_step 26800, total loss is 30.73 , inference loss is 18.42, weight deacy loss is 12.31, training accuracy is 0.093750, time 213.465 samples/sec
+epoch 1, total_step 26820, total loss is 31.08 , inference loss is 18.76, weight deacy loss is 12.31, training accuracy is 0.031250, time 193.115 samples/sec
+epoch 1, total_step 26840, total loss is 29.77 , inference loss is 17.46, weight deacy loss is 12.31, training accuracy is 0.093750, time 219.538 samples/sec
+epoch 1, total_step 26860, total loss is 26.32 , inference loss is 14.01, weight deacy loss is 12.31, training accuracy is 0.156250, time 209.659 samples/sec
+epoch 1, total_step 26880, total loss is 33.20 , inference loss is 20.89, weight deacy loss is 12.31, training accuracy is 0.031250, time 219.324 samples/sec
+epoch 1, total_step 26900, total loss is 33.87 , inference loss is 21.57, weight deacy loss is 12.30, training accuracy is 0.031250, time 232.307 samples/sec
+epoch 1, total_step 26920, total loss is 31.13 , inference loss is 18.82, weight deacy loss is 12.30, training accuracy is 0.093750, time 224.206 samples/sec
+epoch 1, total_step 26940, total loss is 29.96 , inference loss is 17.65, weight deacy loss is 12.30, training accuracy is 0.031250, time 214.672 samples/sec
+epoch 1, total_step 26960, total loss is 27.06 , inference loss is 14.76, weight deacy loss is 12.30, training accuracy is 0.156250, time 207.175 samples/sec
+epoch 1, total_step 26980, total loss is 30.50 , inference loss is 18.20, weight deacy loss is 12.30, training accuracy is 0.031250, time 222.122 samples/sec
+epoch 1, total_step 27000, total loss is 27.86 , inference loss is 15.56, weight deacy loss is 12.30, training accuracy is 0.062500, time 14.132 samples/sec
+epoch 1, total_step 27020, total loss is 30.19 , inference loss is 17.90, weight deacy loss is 12.29, training accuracy is 0.062500, time 222.724 samples/sec
+epoch 1, total_step 27040, total loss is 25.43 , inference loss is 13.14, weight deacy loss is 12.29, training accuracy is 0.031250, time 201.192 samples/sec
+epoch 1, total_step 27060, total loss is 32.88 , inference loss is 20.59, weight deacy loss is 12.29, training accuracy is 0.125000, time 219.073 samples/sec
+epoch 1, total_step 27080, total loss is 35.90 , inference loss is 23.61, weight deacy loss is 12.29, training accuracy is 0.000000, time 214.268 samples/sec
+epoch 1, total_step 27100, total loss is 32.74 , inference loss is 20.45, weight deacy loss is 12.29, training accuracy is 0.062500, time 228.158 samples/sec
+epoch 1, total_step 27120, total loss is 33.29 , inference loss is 21.00, weight deacy loss is 12.29, training accuracy is 0.031250, time 209.553 samples/sec
+epoch 1, total_step 27140, total loss is 33.33 , inference loss is 21.04, weight deacy loss is 12.28, training accuracy is 0.062500, time 209.325 samples/sec
+epoch 1, total_step 27160, total loss is 30.43 , inference loss is 18.15, weight deacy loss is 12.28, training accuracy is 0.093750, time 221.031 samples/sec
+epoch 1, total_step 27180, total loss is 29.13 , inference loss is 16.85, weight deacy loss is 12.28, training accuracy is 0.156250, time 208.138 samples/sec
+epoch 1, total_step 27200, total loss is 29.56 , inference loss is 17.28, weight deacy loss is 12.28, training accuracy is 0.125000, time 209.464 samples/sec
+epoch 1, total_step 27220, total loss is 28.66 , inference loss is 16.39, weight deacy loss is 12.28, training accuracy is 0.125000, time 234.142 samples/sec
+epoch 1, total_step 27240, total loss is 30.36 , inference loss is 18.08, weight deacy loss is 12.28, training accuracy is 0.000000, time 212.899 samples/sec
+epoch 1, total_step 27260, total loss is 31.52 , inference loss is 19.24, weight deacy loss is 12.27, training accuracy is 0.093750, time 298.563 samples/sec
+epoch 1, total_step 27280, total loss is 31.06 , inference loss is 18.79, weight deacy loss is 12.27, training accuracy is 0.125000, time 224.739 samples/sec
+epoch 1, total_step 27300, total loss is 31.63 , inference loss is 19.36, weight deacy loss is 12.27, training accuracy is 0.062500, time 13.564 samples/sec
+epoch 1, total_step 27320, total loss is 31.72 , inference loss is 19.45, weight deacy loss is 12.27, training accuracy is 0.125000, time 223.002 samples/sec
+epoch 1, total_step 27340, total loss is 33.50 , inference loss is 21.23, weight deacy loss is 12.27, training accuracy is 0.000000, time 215.978 samples/sec
+epoch 1, total_step 27360, total loss is 28.05 , inference loss is 15.78, weight deacy loss is 12.27, training accuracy is 0.093750, time 218.449 samples/sec
+epoch 1, total_step 27380, total loss is 29.85 , inference loss is 17.59, weight deacy loss is 12.27, training accuracy is 0.031250, time 220.126 samples/sec
+epoch 1, total_step 27400, total loss is 33.23 , inference loss is 20.97, weight deacy loss is 12.26, training accuracy is 0.125000, time 227.032 samples/sec
+epoch 1, total_step 27420, total loss is 27.37 , inference loss is 15.11, weight deacy loss is 12.26, training accuracy is 0.062500, time 216.866 samples/sec
+epoch 1, total_step 27440, total loss is 31.85 , inference loss is 19.59, weight deacy loss is 12.26, training accuracy is 0.031250, time 220.901 samples/sec
+epoch 1, total_step 27460, total loss is 30.29 , inference loss is 18.03, weight deacy loss is 12.26, training accuracy is 0.031250, time 236.895 samples/sec
+epoch 1, total_step 27480, total loss is 31.18 , inference loss is 18.93, weight deacy loss is 12.26, training accuracy is 0.125000, time 215.195 samples/sec
+epoch 1, total_step 27500, total loss is 30.90 , inference loss is 18.64, weight deacy loss is 12.26, training accuracy is 0.031250, time 224.845 samples/sec
+epoch 1, total_step 27520, total loss is 28.33 , inference loss is 16.07, weight deacy loss is 12.25, training accuracy is 0.093750, time 296.987 samples/sec
+epoch 1, total_step 27540, total loss is 31.99 , inference loss is 19.74, weight deacy loss is 12.25, training accuracy is 0.093750, time 212.743 samples/sec
+epoch 1, total_step 27560, total loss is 32.43 , inference loss is 20.18, weight deacy loss is 12.25, training accuracy is 0.093750, time 208.035 samples/sec
+epoch 1, total_step 27580, total loss is 30.41 , inference loss is 18.16, weight deacy loss is 12.25, training accuracy is 0.000000, time 214.236 samples/sec
+epoch 1, total_step 27600, total loss is 30.94 , inference loss is 18.70, weight deacy loss is 12.25, training accuracy is 0.187500, time 13.266 samples/sec
+epoch 1, total_step 27620, total loss is 33.94 , inference loss is 21.69, weight deacy loss is 12.25, training accuracy is 0.000000, time 210.814 samples/sec
+epoch 1, total_step 27640, total loss is 29.57 , inference loss is 17.33, weight deacy loss is 12.24, training accuracy is 0.031250, time 221.192 samples/sec
+epoch 1, total_step 27660, total loss is 29.73 , inference loss is 17.49, weight deacy loss is 12.24, training accuracy is 0.125000, time 212.095 samples/sec
+epoch 1, total_step 27680, total loss is 28.13 , inference loss is 15.89, weight deacy loss is 12.24, training accuracy is 0.093750, time 214.332 samples/sec
+epoch 1, total_step 27700, total loss is 31.77 , inference loss is 19.53, weight deacy loss is 12.24, training accuracy is 0.062500, time 202.611 samples/sec
+epoch 1, total_step 27720, total loss is 27.13 , inference loss is 14.89, weight deacy loss is 12.24, training accuracy is 0.156250, time 217.851 samples/sec
+epoch 1, total_step 27740, total loss is 33.15 , inference loss is 20.91, weight deacy loss is 12.24, training accuracy is 0.031250, time 205.897 samples/sec
+epoch 1, total_step 27760, total loss is 29.41 , inference loss is 17.17, weight deacy loss is 12.23, training accuracy is 0.062500, time 216.404 samples/sec
+epoch 1, total_step 27780, total loss is 31.93 , inference loss is 19.70, weight deacy loss is 12.23, training accuracy is 0.031250, time 297.814 samples/sec
+epoch 1, total_step 27800, total loss is 28.84 , inference loss is 16.61, weight deacy loss is 12.23, training accuracy is 0.062500, time 228.418 samples/sec
+epoch 1, total_step 27820, total loss is 31.53 , inference loss is 19.30, weight deacy loss is 12.23, training accuracy is 0.125000, time 216.124 samples/sec
+epoch 1, total_step 27840, total loss is 29.58 , inference loss is 17.36, weight deacy loss is 12.23, training accuracy is 0.125000, time 223.798 samples/sec
+epoch 1, total_step 27860, total loss is 29.96 , inference loss is 17.73, weight deacy loss is 12.23, training accuracy is 0.062500, time 239.121 samples/sec
+epoch 1, total_step 27880, total loss is 30.84 , inference loss is 18.62, weight deacy loss is 12.22, training accuracy is 0.093750, time 207.807 samples/sec
+epoch 1, total_step 27900, total loss is 31.51 , inference loss is 19.28, weight deacy loss is 12.22, training accuracy is 0.062500, time 13.825 samples/sec
+epoch 1, total_step 27920, total loss is 31.84 , inference loss is 19.62, weight deacy loss is 12.22, training accuracy is 0.000000, time 207.106 samples/sec
+epoch 1, total_step 27940, total loss is 31.83 , inference loss is 19.61, weight deacy loss is 12.22, training accuracy is 0.125000, time 226.067 samples/sec
+epoch 1, total_step 27960, total loss is 27.68 , inference loss is 15.46, weight deacy loss is 12.22, training accuracy is 0.125000, time 209.942 samples/sec
+epoch 1, total_step 27980, total loss is 30.76 , inference loss is 18.55, weight deacy loss is 12.22, training accuracy is 0.093750, time 218.997 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.042339
+best_threshold_index 100 0.9685185185185186
+best_threshold_index 100 0.9677777777777777
+best_threshold_index 100 0.9685185185185186
+best_threshold_index 100 0.9696296296296296
+best_threshold_index 100 0.9687037037037037
+best_threshold_index 100 0.9674074074074074
+best_threshold_index 100 0.9687037037037037
+best_threshold_index 100 0.9679629629629629
+best_threshold_index 100 0.9662962962962963
+best_threshold_index 100 0.9664814814814815
+[lfw][28000]XNorm: 24.494244
+[lfw][28000]Accuracy-Flip: 0.96800+-0.00891
+test accuracy is: 0.968
+epoch 1, total_step 28000, total loss is 30.92 , inference loss is 18.71, weight deacy loss is 12.21, training accuracy is 0.062500, time 16.578 samples/sec
+epoch 1, total_step 28020, total loss is 30.47 , inference loss is 18.26, weight deacy loss is 12.21, training accuracy is 0.093750, time 213.539 samples/sec
+epoch 1, total_step 28040, total loss is 26.57 , inference loss is 14.36, weight deacy loss is 12.21, training accuracy is 0.093750, time 221.739 samples/sec
+epoch 1, total_step 28060, total loss is 29.89 , inference loss is 17.68, weight deacy loss is 12.21, training accuracy is 0.093750, time 240.216 samples/sec
+epoch 1, total_step 28080, total loss is 28.52 , inference loss is 16.31, weight deacy loss is 12.21, training accuracy is 0.218750, time 223.668 samples/sec
+epoch 1, total_step 28100, total loss is 29.37 , inference loss is 17.17, weight deacy loss is 12.21, training accuracy is 0.125000, time 220.549 samples/sec
+epoch 1, total_step 28120, total loss is 30.28 , inference loss is 18.07, weight deacy loss is 12.21, training accuracy is 0.125000, time 196.746 samples/sec
+epoch 1, total_step 28140, total loss is 28.12 , inference loss is 15.92, weight deacy loss is 12.20, training accuracy is 0.125000, time 207.794 samples/sec
+epoch 1, total_step 28160, total loss is 30.70 , inference loss is 18.50, weight deacy loss is 12.20, training accuracy is 0.062500, time 300.351 samples/sec
+epoch 1, total_step 28180, total loss is 33.45 , inference loss is 21.25, weight deacy loss is 12.20, training accuracy is 0.125000, time 207.773 samples/sec
+epoch 1, total_step 28200, total loss is 31.45 , inference loss is 19.25, weight deacy loss is 12.20, training accuracy is 0.125000, time 14.322 samples/sec
+epoch 1, total_step 28220, total loss is 29.03 , inference loss is 16.84, weight deacy loss is 12.20, training accuracy is 0.218750, time 213.360 samples/sec
+epoch 1, total_step 28240, total loss is 31.97 , inference loss is 19.78, weight deacy loss is 12.20, training accuracy is 0.062500, time 239.799 samples/sec
+epoch 1, total_step 28260, total loss is 30.95 , inference loss is 18.75, weight deacy loss is 12.19, training accuracy is 0.093750, time 221.260 samples/sec
+epoch 1, total_step 28280, total loss is 31.36 , inference loss is 19.17, weight deacy loss is 12.19, training accuracy is 0.031250, time 221.565 samples/sec
+epoch 1, total_step 28300, total loss is 31.08 , inference loss is 18.89, weight deacy loss is 12.19, training accuracy is 0.093750, time 231.507 samples/sec
+epoch 1, total_step 28320, total loss is 32.29 , inference loss is 20.10, weight deacy loss is 12.19, training accuracy is 0.000000, time 214.684 samples/sec
+epoch 1, total_step 28340, total loss is 27.73 , inference loss is 15.54, weight deacy loss is 12.19, training accuracy is 0.093750, time 217.735 samples/sec
+epoch 1, total_step 28360, total loss is 29.04 , inference loss is 16.86, weight deacy loss is 12.19, training accuracy is 0.000000, time 236.173 samples/sec
+epoch 1, total_step 28380, total loss is 32.23 , inference loss is 20.04, weight deacy loss is 12.18, training accuracy is 0.062500, time 227.084 samples/sec
+epoch 1, total_step 28400, total loss is 32.69 , inference loss is 20.50, weight deacy loss is 12.18, training accuracy is 0.062500, time 216.644 samples/sec
+epoch 1, total_step 28420, total loss is 29.38 , inference loss is 17.20, weight deacy loss is 12.18, training accuracy is 0.093750, time 216.372 samples/sec
+epoch 1, total_step 28440, total loss is 33.03 , inference loss is 20.85, weight deacy loss is 12.18, training accuracy is 0.062500, time 223.834 samples/sec
+epoch 1, total_step 28460, total loss is 30.99 , inference loss is 18.81, weight deacy loss is 12.18, training accuracy is 0.093750, time 300.805 samples/sec
+epoch 1, total_step 28480, total loss is 31.66 , inference loss is 19.49, weight deacy loss is 12.18, training accuracy is 0.062500, time 299.018 samples/sec
+epoch 1, total_step 28500, total loss is 32.39 , inference loss is 20.21, weight deacy loss is 12.17, training accuracy is 0.031250, time 13.681 samples/sec
+epoch 1, total_step 28520, total loss is 33.84 , inference loss is 21.67, weight deacy loss is 12.17, training accuracy is 0.031250, time 199.557 samples/sec
+epoch 1, total_step 28540, total loss is 29.63 , inference loss is 17.46, weight deacy loss is 12.17, training accuracy is 0.031250, time 215.920 samples/sec
+epoch 1, total_step 28560, total loss is 34.23 , inference loss is 22.06, weight deacy loss is 12.17, training accuracy is 0.125000, time 207.825 samples/sec
+epoch 1, total_step 28580, total loss is 32.72 , inference loss is 20.55, weight deacy loss is 12.17, training accuracy is 0.000000, time 211.169 samples/sec
+epoch 1, total_step 28600, total loss is 34.92 , inference loss is 22.75, weight deacy loss is 12.17, training accuracy is 0.031250, time 206.830 samples/sec
+epoch 1, total_step 28620, total loss is 30.20 , inference loss is 18.03, weight deacy loss is 12.17, training accuracy is 0.062500, time 243.778 samples/sec
+epoch 1, total_step 28640, total loss is 27.00 , inference loss is 14.83, weight deacy loss is 12.16, training accuracy is 0.031250, time 214.654 samples/sec
+epoch 1, total_step 28660, total loss is 32.11 , inference loss is 19.95, weight deacy loss is 12.16, training accuracy is 0.062500, time 210.597 samples/sec
+epoch 1, total_step 28680, total loss is 30.44 , inference loss is 18.28, weight deacy loss is 12.16, training accuracy is 0.093750, time 228.802 samples/sec
+epoch 1, total_step 28700, total loss is 29.00 , inference loss is 16.84, weight deacy loss is 12.16, training accuracy is 0.125000, time 217.194 samples/sec
+epoch 1, total_step 28720, total loss is 28.64 , inference loss is 16.48, weight deacy loss is 12.16, training accuracy is 0.156250, time 232.617 samples/sec
+epoch 1, total_step 28740, total loss is 29.69 , inference loss is 17.53, weight deacy loss is 12.16, training accuracy is 0.062500, time 219.037 samples/sec
+epoch 1, total_step 28760, total loss is 30.52 , inference loss is 18.36, weight deacy loss is 12.15, training accuracy is 0.062500, time 220.203 samples/sec
+epoch 1, total_step 28780, total loss is 37.45 , inference loss is 25.30, weight deacy loss is 12.15, training accuracy is 0.031250, time 224.948 samples/sec
+epoch 1, total_step 28800, total loss is 29.17 , inference loss is 17.02, weight deacy loss is 12.15, training accuracy is 0.093750, time 13.124 samples/sec
+epoch 1, total_step 28820, total loss is 29.70 , inference loss is 17.55, weight deacy loss is 12.15, training accuracy is 0.156250, time 294.058 samples/sec
+epoch 1, total_step 28840, total loss is 33.46 , inference loss is 21.32, weight deacy loss is 12.15, training accuracy is 0.062500, time 200.855 samples/sec
+epoch 1, total_step 28860, total loss is 31.74 , inference loss is 19.60, weight deacy loss is 12.15, training accuracy is 0.062500, time 229.571 samples/sec
+epoch 1, total_step 28880, total loss is 30.90 , inference loss is 18.76, weight deacy loss is 12.14, training accuracy is 0.156250, time 207.997 samples/sec
+epoch 1, total_step 28900, total loss is 28.90 , inference loss is 16.76, weight deacy loss is 12.14, training accuracy is 0.093750, time 206.346 samples/sec
+epoch 1, total_step 28920, total loss is 28.66 , inference loss is 16.52, weight deacy loss is 12.14, training accuracy is 0.062500, time 213.502 samples/sec
+epoch 1, total_step 28940, total loss is 28.28 , inference loss is 16.14, weight deacy loss is 12.14, training accuracy is 0.156250, time 222.256 samples/sec
+epoch 1, total_step 28960, total loss is 33.84 , inference loss is 21.70, weight deacy loss is 12.14, training accuracy is 0.125000, time 208.045 samples/sec
+epoch 1, total_step 28980, total loss is 26.57 , inference loss is 14.43, weight deacy loss is 12.14, training accuracy is 0.093750, time 203.644 samples/sec
+epoch 1, total_step 29000, total loss is 32.07 , inference loss is 19.93, weight deacy loss is 12.13, training accuracy is 0.031250, time 200.264 samples/sec
+epoch 1, total_step 29020, total loss is 30.82 , inference loss is 18.69, weight deacy loss is 12.13, training accuracy is 0.062500, time 229.478 samples/sec
+epoch 1, total_step 29040, total loss is 34.59 , inference loss is 22.46, weight deacy loss is 12.13, training accuracy is 0.031250, time 229.003 samples/sec
+epoch 1, total_step 29060, total loss is 30.85 , inference loss is 18.72, weight deacy loss is 12.13, training accuracy is 0.093750, time 241.665 samples/sec
+epoch 1, total_step 29080, total loss is 28.34 , inference loss is 16.21, weight deacy loss is 12.13, training accuracy is 0.093750, time 209.783 samples/sec
+epoch 1, total_step 29100, total loss is 35.82 , inference loss is 23.70, weight deacy loss is 12.13, training accuracy is 0.000000, time 15.003 samples/sec
+epoch 1, total_step 29120, total loss is 28.72 , inference loss is 16.60, weight deacy loss is 12.13, training accuracy is 0.093750, time 223.226 samples/sec
+epoch 1, total_step 29140, total loss is 33.37 , inference loss is 21.25, weight deacy loss is 12.12, training accuracy is 0.093750, time 210.906 samples/sec
+epoch 1, total_step 29160, total loss is 34.40 , inference loss is 22.28, weight deacy loss is 12.12, training accuracy is 0.062500, time 205.446 samples/sec
+epoch 1, total_step 29180, total loss is 28.07 , inference loss is 15.95, weight deacy loss is 12.12, training accuracy is 0.093750, time 213.200 samples/sec
+epoch 1, total_step 29200, total loss is 30.54 , inference loss is 18.42, weight deacy loss is 12.12, training accuracy is 0.125000, time 227.029 samples/sec
+epoch 1, total_step 29220, total loss is 29.75 , inference loss is 17.64, weight deacy loss is 12.12, training accuracy is 0.031250, time 216.589 samples/sec
+epoch 1, total_step 29240, total loss is 28.55 , inference loss is 16.43, weight deacy loss is 12.12, training accuracy is 0.156250, time 227.861 samples/sec
+epoch 1, total_step 29260, total loss is 31.79 , inference loss is 19.68, weight deacy loss is 12.11, training accuracy is 0.031250, time 217.159 samples/sec
+epoch 1, total_step 29280, total loss is 35.23 , inference loss is 23.12, weight deacy loss is 12.11, training accuracy is 0.062500, time 203.270 samples/sec
+epoch 1, total_step 29300, total loss is 25.53 , inference loss is 13.42, weight deacy loss is 12.11, training accuracy is 0.125000, time 235.927 samples/sec
+epoch 1, total_step 29320, total loss is 35.20 , inference loss is 23.09, weight deacy loss is 12.11, training accuracy is 0.062500, time 215.308 samples/sec
+epoch 1, total_step 29340, total loss is 35.73 , inference loss is 23.62, weight deacy loss is 12.11, training accuracy is 0.031250, time 206.156 samples/sec
+epoch 1, total_step 29360, total loss is 29.01 , inference loss is 16.91, weight deacy loss is 12.11, training accuracy is 0.000000, time 224.986 samples/sec
+epoch 1, total_step 29380, total loss is 30.29 , inference loss is 18.18, weight deacy loss is 12.10, training accuracy is 0.062500, time 215.202 samples/sec
+epoch 1, total_step 29400, total loss is 30.90 , inference loss is 18.80, weight deacy loss is 12.10, training accuracy is 0.000000, time 13.588 samples/sec
+epoch 1, total_step 29420, total loss is 25.07 , inference loss is 12.97, weight deacy loss is 12.10, training accuracy is 0.093750, time 211.330 samples/sec
+epoch 1, total_step 29440, total loss is 30.08 , inference loss is 17.98, weight deacy loss is 12.10, training accuracy is 0.093750, time 218.052 samples/sec
+epoch 1, total_step 29460, total loss is 27.73 , inference loss is 15.63, weight deacy loss is 12.10, training accuracy is 0.093750, time 207.029 samples/sec
+epoch 1, total_step 29480, total loss is 33.78 , inference loss is 21.69, weight deacy loss is 12.10, training accuracy is 0.062500, time 205.267 samples/sec
+epoch 1, total_step 29500, total loss is 29.15 , inference loss is 17.06, weight deacy loss is 12.10, training accuracy is 0.093750, time 220.361 samples/sec
+epoch 1, total_step 29520, total loss is 32.68 , inference loss is 20.58, weight deacy loss is 12.09, training accuracy is 0.062500, time 221.796 samples/sec
+epoch 1, total_step 29540, total loss is 32.46 , inference loss is 20.37, weight deacy loss is 12.09, training accuracy is 0.031250, time 227.689 samples/sec
+epoch 1, total_step 29560, total loss is 29.07 , inference loss is 16.98, weight deacy loss is 12.09, training accuracy is 0.062500, time 235.399 samples/sec
+epoch 1, total_step 29580, total loss is 28.78 , inference loss is 16.69, weight deacy loss is 12.09, training accuracy is 0.062500, time 224.722 samples/sec
+epoch 1, total_step 29600, total loss is 33.04 , inference loss is 20.95, weight deacy loss is 12.09, training accuracy is 0.031250, time 177.349 samples/sec
+epoch 1, total_step 29620, total loss is 30.97 , inference loss is 18.88, weight deacy loss is 12.09, training accuracy is 0.125000, time 197.462 samples/sec
+epoch 1, total_step 29640, total loss is 31.49 , inference loss is 19.41, weight deacy loss is 12.08, training accuracy is 0.062500, time 223.133 samples/sec
+epoch 1, total_step 29660, total loss is 29.71 , inference loss is 17.63, weight deacy loss is 12.08, training accuracy is 0.062500, time 296.068 samples/sec
+epoch 1, total_step 29680, total loss is 27.81 , inference loss is 15.73, weight deacy loss is 12.08, training accuracy is 0.093750, time 202.727 samples/sec
+epoch 1, total_step 29700, total loss is 31.85 , inference loss is 19.77, weight deacy loss is 12.08, training accuracy is 0.062500, time 11.713 samples/sec
+epoch 1, total_step 29720, total loss is 32.35 , inference loss is 20.27, weight deacy loss is 12.08, training accuracy is 0.062500, time 207.637 samples/sec
+epoch 1, total_step 29740, total loss is 32.47 , inference loss is 20.39, weight deacy loss is 12.08, training accuracy is 0.000000, time 217.585 samples/sec
+epoch 1, total_step 29760, total loss is 26.80 , inference loss is 14.73, weight deacy loss is 12.07, training accuracy is 0.093750, time 219.711 samples/sec
+epoch 1, total_step 29780, total loss is 32.77 , inference loss is 20.70, weight deacy loss is 12.07, training accuracy is 0.000000, time 215.001 samples/sec
+epoch 1, total_step 29800, total loss is 28.38 , inference loss is 16.31, weight deacy loss is 12.07, training accuracy is 0.062500, time 220.964 samples/sec
+epoch 1, total_step 29820, total loss is 29.10 , inference loss is 17.03, weight deacy loss is 12.07, training accuracy is 0.031250, time 196.275 samples/sec
+epoch 1, total_step 29840, total loss is 28.39 , inference loss is 16.32, weight deacy loss is 12.07, training accuracy is 0.000000, time 216.648 samples/sec
+epoch 1, total_step 29860, total loss is 30.25 , inference loss is 18.19, weight deacy loss is 12.07, training accuracy is 0.093750, time 214.559 samples/sec
+epoch 1, total_step 29880, total loss is 33.74 , inference loss is 21.67, weight deacy loss is 12.07, training accuracy is 0.062500, time 214.658 samples/sec
+epoch 1, total_step 29900, total loss is 29.14 , inference loss is 17.08, weight deacy loss is 12.06, training accuracy is 0.000000, time 218.819 samples/sec
+epoch 1, total_step 29920, total loss is 31.86 , inference loss is 19.80, weight deacy loss is 12.06, training accuracy is 0.000000, time 295.834 samples/sec
+epoch 1, total_step 29940, total loss is 29.26 , inference loss is 17.20, weight deacy loss is 12.06, training accuracy is 0.031250, time 207.408 samples/sec
+epoch 1, total_step 29960, total loss is 29.38 , inference loss is 17.32, weight deacy loss is 12.06, training accuracy is 0.125000, time 208.918 samples/sec
+epoch 1, total_step 29980, total loss is 26.87 , inference loss is 14.81, weight deacy loss is 12.06, training accuracy is 0.031250, time 203.281 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.820756999999972
+best_threshold_index 98 0.9662962962962963
+best_threshold_index 101 0.9653703703703703
+best_threshold_index 98 0.9666666666666667
+best_threshold_index 101 0.967037037037037
+best_threshold_index 98 0.9662962962962963
+best_threshold_index 101 0.9640740740740741
+best_threshold_index 101 0.9662962962962963
+best_threshold_index 98 0.9651851851851851
+best_threshold_index 101 0.9640740740740741
+best_threshold_index 101 0.9642592592592593
+[lfw][30000]XNorm: 24.312305
+[lfw][30000]Accuracy-Flip: 0.96367+-0.01038
+test accuracy is: 0.9636666666666667
+epoch 1, total_step 30000, total loss is 30.46 , inference loss is 18.41, weight deacy loss is 12.06, training accuracy is 0.062500, time 17.490 samples/sec
+epoch 1, total_step 30020, total loss is 31.43 , inference loss is 19.37, weight deacy loss is 12.05, training accuracy is 0.031250, time 207.195 samples/sec
+epoch 1, total_step 30040, total loss is 29.87 , inference loss is 17.81, weight deacy loss is 12.05, training accuracy is 0.031250, time 282.438 samples/sec
+epoch 1, total_step 30060, total loss is 26.80 , inference loss is 14.75, weight deacy loss is 12.05, training accuracy is 0.187500, time 297.333 samples/sec
+epoch 1, total_step 30080, total loss is 30.61 , inference loss is 18.56, weight deacy loss is 12.05, training accuracy is 0.031250, time 217.101 samples/sec
+epoch 1, total_step 30100, total loss is 30.83 , inference loss is 18.78, weight deacy loss is 12.05, training accuracy is 0.093750, time 218.035 samples/sec
+epoch 1, total_step 30120, total loss is 30.24 , inference loss is 18.20, weight deacy loss is 12.05, training accuracy is 0.062500, time 215.719 samples/sec
+epoch 1, total_step 30140, total loss is 32.53 , inference loss is 20.49, weight deacy loss is 12.04, training accuracy is 0.000000, time 223.021 samples/sec
+epoch 1, total_step 30160, total loss is 28.02 , inference loss is 15.98, weight deacy loss is 12.04, training accuracy is 0.062500, time 211.141 samples/sec
+epoch 1, total_step 30180, total loss is 32.75 , inference loss is 20.71, weight deacy loss is 12.04, training accuracy is 0.000000, time 208.010 samples/sec
+epoch 1, total_step 30200, total loss is 27.11 , inference loss is 15.07, weight deacy loss is 12.04, training accuracy is 0.093750, time 225.703 samples/sec
+epoch 1, total_step 30220, total loss is 31.33 , inference loss is 19.29, weight deacy loss is 12.04, training accuracy is 0.093750, time 292.195 samples/sec
+epoch 1, total_step 30240, total loss is 32.14 , inference loss is 20.10, weight deacy loss is 12.04, training accuracy is 0.031250, time 297.867 samples/sec
+epoch 1, total_step 30260, total loss is 31.51 , inference loss is 19.47, weight deacy loss is 12.04, training accuracy is 0.031250, time 223.619 samples/sec
+epoch 1, total_step 30280, total loss is 32.03 , inference loss is 19.99, weight deacy loss is 12.03, training accuracy is 0.062500, time 225.005 samples/sec2022-12-01 02:14:12.420604: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7472 of 10000
+2022-12-01 02:14:15.188531: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 1, total_step 30300, total loss is 30.89 , inference loss is 18.86, weight deacy loss is 12.03, training accuracy is 0.062500, time 13.844 samples/sec
+epoch 1, total_step 30320, total loss is 30.80 , inference loss is 18.77, weight deacy loss is 12.03, training accuracy is 0.062500, time 218.722 samples/sec
+epoch 1, total_step 30340, total loss is 32.98 , inference loss is 20.95, weight deacy loss is 12.03, training accuracy is 0.031250, time 224.483 samples/sec
+epoch 1, total_step 30360, total loss is 27.75 , inference loss is 15.73, weight deacy loss is 12.03, training accuracy is 0.062500, time 198.662 samples/sec
+epoch 1, total_step 30380, total loss is 30.56 , inference loss is 18.53, weight deacy loss is 12.03, training accuracy is 0.093750, time 209.265 samples/sec
+epoch 1, total_step 30400, total loss is 28.45 , inference loss is 16.43, weight deacy loss is 12.02, training accuracy is 0.062500, time 179.320 samples/sec
+epoch 1, total_step 30420, total loss is 29.75 , inference loss is 17.73, weight deacy loss is 12.02, training accuracy is 0.031250, time 238.921 samples/sec
+epoch 1, total_step 30440, total loss is 28.31 , inference loss is 16.29, weight deacy loss is 12.02, training accuracy is 0.062500, time 296.394 samples/sec
+epoch 1, total_step 30460, total loss is 28.40 , inference loss is 16.38, weight deacy loss is 12.02, training accuracy is 0.156250, time 196.670 samples/sec
+epoch 1, total_step 30480, total loss is 29.95 , inference loss is 17.93, weight deacy loss is 12.02, training accuracy is 0.125000, time 194.031 samples/sec
+epoch 1, total_step 30500, total loss is 31.63 , inference loss is 19.61, weight deacy loss is 12.02, training accuracy is 0.000000, time 215.180 samples/sec
+epoch 1, total_step 30520, total loss is 27.49 , inference loss is 15.48, weight deacy loss is 12.02, training accuracy is 0.093750, time 200.289 samples/sec
+epoch 1, total_step 30540, total loss is 29.39 , inference loss is 17.37, weight deacy loss is 12.01, training accuracy is 0.031250, time 198.545 samples/sec
+epoch 1, total_step 30560, total loss is 31.39 , inference loss is 19.38, weight deacy loss is 12.01, training accuracy is 0.062500, time 175.325 samples/sec
+epoch 1, total_step 30580, total loss is 28.24 , inference loss is 16.23, weight deacy loss is 12.01, training accuracy is 0.062500, time 203.883 samples/sec
+epoch 1, total_step 30600, total loss is 28.27 , inference loss is 16.26, weight deacy loss is 12.01, training accuracy is 0.000000, time 12.438 samples/sec
+epoch 1, total_step 30620, total loss is 31.64 , inference loss is 19.63, weight deacy loss is 12.01, training accuracy is 0.031250, time 186.294 samples/sec
+epoch 1, total_step 30640, total loss is 30.54 , inference loss is 18.54, weight deacy loss is 12.01, training accuracy is 0.031250, time 206.744 samples/sec
+epoch 1, total_step 30660, total loss is 28.98 , inference loss is 16.98, weight deacy loss is 12.00, training accuracy is 0.031250, time 194.950 samples/sec
+End of epoch 1
+epoch 2, total_step 30680, total loss is 35.71 , inference loss is 23.71, weight deacy loss is 12.00, training accuracy is 0.000000, time 216.983 samples/sec
+epoch 2, total_step 30700, total loss is 24.20 , inference loss is 12.20, weight deacy loss is 12.00, training accuracy is 0.281250, time 207.450 samples/sec
+epoch 2, total_step 30720, total loss is 25.12 , inference loss is 13.12, weight deacy loss is 12.00, training accuracy is 0.187500, time 224.445 samples/sec
+epoch 2, total_step 30740, total loss is 26.46 , inference loss is 14.47, weight deacy loss is 12.00, training accuracy is 0.187500, time 220.525 samples/sec
+epoch 2, total_step 30760, total loss is 29.88 , inference loss is 17.88, weight deacy loss is 12.00, training accuracy is 0.093750, time 205.587 samples/sec
+epoch 2, total_step 30780, total loss is 23.87 , inference loss is 11.87, weight deacy loss is 12.00, training accuracy is 0.312500, time 207.547 samples/sec
+epoch 2, total_step 30800, total loss is 22.84 , inference loss is 10.85, weight deacy loss is 11.99, training accuracy is 0.375000, time 219.085 samples/sec
+epoch 2, total_step 30820, total loss is 18.33 , inference loss is 6.34, weight deacy loss is 11.99, training accuracy is 0.562500, time 211.922 samples/sec
+epoch 2, total_step 30840, total loss is 21.21 , inference loss is 9.22, weight deacy loss is 11.99, training accuracy is 0.437500, time 206.380 samples/sec
+epoch 2, total_step 30860, total loss is 23.89 , inference loss is 11.90, weight deacy loss is 11.99, training accuracy is 0.218750, time 210.943 samples/sec
+epoch 2, total_step 30880, total loss is 23.34 , inference loss is 11.35, weight deacy loss is 11.99, training accuracy is 0.343750, time 206.891 samples/sec
+epoch 2, total_step 30900, total loss is 22.32 , inference loss is 10.33, weight deacy loss is 11.98, training accuracy is 0.375000, time 13.399 samples/sec
+epoch 2, total_step 30920, total loss is 27.53 , inference loss is 15.55, weight deacy loss is 11.98, training accuracy is 0.250000, time 297.722 samples/sec
+epoch 2, total_step 30940, total loss is 29.51 , inference loss is 17.53, weight deacy loss is 11.98, training accuracy is 0.125000, time 210.588 samples/sec
+epoch 2, total_step 30960, total loss is 30.87 , inference loss is 18.89, weight deacy loss is 11.98, training accuracy is 0.218750, time 213.653 samples/sec
+epoch 2, total_step 30980, total loss is 28.52 , inference loss is 16.54, weight deacy loss is 11.98, training accuracy is 0.156250, time 225.150 samples/sec
+epoch 2, total_step 31000, total loss is 26.86 , inference loss is 14.89, weight deacy loss is 11.98, training accuracy is 0.218750, time 204.898 samples/sec
+epoch 2, total_step 31020, total loss is 26.63 , inference loss is 14.66, weight deacy loss is 11.97, training accuracy is 0.250000, time 220.556 samples/sec
+epoch 2, total_step 31040, total loss is 23.08 , inference loss is 11.10, weight deacy loss is 11.97, training accuracy is 0.343750, time 232.760 samples/sec
+epoch 2, total_step 31060, total loss is 24.44 , inference loss is 12.47, weight deacy loss is 11.97, training accuracy is 0.281250, time 216.370 samples/sec
+epoch 2, total_step 31080, total loss is 27.86 , inference loss is 15.89, weight deacy loss is 11.97, training accuracy is 0.218750, time 228.707 samples/sec
+epoch 2, total_step 31100, total loss is 31.89 , inference loss is 19.92, weight deacy loss is 11.97, training accuracy is 0.218750, time 220.840 samples/sec
+epoch 2, total_step 31120, total loss is 26.57 , inference loss is 14.60, weight deacy loss is 11.97, training accuracy is 0.187500, time 195.224 samples/sec
+epoch 2, total_step 31140, total loss is 21.69 , inference loss is 9.73, weight deacy loss is 11.96, training accuracy is 0.218750, time 224.079 samples/sec
+epoch 2, total_step 31160, total loss is 27.19 , inference loss is 15.23, weight deacy loss is 11.96, training accuracy is 0.250000, time 212.757 samples/sec
+epoch 2, total_step 31180, total loss is 28.02 , inference loss is 16.06, weight deacy loss is 11.96, training accuracy is 0.281250, time 234.813 samples/sec
+epoch 2, total_step 31200, total loss is 22.04 , inference loss is 10.08, weight deacy loss is 11.96, training accuracy is 0.375000, time 15.186 samples/sec
+epoch 2, total_step 31220, total loss is 26.49 , inference loss is 14.53, weight deacy loss is 11.96, training accuracy is 0.156250, time 225.066 samples/sec
+epoch 2, total_step 31240, total loss is 22.28 , inference loss is 10.33, weight deacy loss is 11.96, training accuracy is 0.375000, time 244.222 samples/sec
+epoch 2, total_step 31260, total loss is 27.30 , inference loss is 15.34, weight deacy loss is 11.96, training accuracy is 0.187500, time 225.819 samples/sec
+epoch 2, total_step 31280, total loss is 26.87 , inference loss is 14.92, weight deacy loss is 11.95, training accuracy is 0.281250, time 212.931 samples/sec
+epoch 2, total_step 31300, total loss is 28.42 , inference loss is 16.47, weight deacy loss is 11.95, training accuracy is 0.156250, time 213.421 samples/sec
+epoch 2, total_step 31320, total loss is 30.20 , inference loss is 18.25, weight deacy loss is 11.95, training accuracy is 0.187500, time 224.438 samples/sec
+epoch 2, total_step 31340, total loss is 29.52 , inference loss is 17.57, weight deacy loss is 11.95, training accuracy is 0.250000, time 225.336 samples/sec
+epoch 2, total_step 31360, total loss is 26.66 , inference loss is 14.72, weight deacy loss is 11.95, training accuracy is 0.187500, time 217.539 samples/sec
+epoch 2, total_step 31380, total loss is 25.04 , inference loss is 13.09, weight deacy loss is 11.95, training accuracy is 0.156250, time 224.403 samples/sec
+epoch 2, total_step 31400, total loss is 26.99 , inference loss is 15.05, weight deacy loss is 11.94, training accuracy is 0.218750, time 230.749 samples/sec
+epoch 2, total_step 31420, total loss is 25.35 , inference loss is 13.41, weight deacy loss is 11.94, training accuracy is 0.156250, time 229.296 samples/sec
+epoch 2, total_step 31440, total loss is 29.32 , inference loss is 17.38, weight deacy loss is 11.94, training accuracy is 0.187500, time 222.774 samples/sec
+epoch 2, total_step 31460, total loss is 25.64 , inference loss is 13.70, weight deacy loss is 11.94, training accuracy is 0.156250, time 199.530 samples/sec
+epoch 2, total_step 31480, total loss is 29.48 , inference loss is 17.54, weight deacy loss is 11.94, training accuracy is 0.093750, time 214.993 samples/sec
+epoch 2, total_step 31500, total loss is 30.98 , inference loss is 19.05, weight deacy loss is 11.94, training accuracy is 0.218750, time 13.630 samples/sec
+epoch 2, total_step 31520, total loss is 26.80 , inference loss is 14.87, weight deacy loss is 11.93, training accuracy is 0.125000, time 218.853 samples/sec
+epoch 2, total_step 31540, total loss is 29.59 , inference loss is 17.66, weight deacy loss is 11.93, training accuracy is 0.093750, time 228.407 samples/sec
+epoch 2, total_step 31560, total loss is 28.54 , inference loss is 16.61, weight deacy loss is 11.93, training accuracy is 0.062500, time 220.003 samples/sec
+epoch 2, total_step 31580, total loss is 27.78 , inference loss is 15.85, weight deacy loss is 11.93, training accuracy is 0.250000, time 215.451 samples/sec
+epoch 2, total_step 31600, total loss is 31.20 , inference loss is 19.28, weight deacy loss is 11.93, training accuracy is 0.125000, time 215.499 samples/sec
+epoch 2, total_step 31620, total loss is 30.30 , inference loss is 18.38, weight deacy loss is 11.93, training accuracy is 0.156250, time 217.469 samples/sec
+epoch 2, total_step 31640, total loss is 26.17 , inference loss is 14.25, weight deacy loss is 11.92, training accuracy is 0.125000, time 196.000 samples/sec
+epoch 2, total_step 31660, total loss is 25.96 , inference loss is 14.04, weight deacy loss is 11.92, training accuracy is 0.218750, time 218.664 samples/sec
+epoch 2, total_step 31680, total loss is 30.42 , inference loss is 18.50, weight deacy loss is 11.92, training accuracy is 0.093750, time 217.967 samples/sec
+epoch 2, total_step 31700, total loss is 25.41 , inference loss is 13.49, weight deacy loss is 11.92, training accuracy is 0.187500, time 216.818 samples/sec
+epoch 2, total_step 31720, total loss is 26.19 , inference loss is 14.27, weight deacy loss is 11.92, training accuracy is 0.281250, time 220.922 samples/sec
+epoch 2, total_step 31740, total loss is 26.10 , inference loss is 14.18, weight deacy loss is 11.92, training accuracy is 0.250000, time 224.182 samples/sec
+epoch 2, total_step 31760, total loss is 25.51 , inference loss is 13.59, weight deacy loss is 11.91, training accuracy is 0.218750, time 293.086 samples/sec
+epoch 2, total_step 31780, total loss is 29.58 , inference loss is 17.67, weight deacy loss is 11.91, training accuracy is 0.125000, time 200.006 samples/sec
+epoch 2, total_step 31800, total loss is 25.81 , inference loss is 13.90, weight deacy loss is 11.91, training accuracy is 0.281250, time 11.854 samples/sec
+epoch 2, total_step 31820, total loss is 28.50 , inference loss is 16.59, weight deacy loss is 11.91, training accuracy is 0.156250, time 218.688 samples/sec
+epoch 2, total_step 31840, total loss is 31.86 , inference loss is 19.95, weight deacy loss is 11.91, training accuracy is 0.125000, time 222.744 samples/sec
+epoch 2, total_step 31860, total loss is 28.36 , inference loss is 16.45, weight deacy loss is 11.91, training accuracy is 0.093750, time 213.316 samples/sec
+epoch 2, total_step 31880, total loss is 25.86 , inference loss is 13.96, weight deacy loss is 11.90, training accuracy is 0.250000, time 215.053 samples/sec
+epoch 2, total_step 31900, total loss is 26.90 , inference loss is 15.00, weight deacy loss is 11.90, training accuracy is 0.250000, time 223.435 samples/sec
+epoch 2, total_step 31920, total loss is 28.03 , inference loss is 16.13, weight deacy loss is 11.90, training accuracy is 0.187500, time 216.700 samples/sec
+epoch 2, total_step 31940, total loss is 28.20 , inference loss is 16.30, weight deacy loss is 11.90, training accuracy is 0.218750, time 216.802 samples/sec
+epoch 2, total_step 31960, total loss is 30.18 , inference loss is 18.28, weight deacy loss is 11.90, training accuracy is 0.093750, time 232.843 samples/sec
+epoch 2, total_step 31980, total loss is 23.79 , inference loss is 11.90, weight deacy loss is 11.90, training accuracy is 0.312500, time 211.393 samples/sec
+testing verification..
+(12000, 512)
+infer time 22.625037000000017
+best_threshold_index 110 0.975925925925926
+best_threshold_index 110 0.9755555555555555
+best_threshold_index 110 0.9757407407407407
+best_threshold_index 110 0.9762962962962963
+best_threshold_index 110 0.9761111111111112
+best_threshold_index 110 0.9744444444444444
+best_threshold_index 105 0.9753703703703703
+best_threshold_index 110 0.9748148148148148
+best_threshold_index 110 0.9733333333333334
+best_threshold_index 110 0.9740740740740741
+[lfw][32000]XNorm: 23.802170
+[lfw][32000]Accuracy-Flip: 0.97467+-0.00852
+test accuracy is: 0.9746666666666666
+epoch 2, total_step 32000, total loss is 25.87 , inference loss is 13.98, weight deacy loss is 11.89, training accuracy is 0.281250, time 20.781 samples/sec
+epoch 2, total_step 32020, total loss is 27.59 , inference loss is 15.70, weight deacy loss is 11.89, training accuracy is 0.156250, time 213.787 samples/sec
+epoch 2, total_step 32040, total loss is 22.98 , inference loss is 11.09, weight deacy loss is 11.89, training accuracy is 0.343750, time 227.012 samples/sec
+epoch 2, total_step 32060, total loss is 26.34 , inference loss is 14.45, weight deacy loss is 11.89, training accuracy is 0.125000, time 205.164 samples/sec
+epoch 2, total_step 32080, total loss is 26.77 , inference loss is 14.88, weight deacy loss is 11.89, training accuracy is 0.031250, time 216.560 samples/sec
+epoch 2, total_step 32100, total loss is 27.45 , inference loss is 15.57, weight deacy loss is 11.89, training accuracy is 0.250000, time 12.498 samples/sec
+epoch 2, total_step 32120, total loss is 27.42 , inference loss is 15.53, weight deacy loss is 11.89, training accuracy is 0.187500, time 220.499 samples/sec
+epoch 2, total_step 32140, total loss is 30.99 , inference loss is 19.11, weight deacy loss is 11.88, training accuracy is 0.187500, time 217.160 samples/sec
+epoch 2, total_step 32160, total loss is 25.28 , inference loss is 13.40, weight deacy loss is 11.88, training accuracy is 0.218750, time 220.694 samples/sec
+epoch 2, total_step 32180, total loss is 28.01 , inference loss is 16.13, weight deacy loss is 11.88, training accuracy is 0.218750, time 216.771 samples/sec
+epoch 2, total_step 32200, total loss is 32.79 , inference loss is 20.91, weight deacy loss is 11.88, training accuracy is 0.093750, time 223.788 samples/sec
+epoch 2, total_step 32220, total loss is 29.44 , inference loss is 17.56, weight deacy loss is 11.88, training accuracy is 0.125000, time 224.405 samples/sec
+epoch 2, total_step 32240, total loss is 31.15 , inference loss is 19.27, weight deacy loss is 11.88, training accuracy is 0.187500, time 204.580 samples/sec
+epoch 2, total_step 32260, total loss is 30.26 , inference loss is 18.39, weight deacy loss is 11.87, training accuracy is 0.062500, time 298.994 samples/sec
+epoch 2, total_step 32280, total loss is 26.64 , inference loss is 14.76, weight deacy loss is 11.87, training accuracy is 0.156250, time 207.349 samples/sec
+epoch 2, total_step 32300, total loss is 27.37 , inference loss is 15.50, weight deacy loss is 11.87, training accuracy is 0.093750, time 218.787 samples/sec
+epoch 2, total_step 32320, total loss is 26.19 , inference loss is 14.32, weight deacy loss is 11.87, training accuracy is 0.093750, time 194.453 samples/sec
+epoch 2, total_step 32340, total loss is 31.04 , inference loss is 19.17, weight deacy loss is 11.87, training accuracy is 0.125000, time 212.436 samples/sec
+epoch 2, total_step 32360, total loss is 24.32 , inference loss is 12.45, weight deacy loss is 11.87, training accuracy is 0.250000, time 213.860 samples/sec
+epoch 2, total_step 32380, total loss is 21.09 , inference loss is 9.23, weight deacy loss is 11.86, training accuracy is 0.343750, time 219.900 samples/sec
+epoch 2, total_step 32400, total loss is 29.56 , inference loss is 17.70, weight deacy loss is 11.86, training accuracy is 0.125000, time 12.951 samples/sec
+epoch 2, total_step 32420, total loss is 28.73 , inference loss is 16.87, weight deacy loss is 11.86, training accuracy is 0.218750, time 214.007 samples/sec
+epoch 2, total_step 32440, total loss is 38.39 , inference loss is 26.53, weight deacy loss is 11.86, training accuracy is 0.062500, time 240.051 samples/sec
+epoch 2, total_step 32460, total loss is 26.55 , inference loss is 14.69, weight deacy loss is 11.86, training accuracy is 0.125000, time 217.095 samples/sec
+epoch 2, total_step 32480, total loss is 25.85 , inference loss is 13.99, weight deacy loss is 11.86, training accuracy is 0.218750, time 215.744 samples/sec
+epoch 2, total_step 32500, total loss is 25.98 , inference loss is 14.12, weight deacy loss is 11.85, training accuracy is 0.156250, time 217.295 samples/sec
+epoch 2, total_step 32520, total loss is 25.88 , inference loss is 14.03, weight deacy loss is 11.85, training accuracy is 0.125000, time 204.286 samples/sec
+epoch 2, total_step 32540, total loss is 25.20 , inference loss is 13.35, weight deacy loss is 11.85, training accuracy is 0.250000, time 235.489 samples/sec
+epoch 2, total_step 32560, total loss is 26.23 , inference loss is 14.38, weight deacy loss is 11.85, training accuracy is 0.218750, time 225.711 samples/sec
+epoch 2, total_step 32580, total loss is 23.09 , inference loss is 11.24, weight deacy loss is 11.85, training accuracy is 0.343750, time 228.749 samples/sec
+epoch 2, total_step 32600, total loss is 25.67 , inference loss is 13.83, weight deacy loss is 11.85, training accuracy is 0.281250, time 218.834 samples/sec
+epoch 2, total_step 32620, total loss is 26.05 , inference loss is 14.21, weight deacy loss is 11.84, training accuracy is 0.250000, time 225.645 samples/sec
+epoch 2, total_step 32640, total loss is 24.15 , inference loss is 12.31, weight deacy loss is 11.84, training accuracy is 0.312500, time 225.005 samples/sec
+epoch 2, total_step 32660, total loss is 29.12 , inference loss is 17.28, weight deacy loss is 11.84, training accuracy is 0.156250, time 212.610 samples/sec
+epoch 2, total_step 32680, total loss is 27.85 , inference loss is 16.02, weight deacy loss is 11.84, training accuracy is 0.281250, time 228.767 samples/sec
+epoch 2, total_step 32700, total loss is 22.54 , inference loss is 10.71, weight deacy loss is 11.84, training accuracy is 0.281250, time 13.440 samples/sec
+epoch 2, total_step 32720, total loss is 26.34 , inference loss is 14.50, weight deacy loss is 11.84, training accuracy is 0.187500, time 210.833 samples/sec
+epoch 2, total_step 32740, total loss is 23.85 , inference loss is 12.01, weight deacy loss is 11.83, training accuracy is 0.281250, time 213.314 samples/sec
+epoch 2, total_step 32760, total loss is 24.90 , inference loss is 13.06, weight deacy loss is 11.83, training accuracy is 0.187500, time 225.549 samples/sec
+epoch 2, total_step 32780, total loss is 28.58 , inference loss is 16.75, weight deacy loss is 11.83, training accuracy is 0.062500, time 214.210 samples/sec
+epoch 2, total_step 32800, total loss is 26.43 , inference loss is 14.60, weight deacy loss is 11.83, training accuracy is 0.125000, time 297.092 samples/sec
+epoch 2, total_step 32820, total loss is 25.52 , inference loss is 13.69, weight deacy loss is 11.83, training accuracy is 0.281250, time 227.463 samples/sec
+epoch 2, total_step 32840, total loss is 27.61 , inference loss is 15.78, weight deacy loss is 11.83, training accuracy is 0.187500, time 218.714 samples/sec
+epoch 2, total_step 32860, total loss is 26.00 , inference loss is 14.18, weight deacy loss is 11.82, training accuracy is 0.218750, time 218.637 samples/sec
+epoch 2, total_step 32880, total loss is 24.58 , inference loss is 12.75, weight deacy loss is 11.82, training accuracy is 0.187500, time 206.024 samples/sec
+epoch 2, total_step 32900, total loss is 20.16 , inference loss is 8.34, weight deacy loss is 11.82, training accuracy is 0.312500, time 203.703 samples/sec
+epoch 2, total_step 32920, total loss is 26.13 , inference loss is 14.31, weight deacy loss is 11.82, training accuracy is 0.218750, time 218.989 samples/sec
+epoch 2, total_step 32940, total loss is 22.16 , inference loss is 10.34, weight deacy loss is 11.82, training accuracy is 0.343750, time 226.428 samples/sec
+epoch 2, total_step 32960, total loss is 24.36 , inference loss is 12.55, weight deacy loss is 11.82, training accuracy is 0.250000, time 213.461 samples/sec
+epoch 2, total_step 32980, total loss is 25.24 , inference loss is 13.42, weight deacy loss is 11.81, training accuracy is 0.250000, time 223.014 samples/sec
+epoch 2, total_step 33000, total loss is 26.12 , inference loss is 14.31, weight deacy loss is 11.81, training accuracy is 0.250000, time 12.242 samples/sec
+epoch 2, total_step 33020, total loss is 25.47 , inference loss is 13.66, weight deacy loss is 11.81, training accuracy is 0.156250, time 208.796 samples/sec
+epoch 2, total_step 33040, total loss is 27.05 , inference loss is 15.24, weight deacy loss is 11.81, training accuracy is 0.281250, time 209.056 samples/sec
+epoch 2, total_step 33060, total loss is 26.66 , inference loss is 14.86, weight deacy loss is 11.81, training accuracy is 0.125000, time 295.376 samples/sec
+epoch 2, total_step 33080, total loss is 25.67 , inference loss is 13.87, weight deacy loss is 11.81, training accuracy is 0.156250, time 210.033 samples/sec
+epoch 2, total_step 33100, total loss is 21.67 , inference loss is 9.87, weight deacy loss is 11.80, training accuracy is 0.218750, time 211.432 samples/sec
+epoch 2, total_step 33120, total loss is 23.12 , inference loss is 11.32, weight deacy loss is 11.80, training accuracy is 0.281250, time 222.910 samples/sec
+epoch 2, total_step 33140, total loss is 25.99 , inference loss is 14.19, weight deacy loss is 11.80, training accuracy is 0.187500, time 225.534 samples/sec
+epoch 2, total_step 33160, total loss is 29.52 , inference loss is 17.72, weight deacy loss is 11.80, training accuracy is 0.093750, time 220.270 samples/sec
+epoch 2, total_step 33180, total loss is 25.39 , inference loss is 13.59, weight deacy loss is 11.80, training accuracy is 0.218750, time 225.369 samples/sec
+epoch 2, total_step 33200, total loss is 21.28 , inference loss is 9.48, weight deacy loss is 11.80, training accuracy is 0.312500, time 203.181 samples/sec
+epoch 2, total_step 33220, total loss is 28.04 , inference loss is 16.25, weight deacy loss is 11.79, training accuracy is 0.125000, time 215.329 samples/sec
+epoch 2, total_step 33240, total loss is 26.59 , inference loss is 14.80, weight deacy loss is 11.79, training accuracy is 0.218750, time 199.010 samples/sec
+epoch 2, total_step 33260, total loss is 21.02 , inference loss is 9.23, weight deacy loss is 11.79, training accuracy is 0.218750, time 226.488 samples/sec
+epoch 2, total_step 33280, total loss is 19.62 , inference loss is 7.83, weight deacy loss is 11.79, training accuracy is 0.281250, time 223.819 samples/sec
+epoch 2, total_step 33300, total loss is 21.72 , inference loss is 9.93, weight deacy loss is 11.79, training accuracy is 0.437500, time 13.425 samples/sec
+epoch 2, total_step 33320, total loss is 24.48 , inference loss is 12.70, weight deacy loss is 11.79, training accuracy is 0.187500, time 301.191 samples/sec
+epoch 2, total_step 33340, total loss is 22.85 , inference loss is 11.07, weight deacy loss is 11.78, training accuracy is 0.343750, time 199.040 samples/sec
+epoch 2, total_step 33360, total loss is 26.64 , inference loss is 14.86, weight deacy loss is 11.78, training accuracy is 0.187500, time 222.222 samples/sec
+epoch 2, total_step 33380, total loss is 28.60 , inference loss is 16.82, weight deacy loss is 11.78, training accuracy is 0.125000, time 227.633 samples/sec
+epoch 2, total_step 33400, total loss is 23.41 , inference loss is 11.63, weight deacy loss is 11.78, training accuracy is 0.281250, time 223.325 samples/sec
+epoch 2, total_step 33420, total loss is 24.28 , inference loss is 12.50, weight deacy loss is 11.78, training accuracy is 0.312500, time 202.323 samples/sec
+epoch 2, total_step 33440, total loss is 21.56 , inference loss is 9.78, weight deacy loss is 11.78, training accuracy is 0.281250, time 212.777 samples/sec
+epoch 2, total_step 33460, total loss is 26.96 , inference loss is 15.18, weight deacy loss is 11.77, training accuracy is 0.218750, time 235.011 samples/sec
+epoch 2, total_step 33480, total loss is 26.98 , inference loss is 15.21, weight deacy loss is 11.77, training accuracy is 0.125000, time 211.909 samples/sec
+epoch 2, total_step 33500, total loss is 24.49 , inference loss is 12.72, weight deacy loss is 11.77, training accuracy is 0.250000, time 199.834 samples/sec
+epoch 2, total_step 33520, total loss is 25.87 , inference loss is 14.10, weight deacy loss is 11.77, training accuracy is 0.250000, time 213.893 samples/sec
+epoch 2, total_step 33540, total loss is 23.21 , inference loss is 11.44, weight deacy loss is 11.77, training accuracy is 0.187500, time 205.390 samples/sec
+epoch 2, total_step 33560, total loss is 26.12 , inference loss is 14.36, weight deacy loss is 11.77, training accuracy is 0.031250, time 227.286 samples/sec
+epoch 2, total_step 33580, total loss is 21.46 , inference loss is 9.69, weight deacy loss is 11.77, training accuracy is 0.343750, time 216.028 samples/sec
+epoch 2, total_step 33600, total loss is 23.83 , inference loss is 12.07, weight deacy loss is 11.76, training accuracy is 0.250000, time 13.773 samples/sec
+epoch 2, total_step 33620, total loss is 28.35 , inference loss is 16.59, weight deacy loss is 11.76, training accuracy is 0.250000, time 208.230 samples/sec
+epoch 2, total_step 33640, total loss is 25.03 , inference loss is 13.27, weight deacy loss is 11.76, training accuracy is 0.281250, time 219.419 samples/sec
+epoch 2, total_step 33660, total loss is 30.59 , inference loss is 18.83, weight deacy loss is 11.76, training accuracy is 0.218750, time 214.776 samples/sec
+epoch 2, total_step 33680, total loss is 26.21 , inference loss is 14.45, weight deacy loss is 11.76, training accuracy is 0.218750, time 173.970 samples/sec
+epoch 2, total_step 33700, total loss is 25.26 , inference loss is 13.50, weight deacy loss is 11.76, training accuracy is 0.156250, time 212.502 samples/sec
+epoch 2, total_step 33720, total loss is 20.54 , inference loss is 8.78, weight deacy loss is 11.75, training accuracy is 0.312500, time 210.708 samples/sec
+epoch 2, total_step 33740, total loss is 24.04 , inference loss is 12.28, weight deacy loss is 11.75, training accuracy is 0.281250, time 214.799 samples/sec
+epoch 2, total_step 33760, total loss is 25.16 , inference loss is 13.41, weight deacy loss is 11.75, training accuracy is 0.250000, time 230.476 samples/sec
+epoch 2, total_step 33780, total loss is 25.90 , inference loss is 14.15, weight deacy loss is 11.75, training accuracy is 0.218750, time 217.313 samples/sec
+epoch 2, total_step 33800, total loss is 29.96 , inference loss is 18.22, weight deacy loss is 11.75, training accuracy is 0.218750, time 209.801 samples/sec
+epoch 2, total_step 33820, total loss is 26.18 , inference loss is 14.43, weight deacy loss is 11.75, training accuracy is 0.218750, time 212.738 samples/sec
+epoch 2, total_step 33840, total loss is 24.79 , inference loss is 13.04, weight deacy loss is 11.74, training accuracy is 0.343750, time 237.551 samples/sec
+epoch 2, total_step 33860, total loss is 21.03 , inference loss is 9.29, weight deacy loss is 11.74, training accuracy is 0.250000, time 215.959 samples/sec
+epoch 2, total_step 33880, total loss is 28.66 , inference loss is 16.92, weight deacy loss is 11.74, training accuracy is 0.156250, time 217.759 samples/sec
+epoch 2, total_step 33900, total loss is 28.93 , inference loss is 17.19, weight deacy loss is 11.74, training accuracy is 0.218750, time 14.266 samples/sec
+epoch 2, total_step 33920, total loss is 30.73 , inference loss is 19.00, weight deacy loss is 11.74, training accuracy is 0.062500, time 295.484 samples/sec
+epoch 2, total_step 33940, total loss is 27.49 , inference loss is 15.75, weight deacy loss is 11.74, training accuracy is 0.218750, time 204.673 samples/sec
+epoch 2, total_step 33960, total loss is 24.26 , inference loss is 12.52, weight deacy loss is 11.73, training accuracy is 0.250000, time 204.154 samples/sec
+epoch 2, total_step 33980, total loss is 26.29 , inference loss is 14.56, weight deacy loss is 11.73, training accuracy is 0.218750, time 219.567 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.268443000000005
+best_threshold_index 107 0.9788888888888889
+best_threshold_index 108 0.9785185185185186
+best_threshold_index 108 0.9798148148148148
+best_threshold_index 108 0.9792592592592593
+best_threshold_index 108 0.9792592592592593
+best_threshold_index 108 0.9781481481481481
+best_threshold_index 108 0.9788888888888889
+best_threshold_index 107 0.9787037037037037
+best_threshold_index 107 0.9774074074074074
+best_threshold_index 108 0.9777777777777777
+[lfw][34000]XNorm: 23.754778
+[lfw][34000]Accuracy-Flip: 0.97817+-0.00603
+test accuracy is: 0.9781666666666666
+epoch 2, total_step 34000, total loss is 27.89 , inference loss is 16.16, weight deacy loss is 11.73, training accuracy is 0.125000, time 16.778 samples/sec
+epoch 2, total_step 34020, total loss is 24.38 , inference loss is 12.65, weight deacy loss is 11.73, training accuracy is 0.250000, time 176.780 samples/sec
+epoch 2, total_step 34040, total loss is 25.75 , inference loss is 14.03, weight deacy loss is 11.73, training accuracy is 0.218750, time 211.069 samples/sec
+epoch 2, total_step 34060, total loss is 26.47 , inference loss is 14.74, weight deacy loss is 11.73, training accuracy is 0.156250, time 208.428 samples/sec
+epoch 2, total_step 34080, total loss is 26.77 , inference loss is 15.05, weight deacy loss is 11.72, training accuracy is 0.156250, time 219.589 samples/sec
+epoch 2, total_step 34100, total loss is 28.75 , inference loss is 17.03, weight deacy loss is 11.72, training accuracy is 0.125000, time 213.336 samples/sec
+epoch 2, total_step 34120, total loss is 27.58 , inference loss is 15.86, weight deacy loss is 11.72, training accuracy is 0.187500, time 224.400 samples/sec
+epoch 2, total_step 34140, total loss is 25.13 , inference loss is 13.41, weight deacy loss is 11.72, training accuracy is 0.281250, time 299.289 samples/sec
+epoch 2, total_step 34160, total loss is 28.40 , inference loss is 16.69, weight deacy loss is 11.72, training accuracy is 0.093750, time 213.874 samples/sec
+epoch 2, total_step 34180, total loss is 30.07 , inference loss is 18.35, weight deacy loss is 11.72, training accuracy is 0.093750, time 239.652 samples/sec
+epoch 2, total_step 34200, total loss is 27.31 , inference loss is 15.59, weight deacy loss is 11.72, training accuracy is 0.156250, time 14.154 samples/sec
+epoch 2, total_step 34220, total loss is 26.28 , inference loss is 14.56, weight deacy loss is 11.71, training accuracy is 0.156250, time 233.256 samples/sec
+epoch 2, total_step 34240, total loss is 27.75 , inference loss is 16.04, weight deacy loss is 11.71, training accuracy is 0.218750, time 208.202 samples/sec
+epoch 2, total_step 34260, total loss is 29.27 , inference loss is 17.56, weight deacy loss is 11.71, training accuracy is 0.218750, time 225.055 samples/sec
+epoch 2, total_step 34280, total loss is 22.36 , inference loss is 10.65, weight deacy loss is 11.71, training accuracy is 0.218750, time 211.430 samples/sec
+epoch 2, total_step 34300, total loss is 22.48 , inference loss is 10.77, weight deacy loss is 11.71, training accuracy is 0.218750, time 228.187 samples/sec
+epoch 2, total_step 34320, total loss is 28.00 , inference loss is 16.30, weight deacy loss is 11.71, training accuracy is 0.187500, time 224.418 samples/sec
+epoch 2, total_step 34340, total loss is 27.27 , inference loss is 15.56, weight deacy loss is 11.70, training accuracy is 0.187500, time 202.376 samples/sec
+epoch 2, total_step 34360, total loss is 27.68 , inference loss is 15.98, weight deacy loss is 11.70, training accuracy is 0.187500, time 203.533 samples/sec
+epoch 2, total_step 34380, total loss is 32.02 , inference loss is 20.32, weight deacy loss is 11.70, training accuracy is 0.062500, time 228.781 samples/sec
+epoch 2, total_step 34400, total loss is 25.65 , inference loss is 13.95, weight deacy loss is 11.70, training accuracy is 0.218750, time 298.164 samples/sec
+epoch 2, total_step 34420, total loss is 25.63 , inference loss is 13.93, weight deacy loss is 11.70, training accuracy is 0.250000, time 225.700 samples/sec
+epoch 2, total_step 34440, total loss is 26.29 , inference loss is 14.59, weight deacy loss is 11.70, training accuracy is 0.218750, time 202.584 samples/sec
+epoch 2, total_step 34460, total loss is 26.00 , inference loss is 14.31, weight deacy loss is 11.69, training accuracy is 0.156250, time 210.483 samples/sec
+epoch 2, total_step 34480, total loss is 27.31 , inference loss is 15.61, weight deacy loss is 11.69, training accuracy is 0.187500, time 206.046 samples/sec
+epoch 2, total_step 34500, total loss is 29.75 , inference loss is 18.05, weight deacy loss is 11.69, training accuracy is 0.125000, time 13.901 samples/sec
+epoch 2, total_step 34520, total loss is 25.05 , inference loss is 13.36, weight deacy loss is 11.69, training accuracy is 0.156250, time 215.979 samples/sec
+epoch 2, total_step 34540, total loss is 26.42 , inference loss is 14.73, weight deacy loss is 11.69, training accuracy is 0.187500, time 212.920 samples/sec
+epoch 2, total_step 34560, total loss is 27.17 , inference loss is 15.48, weight deacy loss is 11.69, training accuracy is 0.250000, time 211.387 samples/sec
+epoch 2, total_step 34580, total loss is 32.56 , inference loss is 20.88, weight deacy loss is 11.69, training accuracy is 0.062500, time 221.530 samples/sec
+epoch 2, total_step 34600, total loss is 33.40 , inference loss is 21.71, weight deacy loss is 11.68, training accuracy is 0.156250, time 206.644 samples/sec
+epoch 2, total_step 34620, total loss is 25.82 , inference loss is 14.14, weight deacy loss is 11.68, training accuracy is 0.093750, time 217.171 samples/sec
+epoch 2, total_step 34640, total loss is 28.52 , inference loss is 16.84, weight deacy loss is 11.68, training accuracy is 0.156250, time 209.772 samples/sec
+epoch 2, total_step 34660, total loss is 26.96 , inference loss is 15.28, weight deacy loss is 11.68, training accuracy is 0.125000, time 301.216 samples/sec
+epoch 2, total_step 34680, total loss is 29.84 , inference loss is 18.16, weight deacy loss is 11.68, training accuracy is 0.125000, time 224.371 samples/sec
+epoch 2, total_step 34700, total loss is 31.15 , inference loss is 19.48, weight deacy loss is 11.68, training accuracy is 0.125000, time 226.345 samples/sec
+epoch 2, total_step 34720, total loss is 26.50 , inference loss is 14.82, weight deacy loss is 11.67, training accuracy is 0.187500, time 223.977 samples/sec
+epoch 2, total_step 34740, total loss is 26.04 , inference loss is 14.37, weight deacy loss is 11.67, training accuracy is 0.187500, time 220.694 samples/sec
+epoch 2, total_step 34760, total loss is 30.93 , inference loss is 19.26, weight deacy loss is 11.67, training accuracy is 0.031250, time 215.762 samples/sec
+epoch 2, total_step 34780, total loss is 28.75 , inference loss is 17.08, weight deacy loss is 11.67, training accuracy is 0.125000, time 221.412 samples/sec
+epoch 2, total_step 34800, total loss is 31.28 , inference loss is 19.61, weight deacy loss is 11.67, training accuracy is 0.125000, time 12.321 samples/sec
+epoch 2, total_step 34820, total loss is 30.71 , inference loss is 19.04, weight deacy loss is 11.67, training accuracy is 0.125000, time 228.519 samples/sec
+epoch 2, total_step 34840, total loss is 28.09 , inference loss is 16.42, weight deacy loss is 11.67, training accuracy is 0.125000, time 205.626 samples/sec
+epoch 2, total_step 34860, total loss is 27.10 , inference loss is 15.43, weight deacy loss is 11.66, training accuracy is 0.250000, time 228.581 samples/sec
+epoch 2, total_step 34880, total loss is 26.48 , inference loss is 14.82, weight deacy loss is 11.66, training accuracy is 0.156250, time 234.283 samples/sec
+epoch 2, total_step 34900, total loss is 23.32 , inference loss is 11.66, weight deacy loss is 11.66, training accuracy is 0.343750, time 220.132 samples/sec
+epoch 2, total_step 34920, total loss is 22.95 , inference loss is 11.29, weight deacy loss is 11.66, training accuracy is 0.250000, time 212.021 samples/sec
+epoch 2, total_step 34940, total loss is 29.30 , inference loss is 17.65, weight deacy loss is 11.66, training accuracy is 0.093750, time 191.132 samples/sec
+epoch 2, total_step 34960, total loss is 29.76 , inference loss is 18.10, weight deacy loss is 11.66, training accuracy is 0.062500, time 227.354 samples/sec
+epoch 2, total_step 34980, total loss is 27.84 , inference loss is 16.18, weight deacy loss is 11.65, training accuracy is 0.031250, time 214.260 samples/sec
+epoch 2, total_step 35000, total loss is 24.75 , inference loss is 13.10, weight deacy loss is 11.65, training accuracy is 0.062500, time 216.618 samples/sec
+epoch 2, total_step 35020, total loss is 29.20 , inference loss is 17.55, weight deacy loss is 11.65, training accuracy is 0.062500, time 220.931 samples/sec
+epoch 2, total_step 35040, total loss is 22.55 , inference loss is 10.90, weight deacy loss is 11.65, training accuracy is 0.156250, time 207.137 samples/sec
+epoch 2, total_step 35060, total loss is 32.91 , inference loss is 21.26, weight deacy loss is 11.65, training accuracy is 0.062500, time 226.057 samples/sec
+epoch 2, total_step 35080, total loss is 25.87 , inference loss is 14.22, weight deacy loss is 11.65, training accuracy is 0.187500, time 204.746 samples/sec
+epoch 2, total_step 35100, total loss is 29.52 , inference loss is 17.88, weight deacy loss is 11.65, training accuracy is 0.125000, time 13.444 samples/sec
+epoch 2, total_step 35120, total loss is 27.30 , inference loss is 15.65, weight deacy loss is 11.64, training accuracy is 0.125000, time 210.854 samples/sec
+epoch 2, total_step 35140, total loss is 27.14 , inference loss is 15.50, weight deacy loss is 11.64, training accuracy is 0.125000, time 224.657 samples/sec
+epoch 2, total_step 35160, total loss is 27.43 , inference loss is 15.79, weight deacy loss is 11.64, training accuracy is 0.156250, time 244.468 samples/sec
+epoch 2, total_step 35180, total loss is 25.01 , inference loss is 13.37, weight deacy loss is 11.64, training accuracy is 0.250000, time 217.091 samples/sec
+epoch 2, total_step 35200, total loss is 24.08 , inference loss is 12.44, weight deacy loss is 11.64, training accuracy is 0.125000, time 298.915 samples/sec
+epoch 2, total_step 35220, total loss is 24.81 , inference loss is 13.18, weight deacy loss is 11.64, training accuracy is 0.218750, time 212.972 samples/sec
+epoch 2, total_step 35240, total loss is 29.31 , inference loss is 17.68, weight deacy loss is 11.63, training accuracy is 0.093750, time 215.310 samples/sec
+epoch 2, total_step 35260, total loss is 25.46 , inference loss is 13.83, weight deacy loss is 11.63, training accuracy is 0.156250, time 240.144 samples/sec
+epoch 2, total_step 35280, total loss is 26.84 , inference loss is 15.21, weight deacy loss is 11.63, training accuracy is 0.125000, time 218.373 samples/sec
+epoch 2, total_step 35300, total loss is 30.04 , inference loss is 18.41, weight deacy loss is 11.63, training accuracy is 0.125000, time 215.018 samples/sec
+epoch 2, total_step 35320, total loss is 24.24 , inference loss is 12.61, weight deacy loss is 11.63, training accuracy is 0.093750, time 221.791 samples/sec
+epoch 2, total_step 35340, total loss is 25.61 , inference loss is 13.98, weight deacy loss is 11.63, training accuracy is 0.187500, time 224.892 samples/sec
+epoch 2, total_step 35360, total loss is 26.70 , inference loss is 15.08, weight deacy loss is 11.62, training accuracy is 0.125000, time 221.161 samples/sec
+epoch 2, total_step 35380, total loss is 24.68 , inference loss is 13.06, weight deacy loss is 11.62, training accuracy is 0.218750, time 216.680 samples/sec
+epoch 2, total_step 35400, total loss is 24.90 , inference loss is 13.28, weight deacy loss is 11.62, training accuracy is 0.062500, time 14.359 samples/sec
+epoch 2, total_step 35420, total loss is 20.48 , inference loss is 8.86, weight deacy loss is 11.62, training accuracy is 0.312500, time 226.482 samples/sec
+epoch 2, total_step 35440, total loss is 28.59 , inference loss is 16.98, weight deacy loss is 11.62, training accuracy is 0.125000, time 217.462 samples/sec
+epoch 2, total_step 35460, total loss is 31.84 , inference loss is 20.23, weight deacy loss is 11.62, training accuracy is 0.062500, time 298.054 samples/sec
+epoch 2, total_step 35480, total loss is 28.32 , inference loss is 16.70, weight deacy loss is 11.62, training accuracy is 0.156250, time 224.416 samples/sec
+epoch 2, total_step 35500, total loss is 24.05 , inference loss is 12.44, weight deacy loss is 11.61, training accuracy is 0.218750, time 210.508 samples/sec
+epoch 2, total_step 35520, total loss is 32.37 , inference loss is 20.75, weight deacy loss is 11.61, training accuracy is 0.031250, time 225.311 samples/sec
+epoch 2, total_step 35540, total loss is 33.50 , inference loss is 21.89, weight deacy loss is 11.61, training accuracy is 0.125000, time 227.882 samples/sec
+epoch 2, total_step 35560, total loss is 29.84 , inference loss is 18.23, weight deacy loss is 11.61, training accuracy is 0.062500, time 202.389 samples/sec
+epoch 2, total_step 35580, total loss is 27.86 , inference loss is 16.25, weight deacy loss is 11.61, training accuracy is 0.062500, time 212.594 samples/sec
+epoch 2, total_step 35600, total loss is 26.73 , inference loss is 15.12, weight deacy loss is 11.61, training accuracy is 0.093750, time 214.151 samples/sec
+epoch 2, total_step 35620, total loss is 29.33 , inference loss is 17.72, weight deacy loss is 11.60, training accuracy is 0.250000, time 212.652 samples/sec
+epoch 2, total_step 35640, total loss is 26.67 , inference loss is 15.06, weight deacy loss is 11.60, training accuracy is 0.156250, time 222.144 samples/sec
+epoch 2, total_step 35660, total loss is 27.64 , inference loss is 16.04, weight deacy loss is 11.60, training accuracy is 0.156250, time 215.838 samples/sec
+epoch 2, total_step 35680, total loss is 26.09 , inference loss is 14.49, weight deacy loss is 11.60, training accuracy is 0.125000, time 215.880 samples/sec
+epoch 2, total_step 35700, total loss is 25.12 , inference loss is 13.52, weight deacy loss is 11.60, training accuracy is 0.218750, time 14.954 samples/sec
+epoch 2, total_step 35720, total loss is 28.10 , inference loss is 16.51, weight deacy loss is 11.60, training accuracy is 0.156250, time 209.246 samples/sec
+epoch 2, total_step 35740, total loss is 25.08 , inference loss is 13.49, weight deacy loss is 11.60, training accuracy is 0.218750, time 208.333 samples/sec
+epoch 2, total_step 35760, total loss is 28.71 , inference loss is 17.12, weight deacy loss is 11.59, training accuracy is 0.062500, time 218.194 samples/sec
+epoch 2, total_step 35780, total loss is 24.79 , inference loss is 13.20, weight deacy loss is 11.59, training accuracy is 0.187500, time 226.426 samples/sec
+epoch 2, total_step 35800, total loss is 21.65 , inference loss is 10.06, weight deacy loss is 11.59, training accuracy is 0.250000, time 217.417 samples/sec
+epoch 2, total_step 35820, total loss is 28.50 , inference loss is 16.91, weight deacy loss is 11.59, training accuracy is 0.031250, time 219.351 samples/sec
+epoch 2, total_step 35840, total loss is 23.90 , inference loss is 12.31, weight deacy loss is 11.59, training accuracy is 0.187500, time 216.594 samples/sec
+epoch 2, total_step 35860, total loss is 23.17 , inference loss is 11.58, weight deacy loss is 11.59, training accuracy is 0.187500, time 211.101 samples/sec
+epoch 2, total_step 35880, total loss is 26.92 , inference loss is 15.34, weight deacy loss is 11.58, training accuracy is 0.125000, time 210.713 samples/sec
+epoch 2, total_step 35900, total loss is 23.05 , inference loss is 11.47, weight deacy loss is 11.58, training accuracy is 0.218750, time 227.129 samples/sec
+epoch 2, total_step 35920, total loss is 20.96 , inference loss is 9.38, weight deacy loss is 11.58, training accuracy is 0.375000, time 248.166 samples/sec
+epoch 2, total_step 35940, total loss is 23.60 , inference loss is 12.02, weight deacy loss is 11.58, training accuracy is 0.218750, time 224.761 samples/sec
+epoch 2, total_step 35960, total loss is 25.90 , inference loss is 14.32, weight deacy loss is 11.58, training accuracy is 0.250000, time 208.056 samples/sec
+epoch 2, total_step 35980, total loss is 31.71 , inference loss is 20.14, weight deacy loss is 11.58, training accuracy is 0.062500, time 214.415 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.260622999999985
+best_threshold_index 113 0.9812962962962963
+best_threshold_index 113 0.98
+best_threshold_index 113 0.9811111111111112
+best_threshold_index 113 0.9818518518518519
+best_threshold_index 113 0.9812962962962963
+best_threshold_index 113 0.98
+best_threshold_index 113 0.9794444444444445
+best_threshold_index 113 0.9803703703703703
+best_threshold_index 113 0.9783333333333334
+best_threshold_index 113 0.9796296296296296
+[lfw][36000]XNorm: 22.761330
+[lfw][36000]Accuracy-Flip: 0.98033+-0.00912
+test accuracy is: 0.9803333333333333
+epoch 2, total_step 36000, total loss is 25.98 , inference loss is 14.41, weight deacy loss is 11.58, training accuracy is 0.250000, time 17.212 samples/sec
+epoch 2, total_step 36020, total loss is 26.14 , inference loss is 14.56, weight deacy loss is 11.57, training accuracy is 0.281250, time 199.619 samples/sec
+epoch 2, total_step 36040, total loss is 27.32 , inference loss is 15.75, weight deacy loss is 11.57, training accuracy is 0.218750, time 211.837 samples/sec
+epoch 2, total_step 36060, total loss is 25.23 , inference loss is 13.66, weight deacy loss is 11.57, training accuracy is 0.281250, time 225.339 samples/sec
+epoch 2, total_step 36080, total loss is 27.67 , inference loss is 16.10, weight deacy loss is 11.57, training accuracy is 0.062500, time 230.755 samples/sec
+epoch 2, total_step 36100, total loss is 23.27 , inference loss is 11.70, weight deacy loss is 11.57, training accuracy is 0.250000, time 216.528 samples/sec
+epoch 2, total_step 36120, total loss is 28.21 , inference loss is 16.64, weight deacy loss is 11.57, training accuracy is 0.187500, time 196.377 samples/sec
+epoch 2, total_step 36140, total loss is 27.78 , inference loss is 16.22, weight deacy loss is 11.56, training accuracy is 0.281250, time 221.738 samples/sec
+epoch 2, total_step 36160, total loss is 29.57 , inference loss is 18.00, weight deacy loss is 11.56, training accuracy is 0.125000, time 226.874 samples/sec
+epoch 2, total_step 36180, total loss is 23.94 , inference loss is 12.38, weight deacy loss is 11.56, training accuracy is 0.250000, time 202.049 samples/sec
+epoch 2, total_step 36200, total loss is 25.53 , inference loss is 13.97, weight deacy loss is 11.56, training accuracy is 0.093750, time 218.473 samples/sec
+epoch 2, total_step 36220, total loss is 24.87 , inference loss is 13.31, weight deacy loss is 11.56, training accuracy is 0.218750, time 221.794 samples/sec
+epoch 2, total_step 36240, total loss is 26.20 , inference loss is 14.64, weight deacy loss is 11.56, training accuracy is 0.187500, time 218.814 samples/sec
+epoch 2, total_step 36260, total loss is 27.26 , inference loss is 15.71, weight deacy loss is 11.56, training accuracy is 0.187500, time 224.398 samples/sec
+epoch 2, total_step 36280, total loss is 24.89 , inference loss is 13.34, weight deacy loss is 11.55, training accuracy is 0.156250, time 231.995 samples/sec
+epoch 2, total_step 36300, total loss is 28.56 , inference loss is 17.01, weight deacy loss is 11.55, training accuracy is 0.062500, time 12.508 samples/sec
+epoch 2, total_step 36320, total loss is 26.52 , inference loss is 14.97, weight deacy loss is 11.55, training accuracy is 0.250000, time 224.696 samples/sec
+epoch 2, total_step 36340, total loss is 28.32 , inference loss is 16.77, weight deacy loss is 11.55, training accuracy is 0.218750, time 221.309 samples/sec
+epoch 2, total_step 36360, total loss is 26.54 , inference loss is 15.00, weight deacy loss is 11.55, training accuracy is 0.125000, time 208.207 samples/sec
+epoch 2, total_step 36380, total loss is 22.98 , inference loss is 11.43, weight deacy loss is 11.55, training accuracy is 0.312500, time 224.844 samples/sec
+epoch 2, total_step 36400, total loss is 26.96 , inference loss is 15.41, weight deacy loss is 11.54, training accuracy is 0.250000, time 223.407 samples/sec
+epoch 2, total_step 36420, total loss is 28.63 , inference loss is 17.09, weight deacy loss is 11.54, training accuracy is 0.187500, time 228.083 samples/sec
+epoch 2, total_step 36440, total loss is 26.10 , inference loss is 14.55, weight deacy loss is 11.54, training accuracy is 0.156250, time 223.734 samples/sec
+epoch 2, total_step 36460, total loss is 21.78 , inference loss is 10.24, weight deacy loss is 11.54, training accuracy is 0.312500, time 216.430 samples/sec
+epoch 2, total_step 36480, total loss is 25.46 , inference loss is 13.92, weight deacy loss is 11.54, training accuracy is 0.187500, time 237.930 samples/sec
+epoch 2, total_step 36500, total loss is 24.97 , inference loss is 13.43, weight deacy loss is 11.54, training accuracy is 0.218750, time 225.882 samples/sec
+epoch 2, total_step 36520, total loss is 28.82 , inference loss is 17.29, weight deacy loss is 11.53, training accuracy is 0.125000, time 222.640 samples/sec
+epoch 2, total_step 36540, total loss is 28.38 , inference loss is 16.85, weight deacy loss is 11.53, training accuracy is 0.250000, time 219.337 samples/sec
+epoch 2, total_step 36560, total loss is 26.69 , inference loss is 15.16, weight deacy loss is 11.53, training accuracy is 0.156250, time 295.296 samples/sec
+epoch 2, total_step 36580, total loss is 26.63 , inference loss is 15.10, weight deacy loss is 11.53, training accuracy is 0.156250, time 210.135 samples/sec
+epoch 2, total_step 36600, total loss is 31.01 , inference loss is 19.48, weight deacy loss is 11.53, training accuracy is 0.156250, time 12.878 samples/sec
+epoch 2, total_step 36620, total loss is 28.63 , inference loss is 17.10, weight deacy loss is 11.53, training accuracy is 0.218750, time 208.237 samples/sec
+epoch 2, total_step 36640, total loss is 28.84 , inference loss is 17.32, weight deacy loss is 11.53, training accuracy is 0.187500, time 215.610 samples/sec
+epoch 2, total_step 36660, total loss is 25.68 , inference loss is 14.16, weight deacy loss is 11.52, training accuracy is 0.125000, time 207.420 samples/sec
+epoch 2, total_step 36680, total loss is 22.46 , inference loss is 10.93, weight deacy loss is 11.52, training accuracy is 0.250000, time 216.128 samples/sec
+epoch 2, total_step 36700, total loss is 31.94 , inference loss is 20.42, weight deacy loss is 11.52, training accuracy is 0.125000, time 235.583 samples/sec
+epoch 2, total_step 36720, total loss is 29.21 , inference loss is 17.69, weight deacy loss is 11.52, training accuracy is 0.187500, time 195.467 samples/sec
+epoch 2, total_step 36740, total loss is 28.23 , inference loss is 16.71, weight deacy loss is 11.52, training accuracy is 0.156250, time 190.234 samples/sec
+epoch 2, total_step 36760, total loss is 28.55 , inference loss is 17.04, weight deacy loss is 11.52, training accuracy is 0.093750, time 209.835 samples/sec
+epoch 2, total_step 36780, total loss is 28.80 , inference loss is 17.28, weight deacy loss is 11.51, training accuracy is 0.156250, time 228.379 samples/sec
+epoch 2, total_step 36800, total loss is 26.49 , inference loss is 14.98, weight deacy loss is 11.51, training accuracy is 0.062500, time 206.638 samples/sec
+epoch 2, total_step 36820, total loss is 25.06 , inference loss is 13.55, weight deacy loss is 11.51, training accuracy is 0.093750, time 299.617 samples/sec
+epoch 2, total_step 36840, total loss is 24.14 , inference loss is 12.63, weight deacy loss is 11.51, training accuracy is 0.125000, time 210.552 samples/sec
+epoch 2, total_step 36860, total loss is 25.53 , inference loss is 14.02, weight deacy loss is 11.51, training accuracy is 0.093750, time 205.062 samples/sec
+epoch 2, total_step 36880, total loss is 28.13 , inference loss is 16.62, weight deacy loss is 11.51, training accuracy is 0.156250, time 230.529 samples/sec
+epoch 2, total_step 36900, total loss is 29.53 , inference loss is 18.02, weight deacy loss is 11.51, training accuracy is 0.125000, time 13.883 samples/sec
+epoch 2, total_step 36920, total loss is 25.85 , inference loss is 14.34, weight deacy loss is 11.50, training accuracy is 0.250000, time 226.071 samples/sec
+epoch 2, total_step 36940, total loss is 25.27 , inference loss is 13.76, weight deacy loss is 11.50, training accuracy is 0.250000, time 205.948 samples/sec
+epoch 2, total_step 36960, total loss is 23.35 , inference loss is 11.85, weight deacy loss is 11.50, training accuracy is 0.250000, time 222.649 samples/sec
+epoch 2, total_step 36980, total loss is 20.75 , inference loss is 9.25, weight deacy loss is 11.50, training accuracy is 0.406250, time 202.440 samples/sec
+epoch 2, total_step 37000, total loss is 24.29 , inference loss is 12.79, weight deacy loss is 11.50, training accuracy is 0.156250, time 226.483 samples/sec
+epoch 2, total_step 37020, total loss is 32.17 , inference loss is 20.67, weight deacy loss is 11.50, training accuracy is 0.187500, time 216.597 samples/sec
+epoch 2, total_step 37040, total loss is 29.64 , inference loss is 18.15, weight deacy loss is 11.49, training accuracy is 0.218750, time 219.240 samples/sec
+epoch 2, total_step 37060, total loss is 27.23 , inference loss is 15.74, weight deacy loss is 11.49, training accuracy is 0.187500, time 206.757 samples/sec
+epoch 2, total_step 37080, total loss is 25.61 , inference loss is 14.12, weight deacy loss is 11.49, training accuracy is 0.156250, time 184.654 samples/sec
+epoch 2, total_step 37100, total loss is 32.71 , inference loss is 21.22, weight deacy loss is 11.49, training accuracy is 0.250000, time 197.213 samples/sec
+epoch 2, total_step 37120, total loss is 26.96 , inference loss is 15.47, weight deacy loss is 11.49, training accuracy is 0.156250, time 216.666 samples/sec
+epoch 2, total_step 37140, total loss is 30.75 , inference loss is 19.27, weight deacy loss is 11.49, training accuracy is 0.125000, time 222.556 samples/sec
+epoch 2, total_step 37160, total loss is 26.63 , inference loss is 15.14, weight deacy loss is 11.49, training accuracy is 0.156250, time 221.883 samples/sec
+epoch 2, total_step 37180, total loss is 31.80 , inference loss is 20.31, weight deacy loss is 11.48, training accuracy is 0.156250, time 211.381 samples/sec
+epoch 2, total_step 37200, total loss is 26.22 , inference loss is 14.74, weight deacy loss is 11.48, training accuracy is 0.218750, time 10.468 samples/sec
+epoch 2, total_step 37220, total loss is 28.44 , inference loss is 16.96, weight deacy loss is 11.48, training accuracy is 0.093750, time 206.569 samples/sec
+epoch 2, total_step 37240, total loss is 30.57 , inference loss is 19.09, weight deacy loss is 11.48, training accuracy is 0.156250, time 220.446 samples/sec
+epoch 2, total_step 37260, total loss is 28.23 , inference loss is 16.76, weight deacy loss is 11.48, training accuracy is 0.062500, time 204.495 samples/sec
+epoch 2, total_step 37280, total loss is 26.02 , inference loss is 14.54, weight deacy loss is 11.48, training accuracy is 0.187500, time 236.877 samples/sec
+epoch 2, total_step 37300, total loss is 25.18 , inference loss is 13.70, weight deacy loss is 11.48, training accuracy is 0.156250, time 218.013 samples/sec
+epoch 2, total_step 37320, total loss is 28.08 , inference loss is 16.60, weight deacy loss is 11.47, training accuracy is 0.156250, time 207.111 samples/sec
+epoch 2, total_step 37340, total loss is 27.46 , inference loss is 15.99, weight deacy loss is 11.47, training accuracy is 0.187500, time 296.957 samples/sec
+epoch 2, total_step 37360, total loss is 28.66 , inference loss is 17.19, weight deacy loss is 11.47, training accuracy is 0.125000, time 214.502 samples/sec
+epoch 2, total_step 37380, total loss is 24.10 , inference loss is 12.63, weight deacy loss is 11.47, training accuracy is 0.187500, time 221.342 samples/sec
+epoch 2, total_step 37400, total loss is 27.18 , inference loss is 15.71, weight deacy loss is 11.47, training accuracy is 0.156250, time 213.907 samples/sec
+epoch 2, total_step 37420, total loss is 25.99 , inference loss is 14.52, weight deacy loss is 11.47, training accuracy is 0.187500, time 213.433 samples/sec
+epoch 2, total_step 37440, total loss is 31.44 , inference loss is 19.98, weight deacy loss is 11.46, training accuracy is 0.093750, time 216.334 samples/sec
+epoch 2, total_step 37460, total loss is 23.92 , inference loss is 12.46, weight deacy loss is 11.46, training accuracy is 0.187500, time 211.756 samples/sec
+epoch 2, total_step 37480, total loss is 24.39 , inference loss is 12.93, weight deacy loss is 11.46, training accuracy is 0.218750, time 211.705 samples/sec
+epoch 2, total_step 37500, total loss is 25.42 , inference loss is 13.96, weight deacy loss is 11.46, training accuracy is 0.093750, time 15.040 samples/sec
+epoch 2, total_step 37520, total loss is 29.73 , inference loss is 18.27, weight deacy loss is 11.46, training accuracy is 0.062500, time 225.753 samples/sec
+epoch 2, total_step 37540, total loss is 25.77 , inference loss is 14.31, weight deacy loss is 11.46, training accuracy is 0.093750, time 217.802 samples/sec
+epoch 2, total_step 37560, total loss is 28.31 , inference loss is 16.86, weight deacy loss is 11.46, training accuracy is 0.187500, time 216.621 samples/sec
+epoch 2, total_step 37580, total loss is 27.70 , inference loss is 16.24, weight deacy loss is 11.45, training accuracy is 0.093750, time 210.931 samples/sec
+epoch 2, total_step 37600, total loss is 26.14 , inference loss is 14.69, weight deacy loss is 11.45, training accuracy is 0.093750, time 232.355 samples/sec
+epoch 2, total_step 37620, total loss is 29.36 , inference loss is 17.91, weight deacy loss is 11.45, training accuracy is 0.093750, time 238.332 samples/sec
+epoch 2, total_step 37640, total loss is 27.25 , inference loss is 15.80, weight deacy loss is 11.45, training accuracy is 0.156250, time 227.688 samples/sec
+epoch 2, total_step 37660, total loss is 24.68 , inference loss is 13.23, weight deacy loss is 11.45, training accuracy is 0.125000, time 227.786 samples/sec
+epoch 2, total_step 37680, total loss is 28.38 , inference loss is 16.93, weight deacy loss is 11.45, training accuracy is 0.125000, time 203.931 samples/sec
+epoch 2, total_step 37700, total loss is 24.41 , inference loss is 12.96, weight deacy loss is 11.45, training accuracy is 0.125000, time 214.772 samples/sec
+epoch 2, total_step 37720, total loss is 28.01 , inference loss is 16.57, weight deacy loss is 11.44, training accuracy is 0.125000, time 215.383 samples/sec
+epoch 2, total_step 37740, total loss is 25.42 , inference loss is 13.98, weight deacy loss is 11.44, training accuracy is 0.156250, time 223.417 samples/sec
+epoch 2, total_step 37760, total loss is 25.44 , inference loss is 14.00, weight deacy loss is 11.44, training accuracy is 0.093750, time 213.003 samples/sec
+epoch 2, total_step 37780, total loss is 27.80 , inference loss is 16.36, weight deacy loss is 11.44, training accuracy is 0.156250, time 219.241 samples/sec
+epoch 2, total_step 37800, total loss is 24.31 , inference loss is 12.87, weight deacy loss is 11.44, training accuracy is 0.093750, time 13.091 samples/sec
+epoch 2, total_step 37820, total loss is 23.86 , inference loss is 12.42, weight deacy loss is 11.44, training accuracy is 0.187500, time 216.251 samples/sec
+epoch 2, total_step 37840, total loss is 23.97 , inference loss is 12.54, weight deacy loss is 11.43, training accuracy is 0.218750, time 216.385 samples/sec
+epoch 2, total_step 37860, total loss is 28.16 , inference loss is 16.73, weight deacy loss is 11.43, training accuracy is 0.156250, time 216.314 samples/sec
+epoch 2, total_step 37880, total loss is 26.94 , inference loss is 15.51, weight deacy loss is 11.43, training accuracy is 0.031250, time 214.031 samples/sec
+epoch 2, total_step 37900, total loss is 26.22 , inference loss is 14.79, weight deacy loss is 11.43, training accuracy is 0.125000, time 208.901 samples/sec
+epoch 2, total_step 37920, total loss is 26.26 , inference loss is 14.84, weight deacy loss is 11.43, training accuracy is 0.156250, time 204.816 samples/sec
+epoch 2, total_step 37940, total loss is 28.01 , inference loss is 16.58, weight deacy loss is 11.43, training accuracy is 0.125000, time 212.521 samples/sec
+epoch 2, total_step 37960, total loss is 25.54 , inference loss is 14.12, weight deacy loss is 11.43, training accuracy is 0.125000, time 216.285 samples/sec
+epoch 2, total_step 37980, total loss is 25.67 , inference loss is 14.24, weight deacy loss is 11.42, training accuracy is 0.156250, time 190.927 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.93752799999998
+best_threshold_index 110 0.982037037037037
+best_threshold_index 110 0.9814814814814815
+best_threshold_index 110 0.9825925925925926
+best_threshold_index 110 0.9822222222222222
+best_threshold_index 110 0.9818518518518519
+best_threshold_index 114 0.9805555555555555
+best_threshold_index 110 0.9812962962962963
+best_threshold_index 110 0.9801851851851852
+best_threshold_index 110 0.9794444444444445
+best_threshold_index 110 0.9801851851851852
+[lfw][38000]XNorm: 23.016690
+[lfw][38000]Accuracy-Flip: 0.98050+-0.00869
+test accuracy is: 0.9804999999999999
+epoch 2, total_step 38000, total loss is 24.47 , inference loss is 13.05, weight deacy loss is 11.42, training accuracy is 0.156250, time 17.136 samples/sec
+epoch 2, total_step 38020, total loss is 24.48 , inference loss is 13.06, weight deacy loss is 11.42, training accuracy is 0.375000, time 223.320 samples/sec
+epoch 2, total_step 38040, total loss is 25.43 , inference loss is 14.01, weight deacy loss is 11.42, training accuracy is 0.281250, time 218.780 samples/sec
+epoch 2, total_step 38060, total loss is 21.79 , inference loss is 10.37, weight deacy loss is 11.42, training accuracy is 0.375000, time 216.071 samples/sec
+epoch 2, total_step 38080, total loss is 26.94 , inference loss is 15.52, weight deacy loss is 11.42, training accuracy is 0.093750, time 240.367 samples/sec
+epoch 2, total_step 38100, total loss is 27.74 , inference loss is 16.33, weight deacy loss is 11.41, training accuracy is 0.156250, time 14.230 samples/sec
+epoch 2, total_step 38120, total loss is 26.74 , inference loss is 15.32, weight deacy loss is 11.41, training accuracy is 0.156250, time 221.452 samples/sec
+epoch 2, total_step 38140, total loss is 26.10 , inference loss is 14.69, weight deacy loss is 11.41, training accuracy is 0.250000, time 216.688 samples/sec
+epoch 2, total_step 38160, total loss is 28.58 , inference loss is 17.17, weight deacy loss is 11.41, training accuracy is 0.125000, time 226.633 samples/sec
+epoch 2, total_step 38180, total loss is 34.03 , inference loss is 22.62, weight deacy loss is 11.41, training accuracy is 0.125000, time 207.112 samples/sec
+epoch 2, total_step 38200, total loss is 26.42 , inference loss is 15.02, weight deacy loss is 11.41, training accuracy is 0.281250, time 207.203 samples/sec
+epoch 2, total_step 38220, total loss is 25.08 , inference loss is 13.67, weight deacy loss is 11.41, training accuracy is 0.187500, time 219.013 samples/sec
+epoch 2, total_step 38240, total loss is 22.07 , inference loss is 10.67, weight deacy loss is 11.40, training accuracy is 0.343750, time 223.375 samples/sec
+epoch 2, total_step 38260, total loss is 30.93 , inference loss is 19.52, weight deacy loss is 11.40, training accuracy is 0.218750, time 232.626 samples/sec
+epoch 2, total_step 38280, total loss is 26.07 , inference loss is 14.67, weight deacy loss is 11.40, training accuracy is 0.156250, time 215.462 samples/sec
+epoch 2, total_step 38300, total loss is 25.62 , inference loss is 14.22, weight deacy loss is 11.40, training accuracy is 0.218750, time 213.611 samples/sec
+epoch 2, total_step 38320, total loss is 23.20 , inference loss is 11.80, weight deacy loss is 11.40, training accuracy is 0.343750, time 208.785 samples/sec
+epoch 2, total_step 38340, total loss is 24.57 , inference loss is 13.17, weight deacy loss is 11.40, training accuracy is 0.156250, time 203.039 samples/sec
+epoch 2, total_step 38360, total loss is 29.38 , inference loss is 17.98, weight deacy loss is 11.40, training accuracy is 0.187500, time 227.999 samples/sec
+epoch 2, total_step 38380, total loss is 25.71 , inference loss is 14.32, weight deacy loss is 11.39, training accuracy is 0.281250, time 215.621 samples/sec
+epoch 2, total_step 38400, total loss is 30.49 , inference loss is 19.10, weight deacy loss is 11.39, training accuracy is 0.125000, time 12.259 samples/sec
+epoch 2, total_step 38420, total loss is 27.85 , inference loss is 16.46, weight deacy loss is 11.39, training accuracy is 0.187500, time 243.127 samples/sec
+epoch 2, total_step 38440, total loss is 27.88 , inference loss is 16.49, weight deacy loss is 11.39, training accuracy is 0.031250, time 211.336 samples/sec
+epoch 2, total_step 38460, total loss is 23.54 , inference loss is 12.15, weight deacy loss is 11.39, training accuracy is 0.250000, time 235.566 samples/sec
+epoch 2, total_step 38480, total loss is 28.81 , inference loss is 17.43, weight deacy loss is 11.39, training accuracy is 0.062500, time 197.801 samples/sec
+epoch 2, total_step 38500, total loss is 23.21 , inference loss is 11.83, weight deacy loss is 11.38, training accuracy is 0.343750, time 211.467 samples/sec
+epoch 2, total_step 38520, total loss is 23.62 , inference loss is 12.24, weight deacy loss is 11.38, training accuracy is 0.250000, time 213.478 samples/sec
+epoch 2, total_step 38540, total loss is 23.44 , inference loss is 12.06, weight deacy loss is 11.38, training accuracy is 0.218750, time 217.429 samples/sec
+epoch 2, total_step 38560, total loss is 24.50 , inference loss is 13.12, weight deacy loss is 11.38, training accuracy is 0.125000, time 198.758 samples/sec
+epoch 2, total_step 38580, total loss is 28.94 , inference loss is 17.56, weight deacy loss is 11.38, training accuracy is 0.156250, time 220.445 samples/sec
+epoch 2, total_step 38600, total loss is 28.56 , inference loss is 17.18, weight deacy loss is 11.38, training accuracy is 0.031250, time 214.959 samples/sec
+epoch 2, total_step 38620, total loss is 24.88 , inference loss is 13.51, weight deacy loss is 11.38, training accuracy is 0.187500, time 226.844 samples/sec
+epoch 2, total_step 38640, total loss is 27.41 , inference loss is 16.04, weight deacy loss is 11.37, training accuracy is 0.125000, time 226.107 samples/sec
+epoch 2, total_step 38660, total loss is 27.42 , inference loss is 16.05, weight deacy loss is 11.37, training accuracy is 0.218750, time 212.009 samples/sec
+epoch 2, total_step 38680, total loss is 26.63 , inference loss is 15.26, weight deacy loss is 11.37, training accuracy is 0.093750, time 299.939 samples/sec
+epoch 2, total_step 38700, total loss is 27.10 , inference loss is 15.73, weight deacy loss is 11.37, training accuracy is 0.218750, time 13.618 samples/sec
+epoch 2, total_step 38720, total loss is 27.58 , inference loss is 16.21, weight deacy loss is 11.37, training accuracy is 0.156250, time 224.841 samples/sec
+epoch 2, total_step 38740, total loss is 27.01 , inference loss is 15.64, weight deacy loss is 11.37, training accuracy is 0.187500, time 233.043 samples/sec
+epoch 2, total_step 38760, total loss is 26.68 , inference loss is 15.32, weight deacy loss is 11.36, training accuracy is 0.187500, time 197.741 samples/sec
+epoch 2, total_step 38780, total loss is 28.11 , inference loss is 16.74, weight deacy loss is 11.36, training accuracy is 0.187500, time 223.354 samples/sec
+epoch 2, total_step 38800, total loss is 24.73 , inference loss is 13.37, weight deacy loss is 11.36, training accuracy is 0.125000, time 217.293 samples/sec
+epoch 2, total_step 38820, total loss is 27.15 , inference loss is 15.79, weight deacy loss is 11.36, training accuracy is 0.156250, time 222.321 samples/sec
+epoch 2, total_step 38840, total loss is 30.42 , inference loss is 19.06, weight deacy loss is 11.36, training accuracy is 0.125000, time 222.362 samples/sec
+epoch 2, total_step 38860, total loss is 22.84 , inference loss is 11.48, weight deacy loss is 11.36, training accuracy is 0.281250, time 219.704 samples/sec
+epoch 2, total_step 38880, total loss is 23.89 , inference loss is 12.53, weight deacy loss is 11.36, training accuracy is 0.281250, time 212.326 samples/sec
+epoch 2, total_step 38900, total loss is 28.10 , inference loss is 16.75, weight deacy loss is 11.35, training accuracy is 0.156250, time 206.570 samples/sec
+epoch 2, total_step 38920, total loss is 29.14 , inference loss is 17.79, weight deacy loss is 11.35, training accuracy is 0.125000, time 212.069 samples/sec
+epoch 2, total_step 38940, total loss is 24.64 , inference loss is 13.29, weight deacy loss is 11.35, training accuracy is 0.218750, time 218.879 samples/sec
+epoch 2, total_step 38960, total loss is 26.00 , inference loss is 14.65, weight deacy loss is 11.35, training accuracy is 0.156250, time 220.097 samples/sec
+epoch 2, total_step 38980, total loss is 25.03 , inference loss is 13.68, weight deacy loss is 11.35, training accuracy is 0.187500, time 202.255 samples/sec
+epoch 2, total_step 39000, total loss is 27.59 , inference loss is 16.24, weight deacy loss is 11.35, training accuracy is 0.062500, time 13.786 samples/sec
+epoch 2, total_step 39020, total loss is 28.50 , inference loss is 17.15, weight deacy loss is 11.35, training accuracy is 0.218750, time 226.682 samples/sec
+epoch 2, total_step 39040, total loss is 27.94 , inference loss is 16.59, weight deacy loss is 11.34, training accuracy is 0.156250, time 214.084 samples/sec
+epoch 2, total_step 39060, total loss is 26.54 , inference loss is 15.19, weight deacy loss is 11.34, training accuracy is 0.156250, time 234.316 samples/sec
+epoch 2, total_step 39080, total loss is 28.40 , inference loss is 17.05, weight deacy loss is 11.34, training accuracy is 0.281250, time 220.562 samples/sec
+epoch 2, total_step 39100, total loss is 30.65 , inference loss is 19.31, weight deacy loss is 11.34, training accuracy is 0.093750, time 225.966 samples/sec
+epoch 2, total_step 39120, total loss is 29.41 , inference loss is 18.07, weight deacy loss is 11.34, training accuracy is 0.125000, time 300.480 samples/sec
+epoch 2, total_step 39140, total loss is 29.84 , inference loss is 18.50, weight deacy loss is 11.34, training accuracy is 0.062500, time 214.578 samples/sec
+epoch 2, total_step 39160, total loss is 26.49 , inference loss is 15.15, weight deacy loss is 11.33, training accuracy is 0.125000, time 211.491 samples/sec
+epoch 2, total_step 39180, total loss is 32.20 , inference loss is 20.86, weight deacy loss is 11.33, training accuracy is 0.093750, time 214.761 samples/sec
+epoch 2, total_step 39200, total loss is 25.47 , inference loss is 14.14, weight deacy loss is 11.33, training accuracy is 0.156250, time 216.376 samples/sec
+epoch 2, total_step 39220, total loss is 24.07 , inference loss is 12.74, weight deacy loss is 11.33, training accuracy is 0.250000, time 213.858 samples/sec
+epoch 2, total_step 39240, total loss is 27.12 , inference loss is 15.79, weight deacy loss is 11.33, training accuracy is 0.156250, time 218.192 samples/sec
+epoch 2, total_step 39260, total loss is 26.76 , inference loss is 15.43, weight deacy loss is 11.33, training accuracy is 0.093750, time 225.225 samples/sec
+epoch 2, total_step 39280, total loss is 30.11 , inference loss is 18.79, weight deacy loss is 11.33, training accuracy is 0.125000, time 298.510 samples/sec
+epoch 2, total_step 39300, total loss is 27.01 , inference loss is 15.68, weight deacy loss is 11.32, training accuracy is 0.281250, time 13.669 samples/sec
+epoch 2, total_step 39320, total loss is 28.26 , inference loss is 16.93, weight deacy loss is 11.32, training accuracy is 0.156250, time 220.837 samples/sec
+epoch 2, total_step 39340, total loss is 23.41 , inference loss is 12.08, weight deacy loss is 11.32, training accuracy is 0.218750, time 222.195 samples/sec
+epoch 2, total_step 39360, total loss is 29.15 , inference loss is 17.83, weight deacy loss is 11.32, training accuracy is 0.062500, time 214.812 samples/sec
+epoch 2, total_step 39380, total loss is 29.10 , inference loss is 17.78, weight deacy loss is 11.32, training accuracy is 0.125000, time 188.454 samples/sec
+epoch 2, total_step 39400, total loss is 28.05 , inference loss is 16.74, weight deacy loss is 11.32, training accuracy is 0.093750, time 227.146 samples/sec
+epoch 2, total_step 39420, total loss is 31.98 , inference loss is 20.66, weight deacy loss is 11.32, training accuracy is 0.062500, time 224.554 samples/sec
+epoch 2, total_step 39440, total loss is 31.47 , inference loss is 20.16, weight deacy loss is 11.31, training accuracy is 0.062500, time 224.896 samples/sec
+epoch 2, total_step 39460, total loss is 26.87 , inference loss is 15.55, weight deacy loss is 11.31, training accuracy is 0.125000, time 215.788 samples/sec
+epoch 2, total_step 39480, total loss is 23.52 , inference loss is 12.20, weight deacy loss is 11.31, training accuracy is 0.218750, time 231.343 samples/sec
+epoch 2, total_step 39500, total loss is 23.58 , inference loss is 12.27, weight deacy loss is 11.31, training accuracy is 0.125000, time 192.734 samples/sec
+epoch 2, total_step 39520, total loss is 28.97 , inference loss is 17.66, weight deacy loss is 11.31, training accuracy is 0.187500, time 211.721 samples/sec
+epoch 2, total_step 39540, total loss is 23.23 , inference loss is 11.92, weight deacy loss is 11.31, training accuracy is 0.187500, time 209.378 samples/sec
+epoch 2, total_step 39560, total loss is 30.41 , inference loss is 19.10, weight deacy loss is 11.31, training accuracy is 0.125000, time 221.438 samples/sec
+epoch 2, total_step 39580, total loss is 25.69 , inference loss is 14.39, weight deacy loss is 11.30, training accuracy is 0.187500, time 213.033 samples/sec
+epoch 2, total_step 39600, total loss is 28.76 , inference loss is 17.46, weight deacy loss is 11.30, training accuracy is 0.093750, time 12.458 samples/sec
+epoch 2, total_step 39620, total loss is 25.20 , inference loss is 13.90, weight deacy loss is 11.30, training accuracy is 0.156250, time 221.042 samples/sec
+epoch 2, total_step 39640, total loss is 29.23 , inference loss is 17.93, weight deacy loss is 11.30, training accuracy is 0.156250, time 182.377 samples/sec
+epoch 2, total_step 39660, total loss is 29.59 , inference loss is 18.29, weight deacy loss is 11.30, training accuracy is 0.062500, time 201.876 samples/sec
+epoch 2, total_step 39680, total loss is 27.10 , inference loss is 15.80, weight deacy loss is 11.30, training accuracy is 0.156250, time 220.769 samples/sec
+epoch 2, total_step 39700, total loss is 31.29 , inference loss is 19.99, weight deacy loss is 11.30, training accuracy is 0.093750, time 205.536 samples/sec
+epoch 2, total_step 39720, total loss is 24.56 , inference loss is 13.27, weight deacy loss is 11.29, training accuracy is 0.156250, time 237.898 samples/sec
+epoch 2, total_step 39740, total loss is 23.66 , inference loss is 12.36, weight deacy loss is 11.29, training accuracy is 0.250000, time 215.966 samples/sec
+epoch 2, total_step 39760, total loss is 26.59 , inference loss is 15.29, weight deacy loss is 11.29, training accuracy is 0.156250, time 218.738 samples/sec
+epoch 2, total_step 39780, total loss is 31.99 , inference loss is 20.70, weight deacy loss is 11.29, training accuracy is 0.062500, time 215.965 samples/sec
+epoch 2, total_step 39800, total loss is 25.41 , inference loss is 14.12, weight deacy loss is 11.29, training accuracy is 0.187500, time 210.656 samples/sec
+epoch 2, total_step 39820, total loss is 30.38 , inference loss is 19.10, weight deacy loss is 11.29, training accuracy is 0.125000, time 218.797 samples/sec
+epoch 2, total_step 39840, total loss is 28.34 , inference loss is 17.05, weight deacy loss is 11.28, training accuracy is 0.156250, time 218.308 samples/sec
+epoch 2, total_step 39860, total loss is 29.01 , inference loss is 17.73, weight deacy loss is 11.28, training accuracy is 0.125000, time 225.496 samples/sec
+epoch 2, total_step 39880, total loss is 28.57 , inference loss is 17.29, weight deacy loss is 11.28, training accuracy is 0.062500, time 229.364 samples/sec
+epoch 2, total_step 39900, total loss is 25.52 , inference loss is 14.24, weight deacy loss is 11.28, training accuracy is 0.156250, time 13.534 samples/sec
+epoch 2, total_step 39920, total loss is 26.28 , inference loss is 15.00, weight deacy loss is 11.28, training accuracy is 0.125000, time 207.895 samples/sec
+epoch 2, total_step 39940, total loss is 24.69 , inference loss is 13.41, weight deacy loss is 11.28, training accuracy is 0.093750, time 222.202 samples/sec
+epoch 2, total_step 39960, total loss is 22.17 , inference loss is 10.89, weight deacy loss is 11.28, training accuracy is 0.250000, time 235.257 samples/sec
+epoch 2, total_step 39980, total loss is 23.13 , inference loss is 11.85, weight deacy loss is 11.27, training accuracy is 0.187500, time 210.554 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.965379000000006
+best_threshold_index 108 0.9783333333333334
+best_threshold_index 108 0.9785185185185186
+best_threshold_index 108 0.9788888888888889
+best_threshold_index 107 0.9790740740740741
+best_threshold_index 108 0.9787037037037037
+best_threshold_index 108 0.9777777777777777
+best_threshold_index 108 0.9781481481481481
+best_threshold_index 108 0.9775925925925926
+best_threshold_index 107 0.9772222222222222
+best_threshold_index 107 0.9777777777777777
+[lfw][40000]XNorm: 23.710506
+[lfw][40000]Accuracy-Flip: 0.97733+-0.00507
+test accuracy is: 0.9773333333333334
+epoch 2, total_step 40000, total loss is 22.98 , inference loss is 11.71, weight deacy loss is 11.27, training accuracy is 0.187500, time 16.411 samples/sec
+epoch 2, total_step 40020, total loss is 30.18 , inference loss is 18.91, weight deacy loss is 11.27, training accuracy is 0.156250, time 216.557 samples/sec
+epoch 2, total_step 40040, total loss is 26.36 , inference loss is 15.09, weight deacy loss is 11.27, training accuracy is 0.187500, time 216.566 samples/sec
+epoch 2, total_step 40060, total loss is 29.58 , inference loss is 18.31, weight deacy loss is 11.27, training accuracy is 0.156250, time 217.105 samples/sec
+epoch 2, total_step 40080, total loss is 26.55 , inference loss is 15.28, weight deacy loss is 11.27, training accuracy is 0.156250, time 226.347 samples/sec
+epoch 2, total_step 40100, total loss is 26.19 , inference loss is 14.92, weight deacy loss is 11.27, training accuracy is 0.187500, time 236.098 samples/sec
+epoch 2, total_step 40120, total loss is 26.95 , inference loss is 15.69, weight deacy loss is 11.26, training accuracy is 0.187500, time 236.007 samples/sec
+epoch 2, total_step 40140, total loss is 28.03 , inference loss is 16.77, weight deacy loss is 11.26, training accuracy is 0.093750, time 243.692 samples/sec
+epoch 2, total_step 40160, total loss is 27.43 , inference loss is 16.17, weight deacy loss is 11.26, training accuracy is 0.125000, time 218.948 samples/sec
+epoch 2, total_step 40180, total loss is 26.81 , inference loss is 15.55, weight deacy loss is 11.26, training accuracy is 0.281250, time 206.183 samples/sec
+epoch 2, total_step 40200, total loss is 20.99 , inference loss is 9.73, weight deacy loss is 11.26, training accuracy is 0.250000, time 13.174 samples/sec
+epoch 2, total_step 40220, total loss is 28.55 , inference loss is 17.30, weight deacy loss is 11.26, training accuracy is 0.125000, time 296.233 samples/sec
+epoch 2, total_step 40240, total loss is 24.27 , inference loss is 13.01, weight deacy loss is 11.26, training accuracy is 0.250000, time 211.662 samples/sec
+epoch 2, total_step 40260, total loss is 26.51 , inference loss is 15.26, weight deacy loss is 11.25, training accuracy is 0.125000, time 201.888 samples/sec
+epoch 2, total_step 40280, total loss is 28.57 , inference loss is 17.32, weight deacy loss is 11.25, training accuracy is 0.187500, time 224.839 samples/sec
+epoch 2, total_step 40300, total loss is 24.61 , inference loss is 13.36, weight deacy loss is 11.25, training accuracy is 0.156250, time 211.181 samples/sec
+epoch 2, total_step 40320, total loss is 23.63 , inference loss is 12.38, weight deacy loss is 11.25, training accuracy is 0.156250, time 217.287 samples/sec
+epoch 2, total_step 40340, total loss is 28.41 , inference loss is 17.16, weight deacy loss is 11.25, training accuracy is 0.250000, time 220.647 samples/sec
+epoch 2, total_step 40360, total loss is 25.88 , inference loss is 14.63, weight deacy loss is 11.25, training accuracy is 0.281250, time 219.882 samples/sec
+epoch 2, total_step 40380, total loss is 24.50 , inference loss is 13.26, weight deacy loss is 11.25, training accuracy is 0.093750, time 217.709 samples/sec
+epoch 2, total_step 40400, total loss is 24.83 , inference loss is 13.59, weight deacy loss is 11.24, training accuracy is 0.187500, time 202.497 samples/sec
+epoch 2, total_step 40420, total loss is 22.03 , inference loss is 10.78, weight deacy loss is 11.24, training accuracy is 0.187500, time 219.686 samples/sec
+epoch 2, total_step 40440, total loss is 28.94 , inference loss is 17.70, weight deacy loss is 11.24, training accuracy is 0.031250, time 219.134 samples/sec
+epoch 2, total_step 40460, total loss is 24.76 , inference loss is 13.52, weight deacy loss is 11.24, training accuracy is 0.250000, time 216.221 samples/sec
+epoch 2, total_step 40480, total loss is 27.31 , inference loss is 16.07, weight deacy loss is 11.24, training accuracy is 0.156250, time 236.007 samples/sec
+epoch 2, total_step 40500, total loss is 21.74 , inference loss is 10.50, weight deacy loss is 11.24, training accuracy is 0.250000, time 13.742 samples/sec
+epoch 2, total_step 40520, total loss is 25.27 , inference loss is 14.03, weight deacy loss is 11.24, training accuracy is 0.218750, time 221.147 samples/sec
+epoch 2, total_step 40540, total loss is 25.18 , inference loss is 13.94, weight deacy loss is 11.23, training accuracy is 0.187500, time 212.499 samples/sec
+epoch 2, total_step 40560, total loss is 24.86 , inference loss is 13.63, weight deacy loss is 11.23, training accuracy is 0.250000, time 218.863 samples/sec
+epoch 2, total_step 40580, total loss is 23.74 , inference loss is 12.51, weight deacy loss is 11.23, training accuracy is 0.218750, time 215.845 samples/sec
+epoch 2, total_step 40600, total loss is 27.76 , inference loss is 16.53, weight deacy loss is 11.23, training accuracy is 0.093750, time 219.174 samples/sec
+epoch 2, total_step 40620, total loss is 26.24 , inference loss is 15.02, weight deacy loss is 11.23, training accuracy is 0.250000, time 214.549 samples/sec
+epoch 2, total_step 40640, total loss is 27.40 , inference loss is 16.18, weight deacy loss is 11.23, training accuracy is 0.156250, time 224.070 samples/sec
+epoch 2, total_step 40660, total loss is 24.20 , inference loss is 12.97, weight deacy loss is 11.23, training accuracy is 0.125000, time 211.467 samples/sec
+epoch 2, total_step 40680, total loss is 22.13 , inference loss is 10.91, weight deacy loss is 11.22, training accuracy is 0.281250, time 228.771 samples/sec
+epoch 2, total_step 40700, total loss is 26.16 , inference loss is 14.94, weight deacy loss is 11.22, training accuracy is 0.156250, time 215.227 samples/sec
+epoch 2, total_step 40720, total loss is 30.43 , inference loss is 19.21, weight deacy loss is 11.22, training accuracy is 0.093750, time 222.348 samples/sec
+epoch 2, total_step 40740, total loss is 26.62 , inference loss is 15.40, weight deacy loss is 11.22, training accuracy is 0.156250, time 218.601 samples/sec
+epoch 2, total_step 40760, total loss is 21.70 , inference loss is 10.48, weight deacy loss is 11.22, training accuracy is 0.218750, time 202.471 samples/sec
+epoch 2, total_step 40780, total loss is 25.81 , inference loss is 14.59, weight deacy loss is 11.22, training accuracy is 0.156250, time 202.088 samples/sec
+epoch 2, total_step 40800, total loss is 24.03 , inference loss is 12.81, weight deacy loss is 11.21, training accuracy is 0.156250, time 13.628 samples/sec
+epoch 2, total_step 40820, total loss is 24.76 , inference loss is 13.55, weight deacy loss is 11.21, training accuracy is 0.156250, time 211.979 samples/sec
+epoch 2, total_step 40840, total loss is 25.28 , inference loss is 14.07, weight deacy loss is 11.21, training accuracy is 0.218750, time 213.057 samples/sec
+epoch 2, total_step 40860, total loss is 29.10 , inference loss is 17.89, weight deacy loss is 11.21, training accuracy is 0.187500, time 212.966 samples/sec
+epoch 2, total_step 40880, total loss is 24.89 , inference loss is 13.68, weight deacy loss is 11.21, training accuracy is 0.125000, time 213.732 samples/sec
+epoch 2, total_step 40900, total loss is 26.33 , inference loss is 15.12, weight deacy loss is 11.21, training accuracy is 0.218750, time 213.406 samples/sec
+epoch 2, total_step 40920, total loss is 22.18 , inference loss is 10.97, weight deacy loss is 11.21, training accuracy is 0.187500, time 211.456 samples/sec
+epoch 2, total_step 40940, total loss is 28.17 , inference loss is 16.97, weight deacy loss is 11.20, training accuracy is 0.156250, time 215.716 samples/sec
+epoch 2, total_step 40960, total loss is 31.54 , inference loss is 20.34, weight deacy loss is 11.20, training accuracy is 0.093750, time 213.995 samples/sec
+epoch 2, total_step 40980, total loss is 24.69 , inference loss is 13.49, weight deacy loss is 11.20, training accuracy is 0.187500, time 219.721 samples/sec
+epoch 2, total_step 41000, total loss is 23.22 , inference loss is 12.02, weight deacy loss is 11.20, training accuracy is 0.343750, time 300.751 samples/sec
+epoch 2, total_step 41020, total loss is 26.77 , inference loss is 15.57, weight deacy loss is 11.20, training accuracy is 0.218750, time 297.856 samples/sec
+epoch 2, total_step 41040, total loss is 27.15 , inference loss is 15.95, weight deacy loss is 11.20, training accuracy is 0.187500, time 211.856 samples/sec
+epoch 2, total_step 41060, total loss is 26.77 , inference loss is 15.57, weight deacy loss is 11.20, training accuracy is 0.156250, time 219.671 samples/sec
+epoch 2, total_step 41080, total loss is 22.07 , inference loss is 10.88, weight deacy loss is 11.19, training accuracy is 0.250000, time 228.586 samples/sec
+epoch 2, total_step 41100, total loss is 26.07 , inference loss is 14.87, weight deacy loss is 11.19, training accuracy is 0.156250, time 13.700 samples/sec
+epoch 2, total_step 41120, total loss is 27.50 , inference loss is 16.31, weight deacy loss is 11.19, training accuracy is 0.125000, time 211.774 samples/sec
+epoch 2, total_step 41140, total loss is 22.76 , inference loss is 11.57, weight deacy loss is 11.19, training accuracy is 0.218750, time 229.970 samples/sec
+epoch 2, total_step 41160, total loss is 25.70 , inference loss is 14.52, weight deacy loss is 11.19, training accuracy is 0.187500, time 221.522 samples/sec
+epoch 2, total_step 41180, total loss is 25.35 , inference loss is 14.16, weight deacy loss is 11.19, training accuracy is 0.156250, time 297.857 samples/sec
+epoch 2, total_step 41200, total loss is 26.72 , inference loss is 15.54, weight deacy loss is 11.19, training accuracy is 0.281250, time 219.257 samples/sec
+epoch 2, total_step 41220, total loss is 24.63 , inference loss is 13.44, weight deacy loss is 11.18, training accuracy is 0.187500, time 218.862 samples/sec
+epoch 2, total_step 41240, total loss is 26.50 , inference loss is 15.32, weight deacy loss is 11.18, training accuracy is 0.125000, time 225.199 samples/sec
+epoch 2, total_step 41260, total loss is 25.13 , inference loss is 13.95, weight deacy loss is 11.18, training accuracy is 0.156250, time 224.711 samples/sec
+epoch 2, total_step 41280, total loss is 29.57 , inference loss is 18.39, weight deacy loss is 11.18, training accuracy is 0.156250, time 219.104 samples/sec
+epoch 2, total_step 41300, total loss is 25.18 , inference loss is 14.00, weight deacy loss is 11.18, training accuracy is 0.125000, time 220.888 samples/sec
+epoch 2, total_step 41320, total loss is 23.28 , inference loss is 12.10, weight deacy loss is 11.18, training accuracy is 0.250000, time 219.555 samples/sec
+epoch 2, total_step 41340, total loss is 24.78 , inference loss is 13.60, weight deacy loss is 11.18, training accuracy is 0.125000, time 210.538 samples/sec
+epoch 2, total_step 41360, total loss is 28.78 , inference loss is 17.60, weight deacy loss is 11.17, training accuracy is 0.156250, time 214.859 samples/sec
+epoch 2, total_step 41380, total loss is 26.33 , inference loss is 15.16, weight deacy loss is 11.17, training accuracy is 0.156250, time 230.957 samples/sec
+epoch 2, total_step 41400, total loss is 23.70 , inference loss is 12.53, weight deacy loss is 11.17, training accuracy is 0.281250, time 13.700 samples/sec
+epoch 2, total_step 41420, total loss is 26.39 , inference loss is 15.22, weight deacy loss is 11.17, training accuracy is 0.187500, time 223.391 samples/sec
+epoch 2, total_step 41440, total loss is 20.64 , inference loss is 9.47, weight deacy loss is 11.17, training accuracy is 0.250000, time 300.149 samples/sec
+epoch 2, total_step 41460, total loss is 27.73 , inference loss is 16.56, weight deacy loss is 11.17, training accuracy is 0.187500, time 195.933 samples/sec
+epoch 2, total_step 41480, total loss is 26.07 , inference loss is 14.91, weight deacy loss is 11.17, training accuracy is 0.218750, time 234.044 samples/sec
+epoch 2, total_step 41500, total loss is 27.36 , inference loss is 16.20, weight deacy loss is 11.16, training accuracy is 0.062500, time 210.032 samples/sec
+epoch 2, total_step 41520, total loss is 24.80 , inference loss is 13.64, weight deacy loss is 11.16, training accuracy is 0.281250, time 209.627 samples/sec
+epoch 2, total_step 41540, total loss is 28.16 , inference loss is 17.00, weight deacy loss is 11.16, training accuracy is 0.156250, time 222.534 samples/sec
+epoch 2, total_step 41560, total loss is 26.92 , inference loss is 15.76, weight deacy loss is 11.16, training accuracy is 0.031250, time 210.098 samples/sec
+epoch 2, total_step 41580, total loss is 27.93 , inference loss is 16.77, weight deacy loss is 11.16, training accuracy is 0.062500, time 213.673 samples/sec
+epoch 2, total_step 41600, total loss is 26.35 , inference loss is 15.19, weight deacy loss is 11.16, training accuracy is 0.156250, time 217.279 samples/sec
+epoch 2, total_step 41620, total loss is 22.48 , inference loss is 11.32, weight deacy loss is 11.16, training accuracy is 0.250000, time 202.424 samples/sec
+epoch 2, total_step 41640, total loss is 27.83 , inference loss is 16.67, weight deacy loss is 11.15, training accuracy is 0.187500, time 213.627 samples/sec
+epoch 2, total_step 41660, total loss is 24.79 , inference loss is 13.64, weight deacy loss is 11.15, training accuracy is 0.250000, time 224.935 samples/sec
+epoch 2, total_step 41680, total loss is 24.73 , inference loss is 13.57, weight deacy loss is 11.15, training accuracy is 0.218750, time 230.031 samples/sec
+epoch 2, total_step 41700, total loss is 20.47 , inference loss is 9.32, weight deacy loss is 11.15, training accuracy is 0.343750, time 14.210 samples/sec
+epoch 2, total_step 41720, total loss is 24.46 , inference loss is 13.31, weight deacy loss is 11.15, training accuracy is 0.187500, time 220.073 samples/sec
+epoch 2, total_step 41740, total loss is 28.54 , inference loss is 17.39, weight deacy loss is 11.15, training accuracy is 0.218750, time 221.848 samples/sec
+epoch 2, total_step 41760, total loss is 22.50 , inference loss is 11.35, weight deacy loss is 11.15, training accuracy is 0.281250, time 221.938 samples/sec
+epoch 2, total_step 41780, total loss is 28.01 , inference loss is 16.86, weight deacy loss is 11.14, training accuracy is 0.093750, time 205.738 samples/sec
+epoch 2, total_step 41800, total loss is 28.34 , inference loss is 17.20, weight deacy loss is 11.14, training accuracy is 0.156250, time 216.605 samples/sec
+epoch 2, total_step 41820, total loss is 27.93 , inference loss is 16.78, weight deacy loss is 11.14, training accuracy is 0.187500, time 212.424 samples/sec
+epoch 2, total_step 41840, total loss is 25.88 , inference loss is 14.74, weight deacy loss is 11.14, training accuracy is 0.281250, time 220.408 samples/sec
+epoch 2, total_step 41860, total loss is 25.96 , inference loss is 14.82, weight deacy loss is 11.14, training accuracy is 0.093750, time 224.444 samples/sec
+epoch 2, total_step 41880, total loss is 26.52 , inference loss is 15.38, weight deacy loss is 11.14, training accuracy is 0.187500, time 201.020 samples/sec
+epoch 2, total_step 41900, total loss is 28.35 , inference loss is 17.21, weight deacy loss is 11.14, training accuracy is 0.093750, time 239.436 samples/sec
+epoch 2, total_step 41920, total loss is 25.59 , inference loss is 14.46, weight deacy loss is 11.13, training accuracy is 0.218750, time 220.015 samples/sec
+epoch 2, total_step 41940, total loss is 24.44 , inference loss is 13.30, weight deacy loss is 11.13, training accuracy is 0.156250, time 215.218 samples/sec
+epoch 2, total_step 41960, total loss is 29.07 , inference loss is 17.94, weight deacy loss is 11.13, training accuracy is 0.125000, time 217.089 samples/sec
+epoch 2, total_step 41980, total loss is 25.75 , inference loss is 14.62, weight deacy loss is 11.13, training accuracy is 0.093750, time 219.052 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.759942999999993
+best_threshold_index 109 0.977037037037037
+best_threshold_index 109 0.9774074074074074
+best_threshold_index 109 0.9781481481481481
+best_threshold_index 109 0.9781481481481481
+best_threshold_index 107 0.9774074074074074
+best_threshold_index 109 0.9764814814814815
+best_threshold_index 109 0.9764814814814815
+best_threshold_index 109 0.975925925925926
+best_threshold_index 109 0.9757407407407407
+best_threshold_index 109 0.9755555555555555
+[lfw][42000]XNorm: 23.254394
+[lfw][42000]Accuracy-Flip: 0.97650+-0.00835
+test accuracy is: 0.9765
+epoch 2, total_step 42000, total loss is 26.44 , inference loss is 15.31, weight deacy loss is 11.13, training accuracy is 0.187500, time 16.476 samples/sec
+epoch 2, total_step 42020, total loss is 33.87 , inference loss is 22.74, weight deacy loss is 11.13, training accuracy is 0.031250, time 214.936 samples/sec
+epoch 2, total_step 42040, total loss is 27.09 , inference loss is 15.96, weight deacy loss is 11.13, training accuracy is 0.031250, time 223.292 samples/sec
+epoch 2, total_step 42060, total loss is 31.56 , inference loss is 20.44, weight deacy loss is 11.12, training accuracy is 0.125000, time 218.249 samples/sec
+epoch 2, total_step 42080, total loss is 23.39 , inference loss is 12.27, weight deacy loss is 11.12, training accuracy is 0.187500, time 216.980 samples/sec
+epoch 2, total_step 42100, total loss is 27.47 , inference loss is 16.35, weight deacy loss is 11.12, training accuracy is 0.062500, time 298.037 samples/sec
+epoch 2, total_step 42120, total loss is 23.29 , inference loss is 12.17, weight deacy loss is 11.12, training accuracy is 0.187500, time 227.841 samples/sec
+epoch 2, total_step 42140, total loss is 34.19 , inference loss is 23.07, weight deacy loss is 11.12, training accuracy is 0.031250, time 209.321 samples/sec
+epoch 2, total_step 42160, total loss is 26.44 , inference loss is 15.32, weight deacy loss is 11.12, training accuracy is 0.125000, time 211.379 samples/sec
+epoch 2, total_step 42180, total loss is 23.00 , inference loss is 11.89, weight deacy loss is 11.12, training accuracy is 0.125000, time 217.773 samples/sec
+epoch 2, total_step 42200, total loss is 29.92 , inference loss is 18.80, weight deacy loss is 11.11, training accuracy is 0.031250, time 228.216 samples/sec
+epoch 2, total_step 42220, total loss is 26.43 , inference loss is 15.32, weight deacy loss is 11.11, training accuracy is 0.062500, time 214.773 samples/sec
+epoch 2, total_step 42240, total loss is 25.61 , inference loss is 14.50, weight deacy loss is 11.11, training accuracy is 0.156250, time 225.727 samples/sec
+epoch 2, total_step 42260, total loss is 24.83 , inference loss is 13.72, weight deacy loss is 11.11, training accuracy is 0.125000, time 221.542 samples/sec
+epoch 2, total_step 42280, total loss is 29.53 , inference loss is 18.42, weight deacy loss is 11.11, training accuracy is 0.125000, time 222.719 samples/sec
+epoch 2, total_step 42300, total loss is 24.78 , inference loss is 13.67, weight deacy loss is 11.11, training accuracy is 0.093750, time 13.015 samples/sec
+epoch 2, total_step 42320, total loss is 27.94 , inference loss is 16.83, weight deacy loss is 11.11, training accuracy is 0.218750, time 237.259 samples/sec
+epoch 2, total_step 42340, total loss is 24.91 , inference loss is 13.81, weight deacy loss is 11.10, training accuracy is 0.093750, time 214.915 samples/sec
+epoch 2, total_step 42360, total loss is 27.77 , inference loss is 16.66, weight deacy loss is 11.10, training accuracy is 0.125000, time 298.217 samples/sec
+epoch 2, total_step 42380, total loss is 24.68 , inference loss is 13.57, weight deacy loss is 11.10, training accuracy is 0.187500, time 214.439 samples/sec
+epoch 2, total_step 42400, total loss is 25.26 , inference loss is 14.16, weight deacy loss is 11.10, training accuracy is 0.187500, time 222.217 samples/sec
+epoch 2, total_step 42420, total loss is 24.55 , inference loss is 13.45, weight deacy loss is 11.10, training accuracy is 0.187500, time 228.990 samples/sec
+epoch 2, total_step 42440, total loss is 26.27 , inference loss is 15.17, weight deacy loss is 11.10, training accuracy is 0.125000, time 226.909 samples/sec
+epoch 2, total_step 42460, total loss is 28.68 , inference loss is 17.58, weight deacy loss is 11.10, training accuracy is 0.093750, time 222.912 samples/sec
+epoch 2, total_step 42480, total loss is 24.59 , inference loss is 13.49, weight deacy loss is 11.09, training accuracy is 0.093750, time 221.279 samples/sec
+epoch 2, total_step 42500, total loss is 25.62 , inference loss is 14.52, weight deacy loss is 11.09, training accuracy is 0.062500, time 214.215 samples/sec
+epoch 2, total_step 42520, total loss is 32.28 , inference loss is 21.19, weight deacy loss is 11.09, training accuracy is 0.156250, time 224.185 samples/sec
+epoch 2, total_step 42540, total loss is 27.89 , inference loss is 16.80, weight deacy loss is 11.09, training accuracy is 0.062500, time 225.804 samples/sec
+epoch 2, total_step 42560, total loss is 26.77 , inference loss is 15.69, weight deacy loss is 11.09, training accuracy is 0.000000, time 208.705 samples/sec
+epoch 2, total_step 42580, total loss is 29.80 , inference loss is 18.71, weight deacy loss is 11.09, training accuracy is 0.062500, time 213.654 samples/sec
+epoch 2, total_step 42600, total loss is 24.93 , inference loss is 13.84, weight deacy loss is 11.09, training accuracy is 0.218750, time 12.068 samples/sec
+epoch 2, total_step 42620, total loss is 26.18 , inference loss is 15.09, weight deacy loss is 11.08, training accuracy is 0.093750, time 294.775 samples/sec
+epoch 2, total_step 42640, total loss is 28.43 , inference loss is 17.35, weight deacy loss is 11.08, training accuracy is 0.031250, time 215.417 samples/sec
+epoch 2, total_step 42660, total loss is 28.49 , inference loss is 17.40, weight deacy loss is 11.08, training accuracy is 0.218750, time 211.729 samples/sec
+epoch 2, total_step 42680, total loss is 27.61 , inference loss is 16.53, weight deacy loss is 11.08, training accuracy is 0.156250, time 225.284 samples/sec
+epoch 2, total_step 42700, total loss is 27.86 , inference loss is 16.79, weight deacy loss is 11.08, training accuracy is 0.125000, time 215.357 samples/sec
+epoch 2, total_step 42720, total loss is 27.93 , inference loss is 16.85, weight deacy loss is 11.08, training accuracy is 0.125000, time 213.790 samples/sec
+epoch 2, total_step 42740, total loss is 24.84 , inference loss is 13.76, weight deacy loss is 11.08, training accuracy is 0.062500, time 215.630 samples/sec
+epoch 2, total_step 42760, total loss is 25.84 , inference loss is 14.76, weight deacy loss is 11.07, training accuracy is 0.093750, time 215.379 samples/sec
+epoch 2, total_step 42780, total loss is 25.94 , inference loss is 14.87, weight deacy loss is 11.07, training accuracy is 0.156250, time 297.569 samples/sec
+epoch 2, total_step 42800, total loss is 24.97 , inference loss is 13.90, weight deacy loss is 11.07, training accuracy is 0.125000, time 298.232 samples/sec
+epoch 2, total_step 42820, total loss is 23.76 , inference loss is 12.69, weight deacy loss is 11.07, training accuracy is 0.187500, time 209.450 samples/sec
+epoch 2, total_step 42840, total loss is 22.99 , inference loss is 11.93, weight deacy loss is 11.07, training accuracy is 0.125000, time 209.427 samples/sec
+epoch 2, total_step 42860, total loss is 28.94 , inference loss is 17.87, weight deacy loss is 11.07, training accuracy is 0.187500, time 218.097 samples/sec
+epoch 2, total_step 42880, total loss is 27.25 , inference loss is 16.19, weight deacy loss is 11.07, training accuracy is 0.187500, time 222.951 samples/sec
+epoch 2, total_step 42900, total loss is 25.76 , inference loss is 14.70, weight deacy loss is 11.07, training accuracy is 0.125000, time 11.354 samples/sec
+epoch 2, total_step 42920, total loss is 28.20 , inference loss is 17.14, weight deacy loss is 11.06, training accuracy is 0.156250, time 207.593 samples/sec
+epoch 2, total_step 42940, total loss is 28.20 , inference loss is 17.14, weight deacy loss is 11.06, training accuracy is 0.093750, time 214.036 samples/sec
+epoch 2, total_step 42960, total loss is 27.52 , inference loss is 16.46, weight deacy loss is 11.06, training accuracy is 0.125000, time 191.415 samples/sec
+epoch 2, total_step 42980, total loss is 29.37 , inference loss is 18.31, weight deacy loss is 11.06, training accuracy is 0.125000, time 222.248 samples/sec
+epoch 2, total_step 43000, total loss is 28.41 , inference loss is 17.35, weight deacy loss is 11.06, training accuracy is 0.125000, time 222.368 samples/sec
+epoch 2, total_step 43020, total loss is 25.11 , inference loss is 14.05, weight deacy loss is 11.06, training accuracy is 0.156250, time 223.483 samples/sec
+epoch 2, total_step 43040, total loss is 25.61 , inference loss is 14.55, weight deacy loss is 11.06, training accuracy is 0.218750, time 213.008 samples/sec
+epoch 2, total_step 43060, total loss is 25.90 , inference loss is 14.84, weight deacy loss is 11.05, training accuracy is 0.218750, time 297.879 samples/sec
+epoch 2, total_step 43080, total loss is 27.17 , inference loss is 16.12, weight deacy loss is 11.05, training accuracy is 0.156250, time 204.884 samples/sec
+epoch 2, total_step 43100, total loss is 30.39 , inference loss is 19.34, weight deacy loss is 11.05, training accuracy is 0.031250, time 212.937 samples/sec
+epoch 2, total_step 43120, total loss is 22.61 , inference loss is 11.56, weight deacy loss is 11.05, training accuracy is 0.281250, time 205.128 samples/sec
+epoch 2, total_step 43140, total loss is 24.23 , inference loss is 13.18, weight deacy loss is 11.05, training accuracy is 0.156250, time 210.919 samples/sec
+epoch 2, total_step 43160, total loss is 29.93 , inference loss is 18.89, weight deacy loss is 11.05, training accuracy is 0.125000, time 203.559 samples/sec
+epoch 2, total_step 43180, total loss is 27.86 , inference loss is 16.81, weight deacy loss is 11.05, training accuracy is 0.093750, time 224.078 samples/sec
+epoch 2, total_step 43200, total loss is 26.42 , inference loss is 15.37, weight deacy loss is 11.04, training accuracy is 0.156250, time 13.991 samples/sec
+epoch 2, total_step 43220, total loss is 26.28 , inference loss is 15.24, weight deacy loss is 11.04, training accuracy is 0.062500, time 232.435 samples/sec
+epoch 2, total_step 43240, total loss is 24.62 , inference loss is 13.57, weight deacy loss is 11.04, training accuracy is 0.093750, time 191.255 samples/sec
+epoch 2, total_step 43260, total loss is 29.32 , inference loss is 18.28, weight deacy loss is 11.04, training accuracy is 0.062500, time 213.939 samples/sec
+epoch 2, total_step 43280, total loss is 25.07 , inference loss is 14.03, weight deacy loss is 11.04, training accuracy is 0.125000, time 222.876 samples/sec
+epoch 2, total_step 43300, total loss is 30.44 , inference loss is 19.41, weight deacy loss is 11.04, training accuracy is 0.125000, time 216.551 samples/sec
+epoch 2, total_step 43320, total loss is 25.80 , inference loss is 14.77, weight deacy loss is 11.04, training accuracy is 0.218750, time 190.764 samples/sec
+epoch 2, total_step 43340, total loss is 23.81 , inference loss is 12.78, weight deacy loss is 11.03, training accuracy is 0.125000, time 189.983 samples/sec
+epoch 2, total_step 43360, total loss is 26.63 , inference loss is 15.60, weight deacy loss is 11.03, training accuracy is 0.156250, time 216.471 samples/sec
+epoch 2, total_step 43380, total loss is 27.09 , inference loss is 16.06, weight deacy loss is 11.03, training accuracy is 0.093750, time 213.902 samples/sec
+epoch 2, total_step 43400, total loss is 28.71 , inference loss is 17.68, weight deacy loss is 11.03, training accuracy is 0.093750, time 194.639 samples/sec
+epoch 2, total_step 43420, total loss is 24.95 , inference loss is 13.92, weight deacy loss is 11.03, training accuracy is 0.312500, time 220.969 samples/sec
+epoch 2, total_step 43440, total loss is 28.31 , inference loss is 17.28, weight deacy loss is 11.03, training accuracy is 0.156250, time 211.762 samples/sec
+epoch 2, total_step 43460, total loss is 27.74 , inference loss is 16.71, weight deacy loss is 11.03, training accuracy is 0.156250, time 213.666 samples/sec
+epoch 2, total_step 43480, total loss is 28.40 , inference loss is 17.38, weight deacy loss is 11.02, training accuracy is 0.062500, time 217.597 samples/sec
+epoch 2, total_step 43500, total loss is 32.77 , inference loss is 21.75, weight deacy loss is 11.02, training accuracy is 0.062500, time 13.356 samples/sec
+epoch 2, total_step 43520, total loss is 26.52 , inference loss is 15.49, weight deacy loss is 11.02, training accuracy is 0.031250, time 222.065 samples/sec
+epoch 2, total_step 43540, total loss is 24.72 , inference loss is 13.70, weight deacy loss is 11.02, training accuracy is 0.062500, time 226.064 samples/sec
+epoch 2, total_step 43560, total loss is 21.06 , inference loss is 10.04, weight deacy loss is 11.02, training accuracy is 0.156250, time 180.038 samples/sec
+epoch 2, total_step 43580, total loss is 27.46 , inference loss is 16.44, weight deacy loss is 11.02, training accuracy is 0.125000, time 210.845 samples/sec
+epoch 2, total_step 43600, total loss is 25.63 , inference loss is 14.61, weight deacy loss is 11.02, training accuracy is 0.187500, time 177.419 samples/sec
+epoch 2, total_step 43620, total loss is 29.13 , inference loss is 18.12, weight deacy loss is 11.01, training accuracy is 0.187500, time 205.283 samples/sec
+epoch 2, total_step 43640, total loss is 28.20 , inference loss is 17.18, weight deacy loss is 11.01, training accuracy is 0.031250, time 215.997 samples/sec
+epoch 2, total_step 43660, total loss is 25.37 , inference loss is 14.36, weight deacy loss is 11.01, training accuracy is 0.125000, time 216.649 samples/sec
+epoch 2, total_step 43680, total loss is 25.17 , inference loss is 14.16, weight deacy loss is 11.01, training accuracy is 0.250000, time 224.735 samples/sec
+epoch 2, total_step 43700, total loss is 24.11 , inference loss is 13.10, weight deacy loss is 11.01, training accuracy is 0.125000, time 226.307 samples/sec
+epoch 2, total_step 43720, total loss is 24.33 , inference loss is 13.33, weight deacy loss is 11.01, training accuracy is 0.062500, time 210.882 samples/sec
+epoch 2, total_step 43740, total loss is 32.97 , inference loss is 21.97, weight deacy loss is 11.01, training accuracy is 0.031250, time 202.592 samples/sec
+epoch 2, total_step 43760, total loss is 25.34 , inference loss is 14.33, weight deacy loss is 11.01, training accuracy is 0.125000, time 226.063 samples/sec
+epoch 2, total_step 43780, total loss is 27.12 , inference loss is 16.12, weight deacy loss is 11.00, training accuracy is 0.125000, time 223.988 samples/sec
+epoch 2, total_step 43800, total loss is 30.23 , inference loss is 19.23, weight deacy loss is 11.00, training accuracy is 0.093750, time 14.221 samples/sec
+epoch 2, total_step 43820, total loss is 29.12 , inference loss is 18.12, weight deacy loss is 11.00, training accuracy is 0.156250, time 221.164 samples/sec
+epoch 2, total_step 43840, total loss is 29.85 , inference loss is 18.85, weight deacy loss is 11.00, training accuracy is 0.125000, time 226.488 samples/sec
+epoch 2, total_step 43860, total loss is 28.24 , inference loss is 17.24, weight deacy loss is 11.00, training accuracy is 0.062500, time 298.863 samples/sec
+epoch 2, total_step 43880, total loss is 28.06 , inference loss is 17.07, weight deacy loss is 11.00, training accuracy is 0.093750, time 217.114 samples/sec
+epoch 2, total_step 43900, total loss is 28.05 , inference loss is 17.06, weight deacy loss is 11.00, training accuracy is 0.031250, time 201.691 samples/sec
+epoch 2, total_step 43920, total loss is 28.41 , inference loss is 17.41, weight deacy loss is 10.99, training accuracy is 0.187500, time 227.371 samples/sec
+epoch 2, total_step 43940, total loss is 23.34 , inference loss is 12.35, weight deacy loss is 10.99, training accuracy is 0.250000, time 228.336 samples/sec
+epoch 2, total_step 43960, total loss is 24.66 , inference loss is 13.67, weight deacy loss is 10.99, training accuracy is 0.187500, time 211.806 samples/sec
+epoch 2, total_step 43980, total loss is 26.71 , inference loss is 15.72, weight deacy loss is 10.99, training accuracy is 0.093750, time 236.938 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.225942999999994
+best_threshold_index 106 0.9748148148148148
+best_threshold_index 106 0.9742592592592593
+best_threshold_index 106 0.9764814814814815
+best_threshold_index 105 0.9764814814814815
+best_threshold_index 106 0.9753703703703703
+best_threshold_index 106 0.9737037037037037
+best_threshold_index 106 0.9737037037037037
+best_threshold_index 105 0.9735185185185186
+best_threshold_index 106 0.9731481481481481
+best_threshold_index 106 0.9735185185185186
+[lfw][44000]XNorm: 23.642517
+[lfw][44000]Accuracy-Flip: 0.97383+-0.01093
+test accuracy is: 0.9738333333333333
+epoch 2, total_step 44000, total loss is 24.48 , inference loss is 13.49, weight deacy loss is 10.99, training accuracy is 0.156250, time 17.082 samples/sec
+epoch 2, total_step 44020, total loss is 23.19 , inference loss is 12.20, weight deacy loss is 10.99, training accuracy is 0.218750, time 218.541 samples/sec
+epoch 2, total_step 44040, total loss is 27.23 , inference loss is 16.24, weight deacy loss is 10.99, training accuracy is 0.031250, time 210.688 samples/sec
+epoch 2, total_step 44060, total loss is 24.84 , inference loss is 13.85, weight deacy loss is 10.98, training accuracy is 0.156250, time 217.179 samples/sec
+epoch 2, total_step 44080, total loss is 30.74 , inference loss is 19.75, weight deacy loss is 10.98, training accuracy is 0.093750, time 224.710 samples/sec
+epoch 2, total_step 44100, total loss is 25.08 , inference loss is 14.10, weight deacy loss is 10.98, training accuracy is 0.125000, time 14.196 samples/sec
+epoch 2, total_step 44120, total loss is 21.82 , inference loss is 10.84, weight deacy loss is 10.98, training accuracy is 0.187500, time 223.955 samples/sec
+epoch 2, total_step 44140, total loss is 24.60 , inference loss is 13.62, weight deacy loss is 10.98, training accuracy is 0.218750, time 196.226 samples/sec
+epoch 2, total_step 44160, total loss is 26.44 , inference loss is 15.46, weight deacy loss is 10.98, training accuracy is 0.156250, time 209.771 samples/sec
+epoch 2, total_step 44180, total loss is 27.05 , inference loss is 16.07, weight deacy loss is 10.98, training accuracy is 0.281250, time 213.299 samples/sec
+epoch 2, total_step 44200, total loss is 24.49 , inference loss is 13.52, weight deacy loss is 10.97, training accuracy is 0.218750, time 222.757 samples/sec
+epoch 2, total_step 44220, total loss is 33.56 , inference loss is 22.59, weight deacy loss is 10.97, training accuracy is 0.031250, time 296.417 samples/sec
+epoch 2, total_step 44240, total loss is 25.78 , inference loss is 14.81, weight deacy loss is 10.97, training accuracy is 0.125000, time 212.823 samples/sec
+epoch 2, total_step 44260, total loss is 25.74 , inference loss is 14.77, weight deacy loss is 10.97, training accuracy is 0.062500, time 218.579 samples/sec
+epoch 2, total_step 44280, total loss is 35.26 , inference loss is 24.29, weight deacy loss is 10.97, training accuracy is 0.062500, time 222.037 samples/sec
+epoch 2, total_step 44300, total loss is 23.44 , inference loss is 12.48, weight deacy loss is 10.97, training accuracy is 0.125000, time 218.194 samples/sec
+epoch 2, total_step 44320, total loss is 24.98 , inference loss is 14.02, weight deacy loss is 10.97, training accuracy is 0.093750, time 226.714 samples/sec
+epoch 2, total_step 44340, total loss is 23.63 , inference loss is 12.67, weight deacy loss is 10.97, training accuracy is 0.125000, time 228.193 samples/sec
+epoch 2, total_step 44360, total loss is 26.56 , inference loss is 15.60, weight deacy loss is 10.96, training accuracy is 0.156250, time 232.996 samples/sec
+epoch 2, total_step 44380, total loss is 24.27 , inference loss is 13.31, weight deacy loss is 10.96, training accuracy is 0.093750, time 240.099 samples/sec
+epoch 2, total_step 44400, total loss is 26.08 , inference loss is 15.12, weight deacy loss is 10.96, training accuracy is 0.156250, time 14.425 samples/sec
+epoch 2, total_step 44420, total loss is 24.44 , inference loss is 13.48, weight deacy loss is 10.96, training accuracy is 0.093750, time 214.421 samples/sec
+epoch 2, total_step 44440, total loss is 28.43 , inference loss is 17.47, weight deacy loss is 10.96, training accuracy is 0.093750, time 224.206 samples/sec
+epoch 2, total_step 44460, total loss is 26.64 , inference loss is 15.68, weight deacy loss is 10.96, training accuracy is 0.125000, time 218.358 samples/sec
+epoch 2, total_step 44480, total loss is 25.02 , inference loss is 14.06, weight deacy loss is 10.96, training accuracy is 0.187500, time 214.061 samples/sec
+epoch 2, total_step 44500, total loss is 25.67 , inference loss is 14.72, weight deacy loss is 10.95, training accuracy is 0.031250, time 197.940 samples/sec
+epoch 2, total_step 44520, total loss is 27.38 , inference loss is 16.43, weight deacy loss is 10.95, training accuracy is 0.093750, time 217.490 samples/sec
+epoch 2, total_step 44540, total loss is 29.58 , inference loss is 18.62, weight deacy loss is 10.95, training accuracy is 0.031250, time 298.159 samples/sec
+epoch 2, total_step 44560, total loss is 24.03 , inference loss is 13.08, weight deacy loss is 10.95, training accuracy is 0.156250, time 219.728 samples/sec
+epoch 2, total_step 44580, total loss is 25.98 , inference loss is 15.04, weight deacy loss is 10.95, training accuracy is 0.125000, time 207.666 samples/sec
+epoch 2, total_step 44600, total loss is 27.53 , inference loss is 16.58, weight deacy loss is 10.95, training accuracy is 0.156250, time 219.973 samples/sec
+epoch 2, total_step 44620, total loss is 25.57 , inference loss is 14.63, weight deacy loss is 10.95, training accuracy is 0.187500, time 219.650 samples/sec
+epoch 2, total_step 44640, total loss is 29.01 , inference loss is 18.07, weight deacy loss is 10.94, training accuracy is 0.062500, time 224.974 samples/sec
+epoch 2, total_step 44660, total loss is 24.77 , inference loss is 13.83, weight deacy loss is 10.94, training accuracy is 0.156250, time 292.033 samples/sec
+epoch 2, total_step 44680, total loss is 26.25 , inference loss is 15.31, weight deacy loss is 10.94, training accuracy is 0.000000, time 229.218 samples/sec
+epoch 2, total_step 44700, total loss is 23.92 , inference loss is 12.98, weight deacy loss is 10.94, training accuracy is 0.156250, time 13.188 samples/sec
+epoch 2, total_step 44720, total loss is 23.48 , inference loss is 12.55, weight deacy loss is 10.94, training accuracy is 0.156250, time 210.739 samples/sec
+epoch 2, total_step 44740, total loss is 26.35 , inference loss is 15.41, weight deacy loss is 10.94, training accuracy is 0.093750, time 224.680 samples/sec
+epoch 2, total_step 44760, total loss is 22.67 , inference loss is 11.73, weight deacy loss is 10.94, training accuracy is 0.156250, time 212.190 samples/sec
+epoch 2, total_step 44780, total loss is 29.03 , inference loss is 18.10, weight deacy loss is 10.94, training accuracy is 0.000000, time 209.297 samples/sec
+epoch 2, total_step 44800, total loss is 28.05 , inference loss is 17.12, weight deacy loss is 10.93, training accuracy is 0.156250, time 213.198 samples/sec
+epoch 2, total_step 44820, total loss is 29.92 , inference loss is 18.98, weight deacy loss is 10.93, training accuracy is 0.000000, time 212.311 samples/sec
+epoch 2, total_step 44840, total loss is 31.44 , inference loss is 20.51, weight deacy loss is 10.93, training accuracy is 0.062500, time 216.483 samples/sec
+epoch 2, total_step 44860, total loss is 25.70 , inference loss is 14.77, weight deacy loss is 10.93, training accuracy is 0.031250, time 226.648 samples/sec
+epoch 2, total_step 44880, total loss is 28.97 , inference loss is 18.04, weight deacy loss is 10.93, training accuracy is 0.093750, time 190.466 samples/sec
+epoch 2, total_step 44900, total loss is 24.52 , inference loss is 13.60, weight deacy loss is 10.93, training accuracy is 0.187500, time 214.038 samples/sec
+epoch 2, total_step 44920, total loss is 29.73 , inference loss is 18.80, weight deacy loss is 10.93, training accuracy is 0.125000, time 212.107 samples/sec
+epoch 2, total_step 44940, total loss is 28.97 , inference loss is 18.05, weight deacy loss is 10.92, training accuracy is 0.125000, time 298.705 samples/sec
+epoch 2, total_step 44960, total loss is 28.14 , inference loss is 17.21, weight deacy loss is 10.92, training accuracy is 0.125000, time 223.500 samples/sec
+epoch 2, total_step 44980, total loss is 26.63 , inference loss is 15.71, weight deacy loss is 10.92, training accuracy is 0.125000, time 220.453 samples/sec
+epoch 2, total_step 45000, total loss is 23.93 , inference loss is 13.01, weight deacy loss is 10.92, training accuracy is 0.281250, time 13.094 samples/sec
+epoch 2, total_step 45020, total loss is 24.56 , inference loss is 13.64, weight deacy loss is 10.92, training accuracy is 0.156250, time 222.383 samples/sec
+epoch 2, total_step 45040, total loss is 25.99 , inference loss is 15.07, weight deacy loss is 10.92, training accuracy is 0.125000, time 209.439 samples/sec
+epoch 2, total_step 45060, total loss is 27.24 , inference loss is 16.33, weight deacy loss is 10.92, training accuracy is 0.062500, time 209.586 samples/sec
+epoch 2, total_step 45080, total loss is 27.58 , inference loss is 16.67, weight deacy loss is 10.91, training accuracy is 0.156250, time 216.354 samples/sec
+epoch 2, total_step 45100, total loss is 28.07 , inference loss is 17.16, weight deacy loss is 10.91, training accuracy is 0.093750, time 219.030 samples/sec
+epoch 2, total_step 45120, total loss is 26.61 , inference loss is 15.70, weight deacy loss is 10.91, training accuracy is 0.062500, time 202.324 samples/sec
+epoch 2, total_step 45140, total loss is 26.15 , inference loss is 15.24, weight deacy loss is 10.91, training accuracy is 0.125000, time 207.607 samples/sec
+epoch 2, total_step 45160, total loss is 28.70 , inference loss is 17.79, weight deacy loss is 10.91, training accuracy is 0.093750, time 235.063 samples/sec
+epoch 2, total_step 45180, total loss is 21.72 , inference loss is 10.81, weight deacy loss is 10.91, training accuracy is 0.187500, time 214.309 samples/sec
+epoch 2, total_step 45200, total loss is 25.34 , inference loss is 14.43, weight deacy loss is 10.91, training accuracy is 0.062500, time 295.621 samples/sec
+epoch 2, total_step 45220, total loss is 24.76 , inference loss is 13.86, weight deacy loss is 10.91, training accuracy is 0.062500, time 207.832 samples/sec
+epoch 2, total_step 45240, total loss is 26.61 , inference loss is 15.71, weight deacy loss is 10.90, training accuracy is 0.031250, time 194.155 samples/sec
+epoch 2, total_step 45260, total loss is 29.16 , inference loss is 18.26, weight deacy loss is 10.90, training accuracy is 0.093750, time 237.971 samples/sec
+epoch 2, total_step 45280, total loss is 25.81 , inference loss is 14.91, weight deacy loss is 10.90, training accuracy is 0.125000, time 216.063 samples/sec
+epoch 2, total_step 45300, total loss is 29.33 , inference loss is 18.43, weight deacy loss is 10.90, training accuracy is 0.031250, time 13.530 samples/sec
+epoch 2, total_step 45320, total loss is 24.63 , inference loss is 13.73, weight deacy loss is 10.90, training accuracy is 0.187500, time 221.718 samples/sec
+epoch 2, total_step 45340, total loss is 25.05 , inference loss is 14.15, weight deacy loss is 10.90, training accuracy is 0.187500, time 217.925 samples/sec
+epoch 2, total_step 45360, total loss is 27.45 , inference loss is 16.56, weight deacy loss is 10.90, training accuracy is 0.031250, time 200.859 samples/sec2022-12-01 03:14:02.230640: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7376 of 10000
+2022-12-01 03:14:05.579513: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 2, total_step 45380, total loss is 24.03 , inference loss is 13.13, weight deacy loss is 10.89, training accuracy is 0.187500, time 222.727 samples/sec
+epoch 2, total_step 45400, total loss is 28.83 , inference loss is 17.94, weight deacy loss is 10.89, training accuracy is 0.093750, time 225.831 samples/sec
+epoch 2, total_step 45420, total loss is 28.38 , inference loss is 17.49, weight deacy loss is 10.89, training accuracy is 0.062500, time 210.361 samples/sec
+epoch 2, total_step 45440, total loss is 30.40 , inference loss is 19.51, weight deacy loss is 10.89, training accuracy is 0.031250, time 197.054 samples/sec
+epoch 2, total_step 45460, total loss is 22.11 , inference loss is 11.22, weight deacy loss is 10.89, training accuracy is 0.187500, time 294.617 samples/sec
+epoch 2, total_step 45480, total loss is 25.36 , inference loss is 14.47, weight deacy loss is 10.89, training accuracy is 0.062500, time 220.740 samples/sec
+epoch 2, total_step 45500, total loss is 28.52 , inference loss is 17.63, weight deacy loss is 10.89, training accuracy is 0.031250, time 212.927 samples/sec
+epoch 2, total_step 45520, total loss is 23.70 , inference loss is 12.81, weight deacy loss is 10.88, training accuracy is 0.093750, time 228.209 samples/sec
+epoch 2, total_step 45540, total loss is 25.83 , inference loss is 14.94, weight deacy loss is 10.88, training accuracy is 0.031250, time 224.012 samples/sec
+epoch 2, total_step 45560, total loss is 29.68 , inference loss is 18.80, weight deacy loss is 10.88, training accuracy is 0.093750, time 219.698 samples/sec
+epoch 2, total_step 45580, total loss is 28.76 , inference loss is 17.88, weight deacy loss is 10.88, training accuracy is 0.062500, time 214.306 samples/sec
+epoch 2, total_step 45600, total loss is 25.75 , inference loss is 14.87, weight deacy loss is 10.88, training accuracy is 0.062500, time 13.560 samples/sec
+epoch 2, total_step 45620, total loss is 30.91 , inference loss is 20.03, weight deacy loss is 10.88, training accuracy is 0.062500, time 214.518 samples/sec
+epoch 2, total_step 45640, total loss is 25.64 , inference loss is 14.76, weight deacy loss is 10.88, training accuracy is 0.093750, time 221.252 samples/sec
+epoch 2, total_step 45660, total loss is 28.23 , inference loss is 17.35, weight deacy loss is 10.88, training accuracy is 0.125000, time 190.292 samples/sec
+epoch 2, total_step 45680, total loss is 26.08 , inference loss is 15.21, weight deacy loss is 10.87, training accuracy is 0.093750, time 222.940 samples/sec
+epoch 2, total_step 45700, total loss is 25.94 , inference loss is 15.07, weight deacy loss is 10.87, training accuracy is 0.125000, time 202.088 samples/sec
+epoch 2, total_step 45720, total loss is 28.47 , inference loss is 17.60, weight deacy loss is 10.87, training accuracy is 0.062500, time 299.953 samples/sec
+epoch 2, total_step 45740, total loss is 24.46 , inference loss is 13.59, weight deacy loss is 10.87, training accuracy is 0.093750, time 176.966 samples/sec
+epoch 2, total_step 45760, total loss is 25.66 , inference loss is 14.79, weight deacy loss is 10.87, training accuracy is 0.187500, time 191.996 samples/sec
+epoch 2, total_step 45780, total loss is 28.40 , inference loss is 17.54, weight deacy loss is 10.87, training accuracy is 0.093750, time 192.718 samples/sec
+epoch 2, total_step 45800, total loss is 21.97 , inference loss is 11.11, weight deacy loss is 10.87, training accuracy is 0.093750, time 193.492 samples/sec
+epoch 2, total_step 45820, total loss is 29.86 , inference loss is 18.99, weight deacy loss is 10.86, training accuracy is 0.000000, time 202.951 samples/sec
+epoch 2, total_step 45840, total loss is 26.99 , inference loss is 16.12, weight deacy loss is 10.86, training accuracy is 0.093750, time 197.187 samples/sec
+epoch 2, total_step 45860, total loss is 27.77 , inference loss is 16.91, weight deacy loss is 10.86, training accuracy is 0.062500, time 220.406 samples/sec
+epoch 2, total_step 45880, total loss is 24.91 , inference loss is 14.05, weight deacy loss is 10.86, training accuracy is 0.156250, time 216.407 samples/sec
+epoch 2, total_step 45900, total loss is 27.63 , inference loss is 16.77, weight deacy loss is 10.86, training accuracy is 0.125000, time 12.206 samples/sec
+epoch 2, total_step 45920, total loss is 25.13 , inference loss is 14.27, weight deacy loss is 10.86, training accuracy is 0.125000, time 179.046 samples/sec
+epoch 2, total_step 45940, total loss is 28.34 , inference loss is 17.48, weight deacy loss is 10.86, training accuracy is 0.093750, time 203.183 samples/sec
+epoch 2, total_step 45960, total loss is 25.52 , inference loss is 14.67, weight deacy loss is 10.86, training accuracy is 0.156250, time 192.430 samples/sec
+epoch 2, total_step 45980, total loss is 25.47 , inference loss is 14.62, weight deacy loss is 10.85, training accuracy is 0.093750, time 216.788 samples/sec
+End of epoch 2
+testing verification..
+(12000, 512)
+infer time 25.18361599999998
+best_threshold_index 121 0.9805555555555555
+best_threshold_index 121 0.98
+best_threshold_index 121 0.9807407407407407
+best_threshold_index 121 0.980925925925926
+best_threshold_index 121 0.9803703703703703
+best_threshold_index 121 0.9792592592592593
+best_threshold_index 121 0.9801851851851852
+best_threshold_index 121 0.98
+best_threshold_index 121 0.9785185185185186
+best_threshold_index 121 0.9794444444444445
+[lfw][46000]XNorm: 22.070035
+[lfw][46000]Accuracy-Flip: 0.98000+-0.00632
+test accuracy is: 0.9800000000000001
+epoch 3, total_step 46000, total loss is 39.35 , inference loss is 28.50, weight deacy loss is 10.85, training accuracy is 0.000000, time 17.950 samples/sec
+epoch 3, total_step 46020, total loss is 27.54 , inference loss is 16.68, weight deacy loss is 10.85, training accuracy is 0.093750, time 219.626 samples/sec
+epoch 3, total_step 46040, total loss is 19.72 , inference loss is 8.87, weight deacy loss is 10.85, training accuracy is 0.312500, time 202.468 samples/sec
+epoch 3, total_step 46060, total loss is 25.30 , inference loss is 14.45, weight deacy loss is 10.85, training accuracy is 0.187500, time 296.915 samples/sec
+epoch 3, total_step 46080, total loss is 21.36 , inference loss is 10.51, weight deacy loss is 10.85, training accuracy is 0.281250, time 224.337 samples/sec
+epoch 3, total_step 46100, total loss is 28.54 , inference loss is 17.70, weight deacy loss is 10.85, training accuracy is 0.156250, time 227.302 samples/sec
+epoch 3, total_step 46120, total loss is 20.92 , inference loss is 10.07, weight deacy loss is 10.84, training accuracy is 0.312500, time 221.655 samples/sec
+epoch 3, total_step 46140, total loss is 19.73 , inference loss is 8.89, weight deacy loss is 10.84, training accuracy is 0.468750, time 214.535 samples/sec
+epoch 3, total_step 46160, total loss is 22.27 , inference loss is 11.43, weight deacy loss is 10.84, training accuracy is 0.343750, time 213.008 samples/sec
+epoch 3, total_step 46180, total loss is 20.61 , inference loss is 9.76, weight deacy loss is 10.84, training accuracy is 0.500000, time 208.679 samples/sec
+epoch 3, total_step 46200, total loss is 23.73 , inference loss is 12.89, weight deacy loss is 10.84, training accuracy is 0.343750, time 13.512 samples/sec
+epoch 3, total_step 46220, total loss is 19.05 , inference loss is 8.21, weight deacy loss is 10.84, training accuracy is 0.531250, time 214.573 samples/sec
+epoch 3, total_step 46240, total loss is 22.10 , inference loss is 11.26, weight deacy loss is 10.84, training accuracy is 0.281250, time 215.024 samples/sec
+epoch 3, total_step 46260, total loss is 19.70 , inference loss is 8.86, weight deacy loss is 10.83, training accuracy is 0.437500, time 221.055 samples/sec
+epoch 3, total_step 46280, total loss is 22.57 , inference loss is 11.73, weight deacy loss is 10.83, training accuracy is 0.250000, time 298.344 samples/sec
+epoch 3, total_step 46300, total loss is 25.63 , inference loss is 14.80, weight deacy loss is 10.83, training accuracy is 0.218750, time 299.886 samples/sec
+epoch 3, total_step 46320, total loss is 22.07 , inference loss is 11.24, weight deacy loss is 10.83, training accuracy is 0.437500, time 209.391 samples/sec
+epoch 3, total_step 46340, total loss is 22.93 , inference loss is 12.10, weight deacy loss is 10.83, training accuracy is 0.281250, time 219.490 samples/sec
+epoch 3, total_step 46360, total loss is 21.42 , inference loss is 10.59, weight deacy loss is 10.83, training accuracy is 0.312500, time 219.972 samples/sec
+epoch 3, total_step 46380, total loss is 22.62 , inference loss is 11.80, weight deacy loss is 10.83, training accuracy is 0.250000, time 216.340 samples/sec
+epoch 3, total_step 46400, total loss is 22.90 , inference loss is 12.07, weight deacy loss is 10.82, training accuracy is 0.125000, time 204.805 samples/sec
+epoch 3, total_step 46420, total loss is 23.93 , inference loss is 13.11, weight deacy loss is 10.82, training accuracy is 0.312500, time 202.134 samples/sec
+epoch 3, total_step 46440, total loss is 21.70 , inference loss is 10.88, weight deacy loss is 10.82, training accuracy is 0.218750, time 202.617 samples/sec
+epoch 3, total_step 46460, total loss is 23.01 , inference loss is 12.19, weight deacy loss is 10.82, training accuracy is 0.375000, time 214.553 samples/sec
+epoch 3, total_step 46480, total loss is 25.30 , inference loss is 14.48, weight deacy loss is 10.82, training accuracy is 0.218750, time 217.921 samples/sec
+epoch 3, total_step 46500, total loss is 22.42 , inference loss is 11.60, weight deacy loss is 10.82, training accuracy is 0.343750, time 13.341 samples/sec
+epoch 3, total_step 46520, total loss is 23.76 , inference loss is 12.94, weight deacy loss is 10.82, training accuracy is 0.343750, time 226.119 samples/sec
+epoch 3, total_step 46540, total loss is 22.72 , inference loss is 11.91, weight deacy loss is 10.82, training accuracy is 0.187500, time 215.411 samples/sec
+epoch 3, total_step 46560, total loss is 20.85 , inference loss is 10.04, weight deacy loss is 10.81, training accuracy is 0.375000, time 218.335 samples/sec
+epoch 3, total_step 46580, total loss is 22.41 , inference loss is 11.60, weight deacy loss is 10.81, training accuracy is 0.343750, time 214.805 samples/sec
+epoch 3, total_step 46600, total loss is 21.44 , inference loss is 10.63, weight deacy loss is 10.81, training accuracy is 0.312500, time 204.870 samples/sec
+epoch 3, total_step 46620, total loss is 21.06 , inference loss is 10.25, weight deacy loss is 10.81, training accuracy is 0.343750, time 221.373 samples/sec
+epoch 3, total_step 46640, total loss is 21.39 , inference loss is 10.58, weight deacy loss is 10.81, training accuracy is 0.281250, time 220.361 samples/sec
+epoch 3, total_step 46660, total loss is 21.89 , inference loss is 11.08, weight deacy loss is 10.81, training accuracy is 0.312500, time 211.268 samples/sec
+epoch 3, total_step 46680, total loss is 18.73 , inference loss is 7.93, weight deacy loss is 10.81, training accuracy is 0.375000, time 220.388 samples/sec
+epoch 3, total_step 46700, total loss is 20.99 , inference loss is 10.18, weight deacy loss is 10.80, training accuracy is 0.343750, time 225.501 samples/sec
+epoch 3, total_step 46720, total loss is 24.47 , inference loss is 13.67, weight deacy loss is 10.80, training accuracy is 0.281250, time 215.475 samples/sec
+epoch 3, total_step 46740, total loss is 25.21 , inference loss is 14.41, weight deacy loss is 10.80, training accuracy is 0.218750, time 198.949 samples/sec
+epoch 3, total_step 46760, total loss is 22.61 , inference loss is 11.81, weight deacy loss is 10.80, training accuracy is 0.156250, time 213.971 samples/sec
+epoch 3, total_step 46780, total loss is 26.45 , inference loss is 15.66, weight deacy loss is 10.80, training accuracy is 0.250000, time 296.382 samples/sec
+epoch 3, total_step 46800, total loss is 25.68 , inference loss is 14.88, weight deacy loss is 10.80, training accuracy is 0.156250, time 14.318 samples/sec
+epoch 3, total_step 46820, total loss is 25.97 , inference loss is 15.18, weight deacy loss is 10.80, training accuracy is 0.250000, time 234.163 samples/sec
+epoch 3, total_step 46840, total loss is 27.28 , inference loss is 16.49, weight deacy loss is 10.79, training accuracy is 0.218750, time 178.095 samples/sec
+epoch 3, total_step 46860, total loss is 22.79 , inference loss is 12.00, weight deacy loss is 10.79, training accuracy is 0.281250, time 176.282 samples/sec
+epoch 3, total_step 46880, total loss is 23.28 , inference loss is 12.49, weight deacy loss is 10.79, training accuracy is 0.250000, time 182.232 samples/sec
+epoch 3, total_step 46900, total loss is 23.42 , inference loss is 12.63, weight deacy loss is 10.79, training accuracy is 0.218750, time 181.340 samples/sec
+epoch 3, total_step 46920, total loss is 22.33 , inference loss is 11.54, weight deacy loss is 10.79, training accuracy is 0.250000, time 173.885 samples/sec
+epoch 3, total_step 46940, total loss is 25.41 , inference loss is 14.62, weight deacy loss is 10.79, training accuracy is 0.156250, time 177.261 samples/sec
+epoch 3, total_step 46960, total loss is 22.06 , inference loss is 11.28, weight deacy loss is 10.79, training accuracy is 0.218750, time 178.571 samples/sec
+epoch 3, total_step 46980, total loss is 21.03 , inference loss is 10.24, weight deacy loss is 10.79, training accuracy is 0.343750, time 297.370 samples/sec
+epoch 3, total_step 47000, total loss is 19.63 , inference loss is 8.84, weight deacy loss is 10.78, training accuracy is 0.281250, time 180.412 samples/sec
+epoch 3, total_step 47020, total loss is 25.92 , inference loss is 15.14, weight deacy loss is 10.78, training accuracy is 0.156250, time 181.373 samples/sec
+epoch 3, total_step 47040, total loss is 25.90 , inference loss is 15.12, weight deacy loss is 10.78, training accuracy is 0.187500, time 178.466 samples/sec
+epoch 3, total_step 47060, total loss is 23.79 , inference loss is 13.01, weight deacy loss is 10.78, training accuracy is 0.250000, time 183.395 samples/sec
+epoch 3, total_step 47080, total loss is 27.93 , inference loss is 17.15, weight deacy loss is 10.78, training accuracy is 0.218750, time 276.556 samples/sec
+epoch 3, total_step 47100, total loss is 24.06 , inference loss is 13.28, weight deacy loss is 10.78, training accuracy is 0.187500, time 13.605 samples/sec
+epoch 3, total_step 47120, total loss is 25.08 , inference loss is 14.30, weight deacy loss is 10.78, training accuracy is 0.156250, time 212.329 samples/sec
+epoch 3, total_step 47140, total loss is 24.55 , inference loss is 13.77, weight deacy loss is 10.77, training accuracy is 0.312500, time 222.882 samples/sec
+epoch 3, total_step 47160, total loss is 28.54 , inference loss is 17.77, weight deacy loss is 10.77, training accuracy is 0.125000, time 217.919 samples/sec
+epoch 3, total_step 47180, total loss is 23.17 , inference loss is 12.40, weight deacy loss is 10.77, training accuracy is 0.250000, time 210.887 samples/sec
+epoch 3, total_step 47200, total loss is 20.65 , inference loss is 9.88, weight deacy loss is 10.77, training accuracy is 0.187500, time 220.215 samples/sec
+epoch 3, total_step 47220, total loss is 27.40 , inference loss is 16.63, weight deacy loss is 10.77, training accuracy is 0.093750, time 213.200 samples/sec
+epoch 3, total_step 47240, total loss is 29.06 , inference loss is 18.30, weight deacy loss is 10.77, training accuracy is 0.250000, time 210.300 samples/sec
+epoch 3, total_step 47260, total loss is 23.57 , inference loss is 12.80, weight deacy loss is 10.77, training accuracy is 0.125000, time 211.895 samples/sec
+epoch 3, total_step 47280, total loss is 21.43 , inference loss is 10.67, weight deacy loss is 10.77, training accuracy is 0.156250, time 237.430 samples/sec
+epoch 3, total_step 47300, total loss is 27.49 , inference loss is 16.72, weight deacy loss is 10.76, training accuracy is 0.125000, time 298.920 samples/sec
+epoch 3, total_step 47320, total loss is 21.72 , inference loss is 10.96, weight deacy loss is 10.76, training accuracy is 0.312500, time 217.991 samples/sec
+epoch 3, total_step 47340, total loss is 27.70 , inference loss is 16.94, weight deacy loss is 10.76, training accuracy is 0.187500, time 210.672 samples/sec
+epoch 3, total_step 47360, total loss is 25.82 , inference loss is 15.06, weight deacy loss is 10.76, training accuracy is 0.218750, time 204.222 samples/sec
+epoch 3, total_step 47380, total loss is 23.45 , inference loss is 12.69, weight deacy loss is 10.76, training accuracy is 0.281250, time 208.792 samples/sec
+epoch 3, total_step 47400, total loss is 26.98 , inference loss is 16.22, weight deacy loss is 10.76, training accuracy is 0.281250, time 11.823 samples/sec
+epoch 3, total_step 47420, total loss is 26.97 , inference loss is 16.21, weight deacy loss is 10.76, training accuracy is 0.156250, time 217.473 samples/sec
+epoch 3, total_step 47440, total loss is 24.40 , inference loss is 13.64, weight deacy loss is 10.75, training accuracy is 0.218750, time 216.624 samples/sec
+epoch 3, total_step 47460, total loss is 24.00 , inference loss is 13.25, weight deacy loss is 10.75, training accuracy is 0.250000, time 211.917 samples/sec
+epoch 3, total_step 47480, total loss is 24.82 , inference loss is 14.06, weight deacy loss is 10.75, training accuracy is 0.281250, time 230.474 samples/sec
+epoch 3, total_step 47500, total loss is 22.73 , inference loss is 11.98, weight deacy loss is 10.75, training accuracy is 0.125000, time 214.003 samples/sec
+epoch 3, total_step 47520, total loss is 23.25 , inference loss is 12.50, weight deacy loss is 10.75, training accuracy is 0.218750, time 210.233 samples/sec
+epoch 3, total_step 47540, total loss is 23.98 , inference loss is 13.23, weight deacy loss is 10.75, training accuracy is 0.281250, time 207.845 samples/sec
+epoch 3, total_step 47560, total loss is 23.19 , inference loss is 12.44, weight deacy loss is 10.75, training accuracy is 0.250000, time 297.933 samples/sec
+epoch 3, total_step 47580, total loss is 22.80 , inference loss is 12.06, weight deacy loss is 10.74, training accuracy is 0.281250, time 211.455 samples/sec
+epoch 3, total_step 47600, total loss is 21.34 , inference loss is 10.60, weight deacy loss is 10.74, training accuracy is 0.250000, time 207.990 samples/sec
+epoch 3, total_step 47620, total loss is 22.27 , inference loss is 11.53, weight deacy loss is 10.74, training accuracy is 0.218750, time 213.133 samples/sec
+epoch 3, total_step 47640, total loss is 22.25 , inference loss is 11.51, weight deacy loss is 10.74, training accuracy is 0.250000, time 197.662 samples/sec
+epoch 3, total_step 47660, total loss is 26.55 , inference loss is 15.82, weight deacy loss is 10.74, training accuracy is 0.125000, time 208.451 samples/sec
+epoch 3, total_step 47680, total loss is 21.71 , inference loss is 10.97, weight deacy loss is 10.74, training accuracy is 0.187500, time 213.380 samples/sec
+epoch 3, total_step 47700, total loss is 20.67 , inference loss is 9.93, weight deacy loss is 10.74, training accuracy is 0.250000, time 13.479 samples/sec
+epoch 3, total_step 47720, total loss is 25.56 , inference loss is 14.82, weight deacy loss is 10.74, training accuracy is 0.187500, time 212.814 samples/sec
+epoch 3, total_step 47740, total loss is 24.02 , inference loss is 13.29, weight deacy loss is 10.73, training accuracy is 0.250000, time 207.641 samples/sec
+epoch 3, total_step 47760, total loss is 22.67 , inference loss is 11.94, weight deacy loss is 10.73, training accuracy is 0.218750, time 209.046 samples/sec
+epoch 3, total_step 47780, total loss is 24.34 , inference loss is 13.61, weight deacy loss is 10.73, training accuracy is 0.281250, time 217.536 samples/sec
+epoch 3, total_step 47800, total loss is 22.80 , inference loss is 12.07, weight deacy loss is 10.73, training accuracy is 0.250000, time 213.051 samples/sec
+epoch 3, total_step 47820, total loss is 22.28 , inference loss is 11.56, weight deacy loss is 10.73, training accuracy is 0.250000, time 211.640 samples/sec
+epoch 3, total_step 47840, total loss is 21.77 , inference loss is 11.05, weight deacy loss is 10.73, training accuracy is 0.250000, time 296.947 samples/sec
+epoch 3, total_step 47860, total loss is 21.59 , inference loss is 10.86, weight deacy loss is 10.73, training accuracy is 0.312500, time 216.794 samples/sec
+epoch 3, total_step 47880, total loss is 24.18 , inference loss is 13.46, weight deacy loss is 10.72, training accuracy is 0.312500, time 201.120 samples/sec
+epoch 3, total_step 47900, total loss is 25.67 , inference loss is 14.95, weight deacy loss is 10.72, training accuracy is 0.125000, time 227.271 samples/sec
+epoch 3, total_step 47920, total loss is 25.81 , inference loss is 15.09, weight deacy loss is 10.72, training accuracy is 0.187500, time 204.229 samples/sec
+epoch 3, total_step 47940, total loss is 22.39 , inference loss is 11.67, weight deacy loss is 10.72, training accuracy is 0.281250, time 211.153 samples/sec
+epoch 3, total_step 47960, total loss is 18.64 , inference loss is 7.93, weight deacy loss is 10.72, training accuracy is 0.312500, time 210.089 samples/sec
+epoch 3, total_step 47980, total loss is 25.70 , inference loss is 14.98, weight deacy loss is 10.72, training accuracy is 0.250000, time 212.279 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.176746999999988
+best_threshold_index 111 0.9833333333333333
+best_threshold_index 111 0.9825925925925926
+best_threshold_index 111 0.9827777777777778
+best_threshold_index 111 0.9825925925925926
+best_threshold_index 111 0.9833333333333333
+best_threshold_index 110 0.9814814814814815
+best_threshold_index 111 0.9822222222222222
+best_threshold_index 111 0.9818518518518519
+best_threshold_index 111 0.9818518518518519
+best_threshold_index 111 0.9812962962962963
+[lfw][48000]XNorm: 22.560015
+[lfw][48000]Accuracy-Flip: 0.98183+-0.00565
+test accuracy is: 0.9818333333333336
+epoch 3, total_step 48000, total loss is 22.73 , inference loss is 12.01, weight deacy loss is 10.72, training accuracy is 0.312500, time 16.951 samples/sec
+epoch 3, total_step 48020, total loss is 24.41 , inference loss is 13.70, weight deacy loss is 10.71, training accuracy is 0.156250, time 211.980 samples/sec
+epoch 3, total_step 48040, total loss is 33.99 , inference loss is 23.28, weight deacy loss is 10.71, training accuracy is 0.125000, time 209.598 samples/sec
+epoch 3, total_step 48060, total loss is 24.67 , inference loss is 13.96, weight deacy loss is 10.71, training accuracy is 0.250000, time 214.670 samples/sec
+epoch 3, total_step 48080, total loss is 24.65 , inference loss is 13.94, weight deacy loss is 10.71, training accuracy is 0.312500, time 299.024 samples/sec
+epoch 3, total_step 48100, total loss is 20.30 , inference loss is 9.59, weight deacy loss is 10.71, training accuracy is 0.312500, time 297.886 samples/sec
+epoch 3, total_step 48120, total loss is 21.50 , inference loss is 10.79, weight deacy loss is 10.71, training accuracy is 0.406250, time 196.061 samples/sec
+epoch 3, total_step 48140, total loss is 20.95 , inference loss is 10.25, weight deacy loss is 10.71, training accuracy is 0.437500, time 216.662 samples/sec
+epoch 3, total_step 48160, total loss is 25.20 , inference loss is 14.50, weight deacy loss is 10.70, training accuracy is 0.281250, time 206.926 samples/sec
+epoch 3, total_step 48180, total loss is 20.49 , inference loss is 9.78, weight deacy loss is 10.70, training accuracy is 0.093750, time 216.584 samples/sec
+epoch 3, total_step 48200, total loss is 21.87 , inference loss is 11.17, weight deacy loss is 10.70, training accuracy is 0.156250, time 231.564 samples/sec
+epoch 3, total_step 48220, total loss is 22.45 , inference loss is 11.75, weight deacy loss is 10.70, training accuracy is 0.406250, time 215.680 samples/sec
+epoch 3, total_step 48240, total loss is 21.70 , inference loss is 11.00, weight deacy loss is 10.70, training accuracy is 0.312500, time 215.571 samples/sec
+epoch 3, total_step 48260, total loss is 21.08 , inference loss is 10.39, weight deacy loss is 10.70, training accuracy is 0.406250, time 216.048 samples/sec
+epoch 3, total_step 48280, total loss is 21.39 , inference loss is 10.70, weight deacy loss is 10.70, training accuracy is 0.343750, time 213.567 samples/sec
+epoch 3, total_step 48300, total loss is 22.46 , inference loss is 11.76, weight deacy loss is 10.69, training accuracy is 0.281250, time 14.431 samples/sec
+epoch 3, total_step 48320, total loss is 16.33 , inference loss is 5.64, weight deacy loss is 10.69, training accuracy is 0.500000, time 225.629 samples/sec
+epoch 3, total_step 48340, total loss is 20.35 , inference loss is 9.66, weight deacy loss is 10.69, training accuracy is 0.468750, time 220.835 samples/sec
+epoch 3, total_step 48360, total loss is 27.02 , inference loss is 16.33, weight deacy loss is 10.69, training accuracy is 0.250000, time 216.779 samples/sec
+epoch 3, total_step 48380, total loss is 19.89 , inference loss is 9.20, weight deacy loss is 10.69, training accuracy is 0.312500, time 218.872 samples/sec
+epoch 3, total_step 48400, total loss is 20.08 , inference loss is 9.40, weight deacy loss is 10.69, training accuracy is 0.375000, time 209.299 samples/sec
+epoch 3, total_step 48420, total loss is 17.63 , inference loss is 6.95, weight deacy loss is 10.69, training accuracy is 0.375000, time 200.574 samples/sec
+epoch 3, total_step 48440, total loss is 20.63 , inference loss is 9.94, weight deacy loss is 10.68, training accuracy is 0.312500, time 211.830 samples/sec
+epoch 3, total_step 48460, total loss is 19.09 , inference loss is 8.40, weight deacy loss is 10.68, training accuracy is 0.343750, time 215.704 samples/sec
+epoch 3, total_step 48480, total loss is 21.51 , inference loss is 10.82, weight deacy loss is 10.68, training accuracy is 0.343750, time 209.504 samples/sec
+epoch 3, total_step 48500, total loss is 23.20 , inference loss is 12.52, weight deacy loss is 10.68, training accuracy is 0.250000, time 209.874 samples/sec
+epoch 3, total_step 48520, total loss is 20.83 , inference loss is 10.15, weight deacy loss is 10.68, training accuracy is 0.343750, time 217.621 samples/sec
+epoch 3, total_step 48540, total loss is 18.86 , inference loss is 8.18, weight deacy loss is 10.68, training accuracy is 0.250000, time 209.824 samples/sec
+epoch 3, total_step 48560, total loss is 20.56 , inference loss is 9.88, weight deacy loss is 10.68, training accuracy is 0.375000, time 227.749 samples/sec
+epoch 3, total_step 48580, total loss is 20.58 , inference loss is 9.90, weight deacy loss is 10.67, training accuracy is 0.218750, time 210.064 samples/sec
+epoch 3, total_step 48600, total loss is 20.99 , inference loss is 10.32, weight deacy loss is 10.67, training accuracy is 0.281250, time 13.957 samples/sec
+epoch 3, total_step 48620, total loss is 21.79 , inference loss is 11.12, weight deacy loss is 10.67, training accuracy is 0.343750, time 219.625 samples/sec
+epoch 3, total_step 48640, total loss is 23.09 , inference loss is 12.42, weight deacy loss is 10.67, training accuracy is 0.156250, time 297.294 samples/sec
+epoch 3, total_step 48660, total loss is 22.19 , inference loss is 11.52, weight deacy loss is 10.67, training accuracy is 0.156250, time 208.465 samples/sec
+epoch 3, total_step 48680, total loss is 17.91 , inference loss is 7.24, weight deacy loss is 10.67, training accuracy is 0.500000, time 225.519 samples/sec
+epoch 3, total_step 48700, total loss is 17.18 , inference loss is 6.51, weight deacy loss is 10.67, training accuracy is 0.437500, time 227.215 samples/sec
+epoch 3, total_step 48720, total loss is 23.36 , inference loss is 12.70, weight deacy loss is 10.67, training accuracy is 0.281250, time 220.754 samples/sec
+epoch 3, total_step 48740, total loss is 20.24 , inference loss is 9.57, weight deacy loss is 10.66, training accuracy is 0.343750, time 214.259 samples/sec
+epoch 3, total_step 48760, total loss is 19.85 , inference loss is 9.19, weight deacy loss is 10.66, training accuracy is 0.406250, time 211.276 samples/sec
+epoch 3, total_step 48780, total loss is 21.88 , inference loss is 11.22, weight deacy loss is 10.66, training accuracy is 0.156250, time 225.451 samples/sec
+epoch 3, total_step 48800, total loss is 23.43 , inference loss is 12.77, weight deacy loss is 10.66, training accuracy is 0.218750, time 210.763 samples/sec
+epoch 3, total_step 48820, total loss is 16.73 , inference loss is 6.07, weight deacy loss is 10.66, training accuracy is 0.375000, time 213.043 samples/sec
+epoch 3, total_step 48840, total loss is 19.84 , inference loss is 9.18, weight deacy loss is 10.66, training accuracy is 0.250000, time 213.897 samples/sec
+epoch 3, total_step 48860, total loss is 21.96 , inference loss is 11.31, weight deacy loss is 10.66, training accuracy is 0.343750, time 217.361 samples/sec
+epoch 3, total_step 48880, total loss is 21.10 , inference loss is 10.44, weight deacy loss is 10.65, training accuracy is 0.250000, time 224.918 samples/sec
+epoch 3, total_step 48900, total loss is 22.27 , inference loss is 11.62, weight deacy loss is 10.65, training accuracy is 0.406250, time 13.949 samples/sec
+epoch 3, total_step 48920, total loss is 19.61 , inference loss is 8.96, weight deacy loss is 10.65, training accuracy is 0.406250, time 215.068 samples/sec
+epoch 3, total_step 48940, total loss is 17.46 , inference loss is 6.81, weight deacy loss is 10.65, training accuracy is 0.437500, time 214.120 samples/sec
+epoch 3, total_step 48960, total loss is 20.23 , inference loss is 9.58, weight deacy loss is 10.65, training accuracy is 0.312500, time 202.809 samples/sec
+epoch 3, total_step 48980, total loss is 19.66 , inference loss is 9.01, weight deacy loss is 10.65, training accuracy is 0.500000, time 208.384 samples/sec
+epoch 3, total_step 49000, total loss is 23.22 , inference loss is 12.57, weight deacy loss is 10.65, training accuracy is 0.375000, time 219.095 samples/sec
+epoch 3, total_step 49020, total loss is 24.07 , inference loss is 13.43, weight deacy loss is 10.64, training accuracy is 0.281250, time 215.054 samples/sec
+epoch 3, total_step 49040, total loss is 19.48 , inference loss is 8.84, weight deacy loss is 10.64, training accuracy is 0.343750, time 210.030 samples/sec
+epoch 3, total_step 49060, total loss is 23.30 , inference loss is 12.66, weight deacy loss is 10.64, training accuracy is 0.250000, time 209.649 samples/sec
+epoch 3, total_step 49080, total loss is 20.85 , inference loss is 10.21, weight deacy loss is 10.64, training accuracy is 0.281250, time 212.392 samples/sec
+epoch 3, total_step 49100, total loss is 21.61 , inference loss is 10.97, weight deacy loss is 10.64, training accuracy is 0.343750, time 207.631 samples/sec
+epoch 3, total_step 49120, total loss is 20.71 , inference loss is 10.07, weight deacy loss is 10.64, training accuracy is 0.406250, time 214.056 samples/sec
+epoch 3, total_step 49140, total loss is 22.76 , inference loss is 12.12, weight deacy loss is 10.64, training accuracy is 0.375000, time 207.937 samples/sec
+epoch 3, total_step 49160, total loss is 22.39 , inference loss is 11.76, weight deacy loss is 10.63, training accuracy is 0.187500, time 212.550 samples/sec
+epoch 3, total_step 49180, total loss is 23.76 , inference loss is 13.12, weight deacy loss is 10.63, training accuracy is 0.312500, time 204.577 samples/sec
+epoch 3, total_step 49200, total loss is 21.49 , inference loss is 10.86, weight deacy loss is 10.63, training accuracy is 0.250000, time 13.744 samples/sec
+epoch 3, total_step 49220, total loss is 22.01 , inference loss is 11.38, weight deacy loss is 10.63, training accuracy is 0.312500, time 205.898 samples/sec
+epoch 3, total_step 49240, total loss is 28.06 , inference loss is 17.43, weight deacy loss is 10.63, training accuracy is 0.125000, time 214.881 samples/sec
+epoch 3, total_step 49260, total loss is 25.01 , inference loss is 14.38, weight deacy loss is 10.63, training accuracy is 0.250000, time 214.927 samples/sec
+epoch 3, total_step 49280, total loss is 25.86 , inference loss is 15.23, weight deacy loss is 10.63, training accuracy is 0.312500, time 211.568 samples/sec
+epoch 3, total_step 49300, total loss is 17.82 , inference loss is 7.19, weight deacy loss is 10.63, training accuracy is 0.156250, time 210.970 samples/sec
+epoch 3, total_step 49320, total loss is 21.20 , inference loss is 10.58, weight deacy loss is 10.62, training accuracy is 0.312500, time 221.083 samples/sec
+epoch 3, total_step 49340, total loss is 21.52 , inference loss is 10.90, weight deacy loss is 10.62, training accuracy is 0.281250, time 218.328 samples/sec
+epoch 3, total_step 49360, total loss is 23.66 , inference loss is 13.04, weight deacy loss is 10.62, training accuracy is 0.312500, time 213.002 samples/sec
+epoch 3, total_step 49380, total loss is 23.30 , inference loss is 12.68, weight deacy loss is 10.62, training accuracy is 0.187500, time 212.906 samples/sec
+epoch 3, total_step 49400, total loss is 24.68 , inference loss is 14.06, weight deacy loss is 10.62, training accuracy is 0.218750, time 224.014 samples/sec
+epoch 3, total_step 49420, total loss is 24.18 , inference loss is 13.56, weight deacy loss is 10.62, training accuracy is 0.437500, time 206.958 samples/sec
+epoch 3, total_step 49440, total loss is 19.02 , inference loss is 8.41, weight deacy loss is 10.62, training accuracy is 0.406250, time 208.968 samples/sec
+epoch 3, total_step 49460, total loss is 25.00 , inference loss is 14.38, weight deacy loss is 10.61, training accuracy is 0.218750, time 209.508 samples/sec
+epoch 3, total_step 49480, total loss is 24.39 , inference loss is 13.77, weight deacy loss is 10.61, training accuracy is 0.187500, time 183.822 samples/sec
+epoch 3, total_step 49500, total loss is 20.06 , inference loss is 9.44, weight deacy loss is 10.61, training accuracy is 0.218750, time 13.980 samples/sec
+epoch 3, total_step 49520, total loss is 22.94 , inference loss is 12.33, weight deacy loss is 10.61, training accuracy is 0.343750, time 209.626 samples/sec
+epoch 3, total_step 49540, total loss is 22.68 , inference loss is 12.08, weight deacy loss is 10.61, training accuracy is 0.250000, time 208.811 samples/sec
+epoch 3, total_step 49560, total loss is 26.15 , inference loss is 15.54, weight deacy loss is 10.61, training accuracy is 0.125000, time 211.578 samples/sec
+epoch 3, total_step 49580, total loss is 20.86 , inference loss is 10.26, weight deacy loss is 10.61, training accuracy is 0.343750, time 193.425 samples/sec
+epoch 3, total_step 49600, total loss is 22.56 , inference loss is 11.96, weight deacy loss is 10.61, training accuracy is 0.250000, time 208.204 samples/sec
+epoch 3, total_step 49620, total loss is 25.77 , inference loss is 15.17, weight deacy loss is 10.60, training accuracy is 0.156250, time 203.061 samples/sec
+epoch 3, total_step 49640, total loss is 18.41 , inference loss is 7.81, weight deacy loss is 10.60, training accuracy is 0.312500, time 211.847 samples/sec
+epoch 3, total_step 49660, total loss is 22.87 , inference loss is 12.27, weight deacy loss is 10.60, training accuracy is 0.218750, time 223.893 samples/sec
+epoch 3, total_step 49680, total loss is 19.92 , inference loss is 9.32, weight deacy loss is 10.60, training accuracy is 0.218750, time 204.305 samples/sec
+epoch 3, total_step 49700, total loss is 25.04 , inference loss is 14.44, weight deacy loss is 10.60, training accuracy is 0.250000, time 207.637 samples/sec
+epoch 3, total_step 49720, total loss is 21.10 , inference loss is 10.51, weight deacy loss is 10.60, training accuracy is 0.312500, time 221.095 samples/sec
+epoch 3, total_step 49740, total loss is 28.05 , inference loss is 17.46, weight deacy loss is 10.60, training accuracy is 0.125000, time 210.178 samples/sec
+epoch 3, total_step 49760, total loss is 26.41 , inference loss is 15.81, weight deacy loss is 10.59, training accuracy is 0.375000, time 221.771 samples/sec
+epoch 3, total_step 49780, total loss is 20.43 , inference loss is 9.83, weight deacy loss is 10.59, training accuracy is 0.218750, time 210.459 samples/sec
+epoch 3, total_step 49800, total loss is 23.95 , inference loss is 13.36, weight deacy loss is 10.59, training accuracy is 0.312500, time 14.155 samples/sec
+epoch 3, total_step 49820, total loss is 22.18 , inference loss is 11.59, weight deacy loss is 10.59, training accuracy is 0.281250, time 197.611 samples/sec
+epoch 3, total_step 49840, total loss is 21.00 , inference loss is 10.41, weight deacy loss is 10.59, training accuracy is 0.312500, time 208.770 samples/sec
+epoch 3, total_step 49860, total loss is 22.78 , inference loss is 12.19, weight deacy loss is 10.59, training accuracy is 0.281250, time 224.463 samples/sec
+epoch 3, total_step 49880, total loss is 21.39 , inference loss is 10.81, weight deacy loss is 10.59, training accuracy is 0.312500, time 218.061 samples/sec
+epoch 3, total_step 49900, total loss is 25.25 , inference loss is 14.66, weight deacy loss is 10.59, training accuracy is 0.218750, time 216.875 samples/sec
+epoch 3, total_step 49920, total loss is 22.58 , inference loss is 12.00, weight deacy loss is 10.58, training accuracy is 0.281250, time 213.486 samples/sec
+epoch 3, total_step 49940, total loss is 22.84 , inference loss is 12.26, weight deacy loss is 10.58, training accuracy is 0.250000, time 231.907 samples/sec
+epoch 3, total_step 49960, total loss is 26.82 , inference loss is 16.24, weight deacy loss is 10.58, training accuracy is 0.062500, time 227.136 samples/sec
+epoch 3, total_step 49980, total loss is 23.19 , inference loss is 12.61, weight deacy loss is 10.58, training accuracy is 0.250000, time 229.826 samples/sec
+testing verification..
+(12000, 512)
+infer time 21.998904000000003
+best_threshold_index 117 0.9833333333333333
+best_threshold_index 117 0.9827777777777778
+best_threshold_index 117 0.9840740740740741
+best_threshold_index 117 0.9840740740740741
+best_threshold_index 117 0.9842592592592593
+best_threshold_index 117 0.9824074074074074
+best_threshold_index 117 0.9827777777777778
+best_threshold_index 117 0.9833333333333333
+best_threshold_index 117 0.982037037037037
+best_threshold_index 117 0.9825925925925926
+[lfw][50000]XNorm: 22.703694
+[lfw][50000]Accuracy-Flip: 0.98317+-0.00660
+test accuracy is: 0.9831666666666667
+epoch 3, total_step 50000, total loss is 22.44 , inference loss is 11.86, weight deacy loss is 10.58, training accuracy is 0.250000, time 16.648 samples/sec
+epoch 3, total_step 50020, total loss is 21.51 , inference loss is 10.93, weight deacy loss is 10.58, training accuracy is 0.156250, time 217.473 samples/sec
+epoch 3, total_step 50040, total loss is 23.94 , inference loss is 13.36, weight deacy loss is 10.58, training accuracy is 0.218750, time 202.999 samples/sec
+epoch 3, total_step 50060, total loss is 24.53 , inference loss is 13.95, weight deacy loss is 10.58, training accuracy is 0.156250, time 221.355 samples/sec
+epoch 3, total_step 50080, total loss is 22.02 , inference loss is 11.45, weight deacy loss is 10.57, training accuracy is 0.218750, time 234.148 samples/sec
+epoch 3, total_step 50100, total loss is 27.34 , inference loss is 16.76, weight deacy loss is 10.57, training accuracy is 0.156250, time 11.350 samples/sec
+epoch 3, total_step 50120, total loss is 20.52 , inference loss is 9.95, weight deacy loss is 10.57, training accuracy is 0.187500, time 210.186 samples/sec
+epoch 3, total_step 50140, total loss is 26.09 , inference loss is 15.52, weight deacy loss is 10.57, training accuracy is 0.187500, time 212.493 samples/sec
+epoch 3, total_step 50160, total loss is 21.67 , inference loss is 11.10, weight deacy loss is 10.57, training accuracy is 0.281250, time 214.689 samples/sec
+epoch 3, total_step 50180, total loss is 23.00 , inference loss is 12.44, weight deacy loss is 10.57, training accuracy is 0.187500, time 208.916 samples/sec
+epoch 3, total_step 50200, total loss is 21.97 , inference loss is 11.41, weight deacy loss is 10.57, training accuracy is 0.312500, time 208.470 samples/sec
+epoch 3, total_step 50220, total loss is 23.10 , inference loss is 12.53, weight deacy loss is 10.56, training accuracy is 0.156250, time 297.073 samples/sec
+epoch 3, total_step 50240, total loss is 22.10 , inference loss is 11.54, weight deacy loss is 10.56, training accuracy is 0.218750, time 210.140 samples/sec
+epoch 3, total_step 50260, total loss is 22.03 , inference loss is 11.47, weight deacy loss is 10.56, training accuracy is 0.093750, time 219.868 samples/sec
+epoch 3, total_step 50280, total loss is 20.45 , inference loss is 9.89, weight deacy loss is 10.56, training accuracy is 0.156250, time 217.099 samples/sec
+epoch 3, total_step 50300, total loss is 24.50 , inference loss is 13.95, weight deacy loss is 10.56, training accuracy is 0.250000, time 212.905 samples/sec
+epoch 3, total_step 50320, total loss is 21.79 , inference loss is 11.23, weight deacy loss is 10.56, training accuracy is 0.156250, time 216.711 samples/sec
+epoch 3, total_step 50340, total loss is 22.92 , inference loss is 12.36, weight deacy loss is 10.56, training accuracy is 0.218750, time 205.197 samples/sec
+epoch 3, total_step 50360, total loss is 26.60 , inference loss is 16.04, weight deacy loss is 10.56, training accuracy is 0.156250, time 219.453 samples/sec
+epoch 3, total_step 50380, total loss is 21.30 , inference loss is 10.74, weight deacy loss is 10.55, training accuracy is 0.125000, time 224.872 samples/sec
+epoch 3, total_step 50400, total loss is 26.19 , inference loss is 15.64, weight deacy loss is 10.55, training accuracy is 0.218750, time 13.414 samples/sec
+epoch 3, total_step 50420, total loss is 21.84 , inference loss is 11.29, weight deacy loss is 10.55, training accuracy is 0.187500, time 226.737 samples/sec
+epoch 3, total_step 50440, total loss is 18.32 , inference loss is 7.77, weight deacy loss is 10.55, training accuracy is 0.343750, time 203.741 samples/sec
+epoch 3, total_step 50460, total loss is 22.59 , inference loss is 12.04, weight deacy loss is 10.55, training accuracy is 0.218750, time 197.955 samples/sec
+epoch 3, total_step 50480, total loss is 24.31 , inference loss is 13.76, weight deacy loss is 10.55, training accuracy is 0.156250, time 294.506 samples/sec
+epoch 3, total_step 50500, total loss is 26.73 , inference loss is 16.18, weight deacy loss is 10.55, training accuracy is 0.125000, time 214.666 samples/sec
+epoch 3, total_step 50520, total loss is 23.82 , inference loss is 13.28, weight deacy loss is 10.54, training accuracy is 0.218750, time 212.355 samples/sec
+epoch 3, total_step 50540, total loss is 26.27 , inference loss is 15.73, weight deacy loss is 10.54, training accuracy is 0.125000, time 226.861 samples/sec
+epoch 3, total_step 50560, total loss is 22.87 , inference loss is 12.33, weight deacy loss is 10.54, training accuracy is 0.250000, time 216.522 samples/sec
+epoch 3, total_step 50580, total loss is 21.67 , inference loss is 11.13, weight deacy loss is 10.54, training accuracy is 0.187500, time 215.807 samples/sec
+epoch 3, total_step 50600, total loss is 21.92 , inference loss is 11.38, weight deacy loss is 10.54, training accuracy is 0.312500, time 211.427 samples/sec
+epoch 3, total_step 50620, total loss is 24.28 , inference loss is 13.74, weight deacy loss is 10.54, training accuracy is 0.312500, time 218.359 samples/sec
+epoch 3, total_step 50640, total loss is 24.76 , inference loss is 14.22, weight deacy loss is 10.54, training accuracy is 0.187500, time 202.491 samples/sec
+epoch 3, total_step 50660, total loss is 23.65 , inference loss is 13.12, weight deacy loss is 10.54, training accuracy is 0.156250, time 208.255 samples/sec
+epoch 3, total_step 50680, total loss is 19.86 , inference loss is 9.33, weight deacy loss is 10.53, training accuracy is 0.343750, time 223.845 samples/sec
+epoch 3, total_step 50700, total loss is 26.11 , inference loss is 15.57, weight deacy loss is 10.53, training accuracy is 0.187500, time 13.101 samples/sec
+epoch 3, total_step 50720, total loss is 26.25 , inference loss is 15.72, weight deacy loss is 10.53, training accuracy is 0.156250, time 210.528 samples/sec
+epoch 3, total_step 50740, total loss is 25.91 , inference loss is 15.38, weight deacy loss is 10.53, training accuracy is 0.312500, time 223.487 samples/sec
+epoch 3, total_step 50760, total loss is 24.70 , inference loss is 14.17, weight deacy loss is 10.53, training accuracy is 0.156250, time 298.206 samples/sec
+epoch 3, total_step 50780, total loss is 23.86 , inference loss is 13.33, weight deacy loss is 10.53, training accuracy is 0.375000, time 206.989 samples/sec
+epoch 3, total_step 50800, total loss is 24.27 , inference loss is 13.74, weight deacy loss is 10.53, training accuracy is 0.156250, time 213.973 samples/sec
+epoch 3, total_step 50820, total loss is 19.77 , inference loss is 9.25, weight deacy loss is 10.52, training accuracy is 0.281250, time 218.850 samples/sec
+epoch 3, total_step 50840, total loss is 20.44 , inference loss is 9.91, weight deacy loss is 10.52, training accuracy is 0.156250, time 212.443 samples/sec
+epoch 3, total_step 50860, total loss is 25.95 , inference loss is 15.43, weight deacy loss is 10.52, training accuracy is 0.156250, time 207.379 samples/sec
+epoch 3, total_step 50880, total loss is 24.59 , inference loss is 14.07, weight deacy loss is 10.52, training accuracy is 0.125000, time 218.425 samples/sec
+epoch 3, total_step 50900, total loss is 22.36 , inference loss is 11.84, weight deacy loss is 10.52, training accuracy is 0.250000, time 206.843 samples/sec
+epoch 3, total_step 50920, total loss is 26.04 , inference loss is 15.52, weight deacy loss is 10.52, training accuracy is 0.125000, time 216.011 samples/sec
+epoch 3, total_step 50940, total loss is 22.52 , inference loss is 12.01, weight deacy loss is 10.52, training accuracy is 0.156250, time 211.008 samples/sec
+epoch 3, total_step 50960, total loss is 24.66 , inference loss is 14.15, weight deacy loss is 10.52, training accuracy is 0.125000, time 211.779 samples/sec
+epoch 3, total_step 50980, total loss is 26.90 , inference loss is 16.39, weight deacy loss is 10.51, training accuracy is 0.125000, time 207.184 samples/sec
+epoch 3, total_step 51000, total loss is 24.84 , inference loss is 14.32, weight deacy loss is 10.51, training accuracy is 0.187500, time 12.261 samples/sec
+epoch 3, total_step 51020, total loss is 25.92 , inference loss is 15.41, weight deacy loss is 10.51, training accuracy is 0.187500, time 295.445 samples/sec
+epoch 3, total_step 51040, total loss is 21.73 , inference loss is 11.22, weight deacy loss is 10.51, training accuracy is 0.312500, time 211.618 samples/sec
+epoch 3, total_step 51060, total loss is 26.34 , inference loss is 15.83, weight deacy loss is 10.51, training accuracy is 0.218750, time 225.097 samples/sec
+epoch 3, total_step 51080, total loss is 26.93 , inference loss is 16.42, weight deacy loss is 10.51, training accuracy is 0.125000, time 206.391 samples/sec
+epoch 3, total_step 51100, total loss is 23.62 , inference loss is 13.12, weight deacy loss is 10.51, training accuracy is 0.187500, time 216.805 samples/sec
+epoch 3, total_step 51120, total loss is 24.31 , inference loss is 13.81, weight deacy loss is 10.51, training accuracy is 0.312500, time 205.237 samples/sec
+epoch 3, total_step 51140, total loss is 21.75 , inference loss is 11.25, weight deacy loss is 10.50, training accuracy is 0.156250, time 215.994 samples/sec
+epoch 3, total_step 51160, total loss is 27.45 , inference loss is 16.95, weight deacy loss is 10.50, training accuracy is 0.187500, time 219.770 samples/sec
+epoch 3, total_step 51180, total loss is 24.03 , inference loss is 13.53, weight deacy loss is 10.50, training accuracy is 0.250000, time 211.093 samples/sec
+epoch 3, total_step 51200, total loss is 26.55 , inference loss is 16.05, weight deacy loss is 10.50, training accuracy is 0.218750, time 215.441 samples/sec
+epoch 3, total_step 51220, total loss is 23.75 , inference loss is 13.25, weight deacy loss is 10.50, training accuracy is 0.312500, time 206.222 samples/sec
+epoch 3, total_step 51240, total loss is 22.20 , inference loss is 11.70, weight deacy loss is 10.50, training accuracy is 0.187500, time 213.218 samples/sec
+epoch 3, total_step 51260, total loss is 19.34 , inference loss is 8.84, weight deacy loss is 10.50, training accuracy is 0.406250, time 212.574 samples/sec
+epoch 3, total_step 51280, total loss is 25.77 , inference loss is 15.27, weight deacy loss is 10.49, training accuracy is 0.218750, time 214.108 samples/sec
+epoch 3, total_step 51300, total loss is 24.81 , inference loss is 14.32, weight deacy loss is 10.49, training accuracy is 0.187500, time 11.266 samples/sec
+epoch 3, total_step 51320, total loss is 22.61 , inference loss is 12.12, weight deacy loss is 10.49, training accuracy is 0.343750, time 223.876 samples/sec
+epoch 3, total_step 51340, total loss is 24.75 , inference loss is 14.26, weight deacy loss is 10.49, training accuracy is 0.250000, time 210.287 samples/sec
+epoch 3, total_step 51360, total loss is 21.19 , inference loss is 10.70, weight deacy loss is 10.49, training accuracy is 0.156250, time 216.602 samples/sec
+epoch 3, total_step 51380, total loss is 20.86 , inference loss is 10.37, weight deacy loss is 10.49, training accuracy is 0.218750, time 208.386 samples/sec
+epoch 3, total_step 51400, total loss is 21.01 , inference loss is 10.52, weight deacy loss is 10.49, training accuracy is 0.218750, time 224.336 samples/sec
+epoch 3, total_step 51420, total loss is 24.80 , inference loss is 14.31, weight deacy loss is 10.49, training accuracy is 0.218750, time 219.100 samples/sec
+epoch 3, total_step 51440, total loss is 23.99 , inference loss is 13.51, weight deacy loss is 10.48, training accuracy is 0.156250, time 215.544 samples/sec
+epoch 3, total_step 51460, total loss is 22.09 , inference loss is 11.61, weight deacy loss is 10.48, training accuracy is 0.312500, time 215.423 samples/sec
+epoch 3, total_step 51480, total loss is 23.02 , inference loss is 12.54, weight deacy loss is 10.48, training accuracy is 0.187500, time 243.513 samples/sec
+epoch 3, total_step 51500, total loss is 21.53 , inference loss is 11.05, weight deacy loss is 10.48, training accuracy is 0.312500, time 221.312 samples/sec
+epoch 3, total_step 51520, total loss is 20.86 , inference loss is 10.38, weight deacy loss is 10.48, training accuracy is 0.375000, time 250.049 samples/sec
+epoch 3, total_step 51540, total loss is 20.25 , inference loss is 9.78, weight deacy loss is 10.48, training accuracy is 0.312500, time 213.232 samples/sec
+epoch 3, total_step 51560, total loss is 22.51 , inference loss is 12.03, weight deacy loss is 10.48, training accuracy is 0.250000, time 208.542 samples/sec
+epoch 3, total_step 51580, total loss is 25.03 , inference loss is 14.56, weight deacy loss is 10.48, training accuracy is 0.250000, time 220.290 samples/sec
+epoch 3, total_step 51600, total loss is 24.37 , inference loss is 13.90, weight deacy loss is 10.47, training accuracy is 0.218750, time 13.560 samples/sec
+epoch 3, total_step 51620, total loss is 22.67 , inference loss is 12.19, weight deacy loss is 10.47, training accuracy is 0.343750, time 214.956 samples/sec
+epoch 3, total_step 51640, total loss is 20.75 , inference loss is 10.28, weight deacy loss is 10.47, training accuracy is 0.281250, time 199.792 samples/sec
+epoch 3, total_step 51660, total loss is 24.21 , inference loss is 13.74, weight deacy loss is 10.47, training accuracy is 0.187500, time 213.638 samples/sec
+epoch 3, total_step 51680, total loss is 27.09 , inference loss is 16.62, weight deacy loss is 10.47, training accuracy is 0.218750, time 209.173 samples/sec
+epoch 3, total_step 51700, total loss is 22.19 , inference loss is 11.72, weight deacy loss is 10.47, training accuracy is 0.187500, time 213.757 samples/sec
+epoch 3, total_step 51720, total loss is 24.92 , inference loss is 14.45, weight deacy loss is 10.47, training accuracy is 0.250000, time 226.300 samples/sec
+epoch 3, total_step 51740, total loss is 22.39 , inference loss is 11.93, weight deacy loss is 10.46, training accuracy is 0.312500, time 206.220 samples/sec
+epoch 3, total_step 51760, total loss is 19.77 , inference loss is 9.30, weight deacy loss is 10.46, training accuracy is 0.281250, time 212.889 samples/sec
+epoch 3, total_step 51780, total loss is 26.49 , inference loss is 16.02, weight deacy loss is 10.46, training accuracy is 0.218750, time 207.745 samples/sec
+epoch 3, total_step 51800, total loss is 27.29 , inference loss is 16.83, weight deacy loss is 10.46, training accuracy is 0.187500, time 215.613 samples/sec
+epoch 3, total_step 51820, total loss is 25.96 , inference loss is 15.50, weight deacy loss is 10.46, training accuracy is 0.187500, time 297.582 samples/sec
+epoch 3, total_step 51840, total loss is 25.33 , inference loss is 14.87, weight deacy loss is 10.46, training accuracy is 0.187500, time 298.046 samples/sec
+epoch 3, total_step 51860, total loss is 25.08 , inference loss is 14.62, weight deacy loss is 10.46, training accuracy is 0.250000, time 205.691 samples/sec
+epoch 3, total_step 51880, total loss is 22.91 , inference loss is 12.45, weight deacy loss is 10.46, training accuracy is 0.281250, time 205.183 samples/sec
+epoch 3, total_step 51900, total loss is 21.87 , inference loss is 11.41, weight deacy loss is 10.45, training accuracy is 0.312500, time 13.187 samples/sec
+epoch 3, total_step 51920, total loss is 24.87 , inference loss is 14.41, weight deacy loss is 10.45, training accuracy is 0.281250, time 222.735 samples/sec
+epoch 3, total_step 51940, total loss is 25.77 , inference loss is 15.32, weight deacy loss is 10.45, training accuracy is 0.156250, time 216.242 samples/sec
+epoch 3, total_step 51960, total loss is 26.16 , inference loss is 15.71, weight deacy loss is 10.45, training accuracy is 0.281250, time 211.738 samples/sec
+epoch 3, total_step 51980, total loss is 26.31 , inference loss is 15.86, weight deacy loss is 10.45, training accuracy is 0.156250, time 201.035 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.471604000000013
+best_threshold_index 118 0.9846296296296296
+best_threshold_index 118 0.9833333333333333
+best_threshold_index 118 0.9844444444444445
+best_threshold_index 118 0.9848148148148148
+best_threshold_index 118 0.9844444444444445
+best_threshold_index 118 0.9835185185185186
+best_threshold_index 118 0.9837037037037037
+best_threshold_index 118 0.9835185185185186
+best_threshold_index 118 0.9825925925925926
+best_threshold_index 118 0.9833333333333333
+[lfw][52000]XNorm: 21.851396
+[lfw][52000]Accuracy-Flip: 0.98383+-0.00610
+test accuracy is: 0.9838333333333333
+epoch 3, total_step 52000, total loss is 23.41 , inference loss is 12.97, weight deacy loss is 10.45, training accuracy is 0.187500, time 18.774 samples/sec
+epoch 3, total_step 52020, total loss is 23.60 , inference loss is 13.15, weight deacy loss is 10.45, training accuracy is 0.093750, time 217.938 samples/sec
+epoch 3, total_step 52040, total loss is 22.00 , inference loss is 11.56, weight deacy loss is 10.45, training accuracy is 0.343750, time 215.257 samples/sec
+epoch 3, total_step 52060, total loss is 29.34 , inference loss is 18.89, weight deacy loss is 10.44, training accuracy is 0.187500, time 218.530 samples/sec
+epoch 3, total_step 52080, total loss is 24.29 , inference loss is 13.85, weight deacy loss is 10.44, training accuracy is 0.281250, time 211.061 samples/sec
+epoch 3, total_step 52100, total loss is 23.95 , inference loss is 13.51, weight deacy loss is 10.44, training accuracy is 0.187500, time 220.075 samples/sec
+epoch 3, total_step 52120, total loss is 23.78 , inference loss is 13.34, weight deacy loss is 10.44, training accuracy is 0.125000, time 299.180 samples/sec
+epoch 3, total_step 52140, total loss is 23.44 , inference loss is 13.00, weight deacy loss is 10.44, training accuracy is 0.156250, time 220.972 samples/sec
+epoch 3, total_step 52160, total loss is 30.43 , inference loss is 19.99, weight deacy loss is 10.44, training accuracy is 0.125000, time 227.049 samples/sec
+epoch 3, total_step 52180, total loss is 20.91 , inference loss is 10.47, weight deacy loss is 10.44, training accuracy is 0.250000, time 211.204 samples/sec
+epoch 3, total_step 52200, total loss is 23.97 , inference loss is 13.53, weight deacy loss is 10.43, training accuracy is 0.281250, time 13.553 samples/sec
+epoch 3, total_step 52220, total loss is 21.11 , inference loss is 10.68, weight deacy loss is 10.43, training accuracy is 0.343750, time 209.790 samples/sec
+epoch 3, total_step 52240, total loss is 22.33 , inference loss is 11.90, weight deacy loss is 10.43, training accuracy is 0.093750, time 214.014 samples/sec
+epoch 3, total_step 52260, total loss is 25.91 , inference loss is 15.48, weight deacy loss is 10.43, training accuracy is 0.218750, time 226.194 samples/sec
+epoch 3, total_step 52280, total loss is 23.56 , inference loss is 13.13, weight deacy loss is 10.43, training accuracy is 0.250000, time 220.147 samples/sec
+epoch 3, total_step 52300, total loss is 22.54 , inference loss is 12.11, weight deacy loss is 10.43, training accuracy is 0.281250, time 220.948 samples/sec
+epoch 3, total_step 52320, total loss is 23.99 , inference loss is 13.56, weight deacy loss is 10.43, training accuracy is 0.187500, time 212.239 samples/sec
+epoch 3, total_step 52340, total loss is 25.22 , inference loss is 14.79, weight deacy loss is 10.43, training accuracy is 0.281250, time 218.115 samples/sec
+epoch 3, total_step 52360, total loss is 25.73 , inference loss is 15.31, weight deacy loss is 10.42, training accuracy is 0.312500, time 203.224 samples/sec
+epoch 3, total_step 52380, total loss is 24.21 , inference loss is 13.78, weight deacy loss is 10.42, training accuracy is 0.312500, time 299.205 samples/sec
+epoch 3, total_step 52400, total loss is 25.63 , inference loss is 15.21, weight deacy loss is 10.42, training accuracy is 0.187500, time 213.826 samples/sec
+epoch 3, total_step 52420, total loss is 21.59 , inference loss is 11.17, weight deacy loss is 10.42, training accuracy is 0.312500, time 213.919 samples/sec
+epoch 3, total_step 52440, total loss is 26.23 , inference loss is 15.82, weight deacy loss is 10.42, training accuracy is 0.187500, time 201.100 samples/sec
+epoch 3, total_step 52460, total loss is 24.57 , inference loss is 14.15, weight deacy loss is 10.42, training accuracy is 0.281250, time 214.148 samples/sec
+epoch 3, total_step 52480, total loss is 24.06 , inference loss is 13.65, weight deacy loss is 10.42, training accuracy is 0.218750, time 225.373 samples/sec
+epoch 3, total_step 52500, total loss is 23.86 , inference loss is 13.44, weight deacy loss is 10.42, training accuracy is 0.281250, time 12.176 samples/sec
+epoch 3, total_step 52520, total loss is 21.82 , inference loss is 11.40, weight deacy loss is 10.41, training accuracy is 0.125000, time 215.888 samples/sec
+epoch 3, total_step 52540, total loss is 22.50 , inference loss is 12.09, weight deacy loss is 10.41, training accuracy is 0.187500, time 211.140 samples/sec
+epoch 3, total_step 52560, total loss is 21.82 , inference loss is 11.41, weight deacy loss is 10.41, training accuracy is 0.281250, time 211.711 samples/sec
+epoch 3, total_step 52580, total loss is 20.93 , inference loss is 10.52, weight deacy loss is 10.41, training accuracy is 0.375000, time 220.094 samples/sec
+epoch 3, total_step 52600, total loss is 24.74 , inference loss is 14.33, weight deacy loss is 10.41, training accuracy is 0.187500, time 212.135 samples/sec
+epoch 3, total_step 52620, total loss is 23.17 , inference loss is 12.76, weight deacy loss is 10.41, training accuracy is 0.125000, time 235.743 samples/sec
+epoch 3, total_step 52640, total loss is 23.28 , inference loss is 12.87, weight deacy loss is 10.41, training accuracy is 0.187500, time 297.054 samples/sec
+epoch 3, total_step 52660, total loss is 22.76 , inference loss is 12.36, weight deacy loss is 10.40, training accuracy is 0.218750, time 205.568 samples/sec
+epoch 3, total_step 52680, total loss is 23.20 , inference loss is 12.79, weight deacy loss is 10.40, training accuracy is 0.218750, time 204.022 samples/sec
+epoch 3, total_step 52700, total loss is 24.51 , inference loss is 14.11, weight deacy loss is 10.40, training accuracy is 0.156250, time 208.259 samples/sec
+epoch 3, total_step 52720, total loss is 23.37 , inference loss is 12.97, weight deacy loss is 10.40, training accuracy is 0.125000, time 211.061 samples/sec
+epoch 3, total_step 52740, total loss is 22.77 , inference loss is 12.37, weight deacy loss is 10.40, training accuracy is 0.250000, time 211.994 samples/sec
+epoch 3, total_step 52760, total loss is 22.84 , inference loss is 12.44, weight deacy loss is 10.40, training accuracy is 0.281250, time 235.435 samples/sec
+epoch 3, total_step 52780, total loss is 23.21 , inference loss is 12.81, weight deacy loss is 10.40, training accuracy is 0.187500, time 222.151 samples/sec
+epoch 3, total_step 52800, total loss is 23.77 , inference loss is 13.37, weight deacy loss is 10.40, training accuracy is 0.281250, time 11.855 samples/sec
+epoch 3, total_step 52820, total loss is 22.54 , inference loss is 12.14, weight deacy loss is 10.39, training accuracy is 0.187500, time 212.022 samples/sec
+epoch 3, total_step 52840, total loss is 26.30 , inference loss is 15.91, weight deacy loss is 10.39, training accuracy is 0.156250, time 205.393 samples/sec
+epoch 3, total_step 52860, total loss is 26.60 , inference loss is 16.20, weight deacy loss is 10.39, training accuracy is 0.156250, time 212.228 samples/sec
+epoch 3, total_step 52880, total loss is 22.22 , inference loss is 11.83, weight deacy loss is 10.39, training accuracy is 0.250000, time 211.972 samples/sec
+epoch 3, total_step 52900, total loss is 25.72 , inference loss is 15.33, weight deacy loss is 10.39, training accuracy is 0.187500, time 212.177 samples/sec
+epoch 3, total_step 52920, total loss is 28.21 , inference loss is 17.83, weight deacy loss is 10.39, training accuracy is 0.156250, time 198.202 samples/sec
+epoch 3, total_step 52940, total loss is 22.55 , inference loss is 12.16, weight deacy loss is 10.39, training accuracy is 0.156250, time 211.667 samples/sec
+epoch 3, total_step 52960, total loss is 23.19 , inference loss is 12.81, weight deacy loss is 10.39, training accuracy is 0.187500, time 202.607 samples/sec
+epoch 3, total_step 52980, total loss is 21.66 , inference loss is 11.28, weight deacy loss is 10.38, training accuracy is 0.156250, time 208.344 samples/sec
+epoch 3, total_step 53000, total loss is 24.04 , inference loss is 13.65, weight deacy loss is 10.38, training accuracy is 0.218750, time 212.628 samples/sec
+epoch 3, total_step 53020, total loss is 25.14 , inference loss is 14.76, weight deacy loss is 10.38, training accuracy is 0.156250, time 214.488 samples/sec
+epoch 3, total_step 53040, total loss is 28.60 , inference loss is 18.21, weight deacy loss is 10.38, training accuracy is 0.187500, time 210.932 samples/sec
+epoch 3, total_step 53060, total loss is 25.04 , inference loss is 14.66, weight deacy loss is 10.38, training accuracy is 0.187500, time 215.534 samples/sec
+epoch 3, total_step 53080, total loss is 21.54 , inference loss is 11.16, weight deacy loss is 10.38, training accuracy is 0.281250, time 211.358 samples/sec
+epoch 3, total_step 53100, total loss is 22.43 , inference loss is 12.05, weight deacy loss is 10.38, training accuracy is 0.187500, time 12.448 samples/sec
+epoch 3, total_step 53120, total loss is 29.74 , inference loss is 19.36, weight deacy loss is 10.38, training accuracy is 0.218750, time 214.069 samples/sec
+epoch 3, total_step 53140, total loss is 22.15 , inference loss is 11.77, weight deacy loss is 10.37, training accuracy is 0.125000, time 209.775 samples/sec
+epoch 3, total_step 53160, total loss is 24.64 , inference loss is 14.27, weight deacy loss is 10.37, training accuracy is 0.218750, time 210.131 samples/sec
+epoch 3, total_step 53180, total loss is 24.03 , inference loss is 13.66, weight deacy loss is 10.37, training accuracy is 0.281250, time 295.817 samples/sec
+epoch 3, total_step 53200, total loss is 22.83 , inference loss is 12.46, weight deacy loss is 10.37, training accuracy is 0.125000, time 206.920 samples/sec
+epoch 3, total_step 53220, total loss is 27.27 , inference loss is 16.90, weight deacy loss is 10.37, training accuracy is 0.281250, time 211.283 samples/sec
+epoch 3, total_step 53240, total loss is 26.26 , inference loss is 15.89, weight deacy loss is 10.37, training accuracy is 0.187500, time 212.518 samples/sec
+epoch 3, total_step 53260, total loss is 29.13 , inference loss is 18.76, weight deacy loss is 10.37, training accuracy is 0.062500, time 215.327 samples/sec
+epoch 3, total_step 53280, total loss is 25.15 , inference loss is 14.79, weight deacy loss is 10.37, training accuracy is 0.218750, time 208.897 samples/sec
+epoch 3, total_step 53300, total loss is 23.35 , inference loss is 12.99, weight deacy loss is 10.36, training accuracy is 0.093750, time 206.391 samples/sec
+epoch 3, total_step 53320, total loss is 23.52 , inference loss is 13.16, weight deacy loss is 10.36, training accuracy is 0.250000, time 209.235 samples/sec
+epoch 3, total_step 53340, total loss is 23.33 , inference loss is 12.97, weight deacy loss is 10.36, training accuracy is 0.187500, time 204.423 samples/sec
+epoch 3, total_step 53360, total loss is 21.37 , inference loss is 11.01, weight deacy loss is 10.36, training accuracy is 0.250000, time 208.934 samples/sec
+epoch 3, total_step 53380, total loss is 23.91 , inference loss is 13.55, weight deacy loss is 10.36, training accuracy is 0.156250, time 207.577 samples/sec
+epoch 3, total_step 53400, total loss is 26.78 , inference loss is 16.42, weight deacy loss is 10.36, training accuracy is 0.218750, time 14.127 samples/sec
+epoch 3, total_step 53420, total loss is 22.63 , inference loss is 12.27, weight deacy loss is 10.36, training accuracy is 0.250000, time 240.926 samples/sec
+epoch 3, total_step 53440, total loss is 20.39 , inference loss is 10.04, weight deacy loss is 10.35, training accuracy is 0.312500, time 297.640 samples/sec
+epoch 3, total_step 53460, total loss is 23.92 , inference loss is 13.56, weight deacy loss is 10.35, training accuracy is 0.156250, time 214.228 samples/sec
+epoch 3, total_step 53480, total loss is 19.58 , inference loss is 9.23, weight deacy loss is 10.35, training accuracy is 0.375000, time 220.580 samples/sec
+epoch 3, total_step 53500, total loss is 23.52 , inference loss is 13.17, weight deacy loss is 10.35, training accuracy is 0.156250, time 215.558 samples/sec
+epoch 3, total_step 53520, total loss is 21.02 , inference loss is 10.67, weight deacy loss is 10.35, training accuracy is 0.218750, time 215.343 samples/sec
+epoch 3, total_step 53540, total loss is 23.41 , inference loss is 13.06, weight deacy loss is 10.35, training accuracy is 0.250000, time 235.614 samples/sec
+epoch 3, total_step 53560, total loss is 19.39 , inference loss is 9.05, weight deacy loss is 10.35, training accuracy is 0.281250, time 211.733 samples/sec
+epoch 3, total_step 53580, total loss is 23.79 , inference loss is 13.44, weight deacy loss is 10.35, training accuracy is 0.218750, time 219.642 samples/sec
+epoch 3, total_step 53600, total loss is 24.02 , inference loss is 13.67, weight deacy loss is 10.34, training accuracy is 0.125000, time 296.237 samples/sec
+epoch 3, total_step 53620, total loss is 22.56 , inference loss is 12.22, weight deacy loss is 10.34, training accuracy is 0.281250, time 297.810 samples/sec
+epoch 3, total_step 53640, total loss is 23.23 , inference loss is 12.89, weight deacy loss is 10.34, training accuracy is 0.093750, time 214.243 samples/sec
+epoch 3, total_step 53660, total loss is 24.89 , inference loss is 14.55, weight deacy loss is 10.34, training accuracy is 0.156250, time 204.742 samples/sec
+epoch 3, total_step 53680, total loss is 21.87 , inference loss is 11.53, weight deacy loss is 10.34, training accuracy is 0.218750, time 209.953 samples/sec
+epoch 3, total_step 53700, total loss is 19.20 , inference loss is 8.87, weight deacy loss is 10.34, training accuracy is 0.437500, time 14.549 samples/sec
+epoch 3, total_step 53720, total loss is 25.75 , inference loss is 15.42, weight deacy loss is 10.34, training accuracy is 0.125000, time 244.075 samples/sec
+epoch 3, total_step 53740, total loss is 23.29 , inference loss is 12.96, weight deacy loss is 10.34, training accuracy is 0.218750, time 207.479 samples/sec
+epoch 3, total_step 53760, total loss is 17.64 , inference loss is 7.31, weight deacy loss is 10.33, training accuracy is 0.375000, time 211.601 samples/sec
+epoch 3, total_step 53780, total loss is 21.91 , inference loss is 11.57, weight deacy loss is 10.33, training accuracy is 0.281250, time 202.975 samples/sec
+epoch 3, total_step 53800, total loss is 24.60 , inference loss is 14.27, weight deacy loss is 10.33, training accuracy is 0.218750, time 214.654 samples/sec
+epoch 3, total_step 53820, total loss is 22.56 , inference loss is 12.23, weight deacy loss is 10.33, training accuracy is 0.250000, time 209.227 samples/sec
+epoch 3, total_step 53840, total loss is 25.50 , inference loss is 15.17, weight deacy loss is 10.33, training accuracy is 0.250000, time 213.524 samples/sec
+epoch 3, total_step 53860, total loss is 21.70 , inference loss is 11.37, weight deacy loss is 10.33, training accuracy is 0.187500, time 209.898 samples/sec
+epoch 3, total_step 53880, total loss is 21.36 , inference loss is 11.03, weight deacy loss is 10.33, training accuracy is 0.156250, time 206.158 samples/sec
+epoch 3, total_step 53900, total loss is 21.15 , inference loss is 10.82, weight deacy loss is 10.32, training accuracy is 0.406250, time 301.036 samples/sec
+epoch 3, total_step 53920, total loss is 25.05 , inference loss is 14.72, weight deacy loss is 10.32, training accuracy is 0.125000, time 209.366 samples/sec
+epoch 3, total_step 53940, total loss is 24.58 , inference loss is 14.26, weight deacy loss is 10.32, training accuracy is 0.281250, time 209.044 samples/sec
+epoch 3, total_step 53960, total loss is 22.13 , inference loss is 11.81, weight deacy loss is 10.32, training accuracy is 0.281250, time 212.242 samples/sec
+epoch 3, total_step 53980, total loss is 25.07 , inference loss is 14.75, weight deacy loss is 10.32, training accuracy is 0.187500, time 217.779 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.69664699999997
+best_threshold_index 114 0.9835185185185186
+best_threshold_index 117 0.9825925925925926
+best_threshold_index 114 0.9835185185185186
+best_threshold_index 114 0.9833333333333333
+best_threshold_index 114 0.9827777777777778
+best_threshold_index 114 0.9822222222222222
+best_threshold_index 114 0.9827777777777778
+best_threshold_index 114 0.982037037037037
+best_threshold_index 117 0.9812962962962963
+best_threshold_index 117 0.9818518518518519
+[lfw][54000]XNorm: 22.598574
+[lfw][54000]Accuracy-Flip: 0.98167+-0.00637
+test accuracy is: 0.9816666666666667
+epoch 3, total_step 54000, total loss is 22.37 , inference loss is 12.05, weight deacy loss is 10.32, training accuracy is 0.375000, time 17.525 samples/sec
+epoch 3, total_step 54020, total loss is 19.46 , inference loss is 9.14, weight deacy loss is 10.32, training accuracy is 0.343750, time 215.059 samples/sec
+epoch 3, total_step 54040, total loss is 25.59 , inference loss is 15.28, weight deacy loss is 10.32, training accuracy is 0.218750, time 230.942 samples/sec
+epoch 3, total_step 54060, total loss is 21.58 , inference loss is 11.26, weight deacy loss is 10.31, training accuracy is 0.093750, time 214.096 samples/sec
+epoch 3, total_step 54080, total loss is 22.69 , inference loss is 12.38, weight deacy loss is 10.31, training accuracy is 0.250000, time 208.262 samples/sec
+epoch 3, total_step 54100, total loss is 19.77 , inference loss is 9.46, weight deacy loss is 10.31, training accuracy is 0.187500, time 214.040 samples/sec
+epoch 3, total_step 54120, total loss is 23.83 , inference loss is 13.52, weight deacy loss is 10.31, training accuracy is 0.218750, time 215.472 samples/sec
+epoch 3, total_step 54140, total loss is 24.52 , inference loss is 14.21, weight deacy loss is 10.31, training accuracy is 0.250000, time 213.635 samples/sec
+epoch 3, total_step 54160, total loss is 23.18 , inference loss is 12.87, weight deacy loss is 10.31, training accuracy is 0.250000, time 224.994 samples/sec
+epoch 3, total_step 54180, total loss is 26.15 , inference loss is 15.84, weight deacy loss is 10.31, training accuracy is 0.187500, time 221.228 samples/sec
+epoch 3, total_step 54200, total loss is 24.37 , inference loss is 14.06, weight deacy loss is 10.31, training accuracy is 0.281250, time 216.748 samples/sec
+epoch 3, total_step 54220, total loss is 22.53 , inference loss is 12.22, weight deacy loss is 10.30, training accuracy is 0.062500, time 218.529 samples/sec
+epoch 3, total_step 54240, total loss is 24.39 , inference loss is 14.09, weight deacy loss is 10.30, training accuracy is 0.250000, time 209.317 samples/sec
+epoch 3, total_step 54260, total loss is 27.16 , inference loss is 16.86, weight deacy loss is 10.30, training accuracy is 0.156250, time 213.257 samples/sec
+epoch 3, total_step 54280, total loss is 20.23 , inference loss is 9.93, weight deacy loss is 10.30, training accuracy is 0.281250, time 298.230 samples/sec
+epoch 3, total_step 54300, total loss is 21.89 , inference loss is 11.59, weight deacy loss is 10.30, training accuracy is 0.187500, time 14.153 samples/sec
+epoch 3, total_step 54320, total loss is 25.19 , inference loss is 14.89, weight deacy loss is 10.30, training accuracy is 0.156250, time 212.431 samples/sec
+epoch 3, total_step 54340, total loss is 25.85 , inference loss is 15.55, weight deacy loss is 10.30, training accuracy is 0.187500, time 226.201 samples/sec
+epoch 3, total_step 54360, total loss is 25.78 , inference loss is 15.48, weight deacy loss is 10.30, training accuracy is 0.218750, time 208.094 samples/sec
+epoch 3, total_step 54380, total loss is 20.89 , inference loss is 10.60, weight deacy loss is 10.29, training accuracy is 0.312500, time 212.146 samples/sec
+epoch 3, total_step 54400, total loss is 24.20 , inference loss is 13.91, weight deacy loss is 10.29, training accuracy is 0.156250, time 207.884 samples/sec
+epoch 3, total_step 54420, total loss is 21.86 , inference loss is 11.56, weight deacy loss is 10.29, training accuracy is 0.156250, time 225.779 samples/sec
+epoch 3, total_step 54440, total loss is 21.02 , inference loss is 10.73, weight deacy loss is 10.29, training accuracy is 0.250000, time 213.528 samples/sec
+epoch 3, total_step 54460, total loss is 27.75 , inference loss is 17.46, weight deacy loss is 10.29, training accuracy is 0.125000, time 204.679 samples/sec
+epoch 3, total_step 54480, total loss is 23.17 , inference loss is 12.88, weight deacy loss is 10.29, training accuracy is 0.156250, time 208.204 samples/sec
+epoch 3, total_step 54500, total loss is 26.12 , inference loss is 15.84, weight deacy loss is 10.29, training accuracy is 0.187500, time 217.092 samples/sec
+epoch 3, total_step 54520, total loss is 27.84 , inference loss is 17.56, weight deacy loss is 10.29, training accuracy is 0.125000, time 219.811 samples/sec
+epoch 3, total_step 54540, total loss is 22.36 , inference loss is 12.08, weight deacy loss is 10.28, training accuracy is 0.250000, time 202.088 samples/sec
+epoch 3, total_step 54560, total loss is 23.88 , inference loss is 13.59, weight deacy loss is 10.28, training accuracy is 0.187500, time 299.039 samples/sec
+epoch 3, total_step 54580, total loss is 23.65 , inference loss is 13.37, weight deacy loss is 10.28, training accuracy is 0.156250, time 209.320 samples/sec
+epoch 3, total_step 54600, total loss is 21.28 , inference loss is 11.00, weight deacy loss is 10.28, training accuracy is 0.250000, time 14.405 samples/sec
+epoch 3, total_step 54620, total loss is 24.07 , inference loss is 13.79, weight deacy loss is 10.28, training accuracy is 0.250000, time 209.485 samples/sec
+epoch 3, total_step 54640, total loss is 22.72 , inference loss is 12.44, weight deacy loss is 10.28, training accuracy is 0.156250, time 212.514 samples/sec
+epoch 3, total_step 54660, total loss is 25.66 , inference loss is 15.38, weight deacy loss is 10.28, training accuracy is 0.187500, time 206.666 samples/sec
+epoch 3, total_step 54680, total loss is 22.50 , inference loss is 12.22, weight deacy loss is 10.28, training accuracy is 0.218750, time 212.976 samples/sec
+epoch 3, total_step 54700, total loss is 26.44 , inference loss is 16.17, weight deacy loss is 10.27, training accuracy is 0.093750, time 205.637 samples/sec
+epoch 3, total_step 54720, total loss is 24.70 , inference loss is 14.43, weight deacy loss is 10.27, training accuracy is 0.187500, time 225.320 samples/sec
+epoch 3, total_step 54740, total loss is 21.79 , inference loss is 11.52, weight deacy loss is 10.27, training accuracy is 0.187500, time 213.376 samples/sec
+epoch 3, total_step 54760, total loss is 21.77 , inference loss is 11.50, weight deacy loss is 10.27, training accuracy is 0.218750, time 200.412 samples/sec
+epoch 3, total_step 54780, total loss is 22.34 , inference loss is 12.07, weight deacy loss is 10.27, training accuracy is 0.187500, time 208.782 samples/sec
+epoch 3, total_step 54800, total loss is 22.30 , inference loss is 12.03, weight deacy loss is 10.27, training accuracy is 0.125000, time 221.778 samples/sec
+epoch 3, total_step 54820, total loss is 22.79 , inference loss is 12.53, weight deacy loss is 10.27, training accuracy is 0.281250, time 297.841 samples/sec
+epoch 3, total_step 54840, total loss is 24.89 , inference loss is 14.62, weight deacy loss is 10.27, training accuracy is 0.156250, time 220.165 samples/sec
+epoch 3, total_step 54860, total loss is 27.00 , inference loss is 16.74, weight deacy loss is 10.26, training accuracy is 0.125000, time 216.473 samples/sec
+epoch 3, total_step 54880, total loss is 24.91 , inference loss is 14.65, weight deacy loss is 10.26, training accuracy is 0.093750, time 206.077 samples/sec
+epoch 3, total_step 54900, total loss is 26.96 , inference loss is 16.70, weight deacy loss is 10.26, training accuracy is 0.218750, time 13.878 samples/sec
+epoch 3, total_step 54920, total loss is 22.53 , inference loss is 12.27, weight deacy loss is 10.26, training accuracy is 0.156250, time 209.229 samples/sec
+epoch 3, total_step 54940, total loss is 20.62 , inference loss is 10.36, weight deacy loss is 10.26, training accuracy is 0.375000, time 206.224 samples/sec
+epoch 3, total_step 54960, total loss is 24.91 , inference loss is 14.65, weight deacy loss is 10.26, training accuracy is 0.156250, time 215.993 samples/sec
+epoch 3, total_step 54980, total loss is 22.76 , inference loss is 12.50, weight deacy loss is 10.26, training accuracy is 0.093750, time 212.520 samples/sec
+epoch 3, total_step 55000, total loss is 21.62 , inference loss is 11.36, weight deacy loss is 10.26, training accuracy is 0.218750, time 202.789 samples/sec
+epoch 3, total_step 55020, total loss is 23.84 , inference loss is 13.58, weight deacy loss is 10.25, training accuracy is 0.218750, time 242.246 samples/sec
+epoch 3, total_step 55040, total loss is 25.03 , inference loss is 14.78, weight deacy loss is 10.25, training accuracy is 0.218750, time 204.172 samples/sec
+epoch 3, total_step 55060, total loss is 24.75 , inference loss is 14.50, weight deacy loss is 10.25, training accuracy is 0.187500, time 216.733 samples/sec
+epoch 3, total_step 55080, total loss is 22.63 , inference loss is 12.38, weight deacy loss is 10.25, training accuracy is 0.250000, time 211.951 samples/sec
+epoch 3, total_step 55100, total loss is 26.04 , inference loss is 15.79, weight deacy loss is 10.25, training accuracy is 0.250000, time 299.927 samples/sec
+epoch 3, total_step 55120, total loss is 22.10 , inference loss is 11.86, weight deacy loss is 10.25, training accuracy is 0.250000, time 204.530 samples/sec
+epoch 3, total_step 55140, total loss is 23.46 , inference loss is 13.21, weight deacy loss is 10.25, training accuracy is 0.156250, time 220.881 samples/sec
+epoch 3, total_step 55160, total loss is 25.91 , inference loss is 15.66, weight deacy loss is 10.25, training accuracy is 0.093750, time 215.376 samples/sec
+epoch 3, total_step 55180, total loss is 23.31 , inference loss is 13.07, weight deacy loss is 10.24, training accuracy is 0.187500, time 216.372 samples/sec
+epoch 3, total_step 55200, total loss is 22.04 , inference loss is 11.80, weight deacy loss is 10.24, training accuracy is 0.125000, time 12.168 samples/sec
+epoch 3, total_step 55220, total loss is 21.42 , inference loss is 11.18, weight deacy loss is 10.24, training accuracy is 0.218750, time 206.777 samples/sec
+epoch 3, total_step 55240, total loss is 26.29 , inference loss is 16.05, weight deacy loss is 10.24, training accuracy is 0.187500, time 204.868 samples/sec
+epoch 3, total_step 55260, total loss is 24.29 , inference loss is 14.05, weight deacy loss is 10.24, training accuracy is 0.250000, time 219.703 samples/sec
+epoch 3, total_step 55280, total loss is 20.80 , inference loss is 10.57, weight deacy loss is 10.24, training accuracy is 0.343750, time 215.197 samples/sec
+epoch 3, total_step 55300, total loss is 23.28 , inference loss is 13.05, weight deacy loss is 10.24, training accuracy is 0.093750, time 217.974 samples/sec
+epoch 3, total_step 55320, total loss is 20.86 , inference loss is 10.63, weight deacy loss is 10.24, training accuracy is 0.250000, time 220.219 samples/sec
+epoch 3, total_step 55340, total loss is 22.62 , inference loss is 12.39, weight deacy loss is 10.23, training accuracy is 0.218750, time 206.612 samples/sec
+epoch 3, total_step 55360, total loss is 21.88 , inference loss is 11.65, weight deacy loss is 10.23, training accuracy is 0.281250, time 212.205 samples/sec
+epoch 3, total_step 55380, total loss is 19.67 , inference loss is 9.44, weight deacy loss is 10.23, training accuracy is 0.218750, time 217.157 samples/sec
+epoch 3, total_step 55400, total loss is 20.06 , inference loss is 9.83, weight deacy loss is 10.23, training accuracy is 0.156250, time 213.266 samples/sec
+epoch 3, total_step 55420, total loss is 23.70 , inference loss is 13.47, weight deacy loss is 10.23, training accuracy is 0.156250, time 298.262 samples/sec
+epoch 3, total_step 55440, total loss is 21.38 , inference loss is 11.16, weight deacy loss is 10.23, training accuracy is 0.187500, time 208.930 samples/sec
+epoch 3, total_step 55460, total loss is 21.68 , inference loss is 11.45, weight deacy loss is 10.23, training accuracy is 0.281250, time 225.488 samples/sec
+epoch 3, total_step 55480, total loss is 20.34 , inference loss is 10.12, weight deacy loss is 10.23, training accuracy is 0.312500, time 214.900 samples/sec
+epoch 3, total_step 55500, total loss is 25.12 , inference loss is 14.90, weight deacy loss is 10.22, training accuracy is 0.125000, time 14.478 samples/sec
+epoch 3, total_step 55520, total loss is 20.27 , inference loss is 10.04, weight deacy loss is 10.22, training accuracy is 0.218750, time 214.573 samples/sec
+epoch 3, total_step 55540, total loss is 20.42 , inference loss is 10.20, weight deacy loss is 10.22, training accuracy is 0.187500, time 208.561 samples/sec
+epoch 3, total_step 55560, total loss is 24.58 , inference loss is 14.36, weight deacy loss is 10.22, training accuracy is 0.031250, time 215.061 samples/sec
+epoch 3, total_step 55580, total loss is 21.19 , inference loss is 10.97, weight deacy loss is 10.22, training accuracy is 0.312500, time 211.171 samples/sec
+epoch 3, total_step 55600, total loss is 22.08 , inference loss is 11.86, weight deacy loss is 10.22, training accuracy is 0.218750, time 207.322 samples/sec
+epoch 3, total_step 55620, total loss is 24.60 , inference loss is 14.39, weight deacy loss is 10.22, training accuracy is 0.156250, time 210.207 samples/sec
+epoch 3, total_step 55640, total loss is 24.26 , inference loss is 14.05, weight deacy loss is 10.22, training accuracy is 0.312500, time 220.057 samples/sec
+epoch 3, total_step 55660, total loss is 24.40 , inference loss is 14.19, weight deacy loss is 10.21, training accuracy is 0.156250, time 209.300 samples/sec
+epoch 3, total_step 55680, total loss is 22.21 , inference loss is 12.00, weight deacy loss is 10.21, training accuracy is 0.156250, time 216.722 samples/sec
+epoch 3, total_step 55700, total loss is 23.61 , inference loss is 13.39, weight deacy loss is 10.21, training accuracy is 0.312500, time 217.824 samples/sec
+epoch 3, total_step 55720, total loss is 20.35 , inference loss is 10.14, weight deacy loss is 10.21, training accuracy is 0.343750, time 210.072 samples/sec
+epoch 3, total_step 55740, total loss is 27.49 , inference loss is 17.28, weight deacy loss is 10.21, training accuracy is 0.062500, time 228.896 samples/sec
+epoch 3, total_step 55760, total loss is 21.10 , inference loss is 10.89, weight deacy loss is 10.21, training accuracy is 0.156250, time 217.389 samples/sec
+epoch 3, total_step 55780, total loss is 24.63 , inference loss is 14.42, weight deacy loss is 10.21, training accuracy is 0.218750, time 214.795 samples/sec
+epoch 3, total_step 55800, total loss is 21.57 , inference loss is 11.37, weight deacy loss is 10.21, training accuracy is 0.250000, time 12.431 samples/sec
+epoch 3, total_step 55820, total loss is 28.46 , inference loss is 18.25, weight deacy loss is 10.20, training accuracy is 0.187500, time 216.038 samples/sec
+epoch 3, total_step 55840, total loss is 24.28 , inference loss is 14.07, weight deacy loss is 10.20, training accuracy is 0.281250, time 215.529 samples/sec
+epoch 3, total_step 55860, total loss is 21.32 , inference loss is 11.12, weight deacy loss is 10.20, training accuracy is 0.187500, time 200.627 samples/sec
+epoch 3, total_step 55880, total loss is 24.44 , inference loss is 14.24, weight deacy loss is 10.20, training accuracy is 0.125000, time 209.322 samples/sec
+epoch 3, total_step 55900, total loss is 19.51 , inference loss is 9.31, weight deacy loss is 10.20, training accuracy is 0.406250, time 209.620 samples/sec
+epoch 3, total_step 55920, total loss is 26.26 , inference loss is 16.06, weight deacy loss is 10.20, training accuracy is 0.187500, time 217.474 samples/sec
+epoch 3, total_step 55940, total loss is 20.65 , inference loss is 10.45, weight deacy loss is 10.20, training accuracy is 0.250000, time 211.354 samples/sec
+epoch 3, total_step 55960, total loss is 19.23 , inference loss is 9.03, weight deacy loss is 10.20, training accuracy is 0.375000, time 213.944 samples/sec
+epoch 3, total_step 55980, total loss is 23.33 , inference loss is 13.14, weight deacy loss is 10.19, training accuracy is 0.218750, time 208.737 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.40075800000001
+best_threshold_index 107 0.9827777777777778
+best_threshold_index 107 0.9824074074074074
+best_threshold_index 107 0.9835185185185186
+best_threshold_index 107 0.9831481481481481
+best_threshold_index 108 0.9824074074074074
+best_threshold_index 107 0.9814814814814815
+best_threshold_index 107 0.9824074074074074
+best_threshold_index 107 0.9816666666666667
+best_threshold_index 107 0.9807407407407407
+best_threshold_index 107 0.9812962962962963
+[lfw][56000]XNorm: 24.309557
+[lfw][56000]Accuracy-Flip: 0.98183+-0.00751
+test accuracy is: 0.9818333333333333
+epoch 3, total_step 56000, total loss is 27.95 , inference loss is 17.76, weight deacy loss is 10.19, training accuracy is 0.250000, time 15.434 samples/sec
+epoch 3, total_step 56020, total loss is 22.03 , inference loss is 11.84, weight deacy loss is 10.19, training accuracy is 0.250000, time 208.128 samples/sec
+epoch 3, total_step 56040, total loss is 27.94 , inference loss is 17.75, weight deacy loss is 10.19, training accuracy is 0.156250, time 206.196 samples/sec
+epoch 3, total_step 56060, total loss is 18.14 , inference loss is 7.95, weight deacy loss is 10.19, training accuracy is 0.375000, time 213.745 samples/sec
+epoch 3, total_step 56080, total loss is 21.15 , inference loss is 10.96, weight deacy loss is 10.19, training accuracy is 0.281250, time 208.846 samples/sec
+epoch 3, total_step 56100, total loss is 21.45 , inference loss is 11.26, weight deacy loss is 10.19, training accuracy is 0.343750, time 12.493 samples/sec
+epoch 3, total_step 56120, total loss is 21.11 , inference loss is 10.93, weight deacy loss is 10.19, training accuracy is 0.406250, time 223.539 samples/sec
+epoch 3, total_step 56140, total loss is 22.07 , inference loss is 11.89, weight deacy loss is 10.18, training accuracy is 0.093750, time 212.550 samples/sec
+epoch 3, total_step 56160, total loss is 26.77 , inference loss is 16.59, weight deacy loss is 10.18, training accuracy is 0.218750, time 207.386 samples/sec
+epoch 3, total_step 56180, total loss is 26.43 , inference loss is 16.25, weight deacy loss is 10.18, training accuracy is 0.125000, time 225.705 samples/sec
+epoch 3, total_step 56200, total loss is 18.92 , inference loss is 8.74, weight deacy loss is 10.18, training accuracy is 0.343750, time 218.255 samples/sec
+epoch 3, total_step 56220, total loss is 23.29 , inference loss is 13.11, weight deacy loss is 10.18, training accuracy is 0.125000, time 212.813 samples/sec
+epoch 3, total_step 56240, total loss is 22.94 , inference loss is 12.77, weight deacy loss is 10.18, training accuracy is 0.218750, time 208.845 samples/sec
+epoch 3, total_step 56260, total loss is 21.08 , inference loss is 10.90, weight deacy loss is 10.18, training accuracy is 0.312500, time 204.745 samples/sec
+epoch 3, total_step 56280, total loss is 22.13 , inference loss is 11.95, weight deacy loss is 10.18, training accuracy is 0.156250, time 211.962 samples/sec
+epoch 3, total_step 56300, total loss is 19.64 , inference loss is 9.47, weight deacy loss is 10.17, training accuracy is 0.312500, time 218.307 samples/sec
+epoch 3, total_step 56320, total loss is 18.81 , inference loss is 8.64, weight deacy loss is 10.17, training accuracy is 0.375000, time 216.258 samples/sec
+epoch 3, total_step 56340, total loss is 26.43 , inference loss is 16.26, weight deacy loss is 10.17, training accuracy is 0.062500, time 205.202 samples/sec
+epoch 3, total_step 56360, total loss is 23.10 , inference loss is 12.93, weight deacy loss is 10.17, training accuracy is 0.218750, time 207.998 samples/sec
+epoch 3, total_step 56380, total loss is 20.25 , inference loss is 10.08, weight deacy loss is 10.17, training accuracy is 0.250000, time 212.841 samples/sec
+epoch 3, total_step 56400, total loss is 19.53 , inference loss is 9.36, weight deacy loss is 10.17, training accuracy is 0.218750, time 14.053 samples/sec
+epoch 3, total_step 56420, total loss is 25.73 , inference loss is 15.57, weight deacy loss is 10.17, training accuracy is 0.093750, time 295.701 samples/sec
+epoch 3, total_step 56440, total loss is 25.20 , inference loss is 15.03, weight deacy loss is 10.17, training accuracy is 0.312500, time 208.606 samples/sec
+epoch 3, total_step 56460, total loss is 22.67 , inference loss is 12.50, weight deacy loss is 10.16, training accuracy is 0.218750, time 231.120 samples/sec
+epoch 3, total_step 56480, total loss is 25.23 , inference loss is 15.07, weight deacy loss is 10.16, training accuracy is 0.218750, time 216.078 samples/sec
+epoch 3, total_step 56500, total loss is 21.81 , inference loss is 11.64, weight deacy loss is 10.16, training accuracy is 0.281250, time 211.928 samples/sec
+epoch 3, total_step 56520, total loss is 20.50 , inference loss is 10.34, weight deacy loss is 10.16, training accuracy is 0.375000, time 209.744 samples/sec
+epoch 3, total_step 56540, total loss is 20.56 , inference loss is 10.40, weight deacy loss is 10.16, training accuracy is 0.312500, time 210.240 samples/sec
+epoch 3, total_step 56560, total loss is 26.73 , inference loss is 16.58, weight deacy loss is 10.16, training accuracy is 0.093750, time 208.386 samples/sec
+epoch 3, total_step 56580, total loss is 26.65 , inference loss is 16.49, weight deacy loss is 10.16, training accuracy is 0.156250, time 215.379 samples/sec
+epoch 3, total_step 56600, total loss is 22.58 , inference loss is 12.42, weight deacy loss is 10.16, training accuracy is 0.250000, time 212.256 samples/sec
+epoch 3, total_step 56620, total loss is 20.93 , inference loss is 10.77, weight deacy loss is 10.15, training accuracy is 0.312500, time 209.335 samples/sec
+epoch 3, total_step 56640, total loss is 22.80 , inference loss is 12.64, weight deacy loss is 10.15, training accuracy is 0.218750, time 218.445 samples/sec
+epoch 3, total_step 56660, total loss is 24.28 , inference loss is 14.13, weight deacy loss is 10.15, training accuracy is 0.187500, time 215.994 samples/sec
+epoch 3, total_step 56680, total loss is 27.07 , inference loss is 16.92, weight deacy loss is 10.15, training accuracy is 0.187500, time 201.223 samples/sec
+epoch 3, total_step 56700, total loss is 25.04 , inference loss is 14.89, weight deacy loss is 10.15, training accuracy is 0.125000, time 14.578 samples/sec
+epoch 3, total_step 56720, total loss is 23.82 , inference loss is 13.67, weight deacy loss is 10.15, training accuracy is 0.281250, time 225.604 samples/sec
+epoch 3, total_step 56740, total loss is 20.59 , inference loss is 10.44, weight deacy loss is 10.15, training accuracy is 0.187500, time 214.258 samples/sec
+epoch 3, total_step 56760, total loss is 23.86 , inference loss is 13.71, weight deacy loss is 10.15, training accuracy is 0.156250, time 207.882 samples/sec
+epoch 3, total_step 56780, total loss is 28.02 , inference loss is 17.88, weight deacy loss is 10.14, training accuracy is 0.125000, time 217.876 samples/sec
+epoch 3, total_step 56800, total loss is 27.45 , inference loss is 17.31, weight deacy loss is 10.14, training accuracy is 0.093750, time 211.252 samples/sec
+epoch 3, total_step 56820, total loss is 22.29 , inference loss is 12.14, weight deacy loss is 10.14, training accuracy is 0.218750, time 210.081 samples/sec
+epoch 3, total_step 56840, total loss is 23.55 , inference loss is 13.41, weight deacy loss is 10.14, training accuracy is 0.187500, time 215.743 samples/sec
+epoch 3, total_step 56860, total loss is 21.92 , inference loss is 11.78, weight deacy loss is 10.14, training accuracy is 0.187500, time 244.003 samples/sec
+epoch 3, total_step 56880, total loss is 25.13 , inference loss is 15.00, weight deacy loss is 10.14, training accuracy is 0.125000, time 210.589 samples/sec
+epoch 3, total_step 56900, total loss is 27.86 , inference loss is 17.72, weight deacy loss is 10.14, training accuracy is 0.125000, time 203.979 samples/sec
+epoch 3, total_step 56920, total loss is 23.00 , inference loss is 12.87, weight deacy loss is 10.14, training accuracy is 0.218750, time 218.584 samples/sec
+epoch 3, total_step 56940, total loss is 22.05 , inference loss is 11.92, weight deacy loss is 10.14, training accuracy is 0.156250, time 214.970 samples/sec
+epoch 3, total_step 56960, total loss is 22.96 , inference loss is 12.83, weight deacy loss is 10.13, training accuracy is 0.218750, time 211.035 samples/sec
+epoch 3, total_step 56980, total loss is 19.67 , inference loss is 9.54, weight deacy loss is 10.13, training accuracy is 0.218750, time 214.573 samples/sec
+epoch 3, total_step 57000, total loss is 21.10 , inference loss is 10.97, weight deacy loss is 10.13, training accuracy is 0.218750, time 13.956 samples/sec
+epoch 3, total_step 57020, total loss is 21.46 , inference loss is 11.33, weight deacy loss is 10.13, training accuracy is 0.218750, time 212.179 samples/sec
+epoch 3, total_step 57040, total loss is 24.14 , inference loss is 14.01, weight deacy loss is 10.13, training accuracy is 0.218750, time 203.910 samples/sec
+epoch 3, total_step 57060, total loss is 23.14 , inference loss is 13.02, weight deacy loss is 10.13, training accuracy is 0.250000, time 211.323 samples/sec
+epoch 3, total_step 57080, total loss is 23.42 , inference loss is 13.29, weight deacy loss is 10.13, training accuracy is 0.125000, time 296.582 samples/sec
+epoch 3, total_step 57100, total loss is 21.88 , inference loss is 11.76, weight deacy loss is 10.13, training accuracy is 0.250000, time 215.587 samples/sec
+epoch 3, total_step 57120, total loss is 22.71 , inference loss is 12.59, weight deacy loss is 10.12, training accuracy is 0.187500, time 219.341 samples/sec
+epoch 3, total_step 57140, total loss is 27.46 , inference loss is 17.33, weight deacy loss is 10.12, training accuracy is 0.093750, time 210.312 samples/sec
+epoch 3, total_step 57160, total loss is 22.87 , inference loss is 12.75, weight deacy loss is 10.12, training accuracy is 0.250000, time 225.758 samples/sec
+epoch 3, total_step 57180, total loss is 25.38 , inference loss is 15.26, weight deacy loss is 10.12, training accuracy is 0.156250, time 213.356 samples/sec
+epoch 3, total_step 57200, total loss is 27.13 , inference loss is 17.01, weight deacy loss is 10.12, training accuracy is 0.187500, time 214.814 samples/sec
+epoch 3, total_step 57220, total loss is 23.86 , inference loss is 13.74, weight deacy loss is 10.12, training accuracy is 0.156250, time 211.684 samples/sec
+epoch 3, total_step 57240, total loss is 22.80 , inference loss is 12.68, weight deacy loss is 10.12, training accuracy is 0.281250, time 225.127 samples/sec
+epoch 3, total_step 57260, total loss is 23.22 , inference loss is 13.10, weight deacy loss is 10.12, training accuracy is 0.218750, time 298.980 samples/sec
+epoch 3, total_step 57280, total loss is 21.68 , inference loss is 11.56, weight deacy loss is 10.11, training accuracy is 0.218750, time 210.226 samples/sec
+epoch 3, total_step 57300, total loss is 22.50 , inference loss is 12.39, weight deacy loss is 10.11, training accuracy is 0.218750, time 14.115 samples/sec
+epoch 3, total_step 57320, total loss is 22.30 , inference loss is 12.19, weight deacy loss is 10.11, training accuracy is 0.250000, time 214.450 samples/sec
+epoch 3, total_step 57340, total loss is 23.15 , inference loss is 13.04, weight deacy loss is 10.11, training accuracy is 0.281250, time 220.652 samples/sec
+epoch 3, total_step 57360, total loss is 22.40 , inference loss is 12.29, weight deacy loss is 10.11, training accuracy is 0.187500, time 202.550 samples/sec
+epoch 3, total_step 57380, total loss is 26.51 , inference loss is 16.40, weight deacy loss is 10.11, training accuracy is 0.125000, time 218.540 samples/sec
+epoch 3, total_step 57400, total loss is 20.97 , inference loss is 10.86, weight deacy loss is 10.11, training accuracy is 0.125000, time 207.886 samples/sec
+epoch 3, total_step 57420, total loss is 25.09 , inference loss is 14.98, weight deacy loss is 10.11, training accuracy is 0.218750, time 208.673 samples/sec
+epoch 3, total_step 57440, total loss is 24.27 , inference loss is 14.16, weight deacy loss is 10.10, training accuracy is 0.156250, time 297.035 samples/sec
+epoch 3, total_step 57460, total loss is 25.03 , inference loss is 14.93, weight deacy loss is 10.10, training accuracy is 0.156250, time 211.489 samples/sec
+epoch 3, total_step 57480, total loss is 22.61 , inference loss is 12.50, weight deacy loss is 10.10, training accuracy is 0.343750, time 208.289 samples/sec
+epoch 3, total_step 57500, total loss is 22.35 , inference loss is 12.24, weight deacy loss is 10.10, training accuracy is 0.218750, time 215.125 samples/sec
+epoch 3, total_step 57520, total loss is 24.49 , inference loss is 14.39, weight deacy loss is 10.10, training accuracy is 0.187500, time 207.918 samples/sec
+epoch 3, total_step 57540, total loss is 21.97 , inference loss is 11.87, weight deacy loss is 10.10, training accuracy is 0.281250, time 220.018 samples/sec
+epoch 3, total_step 57560, total loss is 27.52 , inference loss is 17.42, weight deacy loss is 10.10, training accuracy is 0.218750, time 214.933 samples/sec
+epoch 3, total_step 57580, total loss is 22.38 , inference loss is 12.28, weight deacy loss is 10.10, training accuracy is 0.187500, time 215.984 samples/sec
+epoch 3, total_step 57600, total loss is 25.17 , inference loss is 15.07, weight deacy loss is 10.10, training accuracy is 0.125000, time 13.530 samples/sec
+epoch 3, total_step 57620, total loss is 21.72 , inference loss is 11.63, weight deacy loss is 10.09, training accuracy is 0.093750, time 212.542 samples/sec
+epoch 3, total_step 57640, total loss is 22.71 , inference loss is 12.62, weight deacy loss is 10.09, training accuracy is 0.343750, time 212.636 samples/sec
+epoch 3, total_step 57660, total loss is 24.24 , inference loss is 14.15, weight deacy loss is 10.09, training accuracy is 0.187500, time 206.430 samples/sec
+epoch 3, total_step 57680, total loss is 26.57 , inference loss is 16.48, weight deacy loss is 10.09, training accuracy is 0.250000, time 201.544 samples/sec
+epoch 3, total_step 57700, total loss is 23.17 , inference loss is 13.08, weight deacy loss is 10.09, training accuracy is 0.156250, time 299.500 samples/sec
+epoch 3, total_step 57720, total loss is 24.62 , inference loss is 14.53, weight deacy loss is 10.09, training accuracy is 0.218750, time 205.810 samples/sec
+epoch 3, total_step 57740, total loss is 25.80 , inference loss is 15.72, weight deacy loss is 10.09, training accuracy is 0.125000, time 214.190 samples/sec
+epoch 3, total_step 57760, total loss is 20.34 , inference loss is 10.25, weight deacy loss is 10.09, training accuracy is 0.312500, time 213.708 samples/sec
+epoch 3, total_step 57780, total loss is 22.04 , inference loss is 11.96, weight deacy loss is 10.08, training accuracy is 0.343750, time 212.810 samples/sec
+epoch 3, total_step 57800, total loss is 27.83 , inference loss is 17.75, weight deacy loss is 10.08, training accuracy is 0.062500, time 209.979 samples/sec
+epoch 3, total_step 57820, total loss is 30.03 , inference loss is 19.95, weight deacy loss is 10.08, training accuracy is 0.187500, time 199.471 samples/sec
+epoch 3, total_step 57840, total loss is 22.18 , inference loss is 12.10, weight deacy loss is 10.08, training accuracy is 0.187500, time 212.807 samples/sec
+epoch 3, total_step 57860, total loss is 21.70 , inference loss is 11.62, weight deacy loss is 10.08, training accuracy is 0.187500, time 209.018 samples/sec
+epoch 3, total_step 57880, total loss is 23.09 , inference loss is 13.01, weight deacy loss is 10.08, training accuracy is 0.250000, time 207.530 samples/sec
+epoch 3, total_step 57900, total loss is 25.01 , inference loss is 14.93, weight deacy loss is 10.08, training accuracy is 0.187500, time 12.366 samples/sec
+epoch 3, total_step 57920, total loss is 25.12 , inference loss is 15.04, weight deacy loss is 10.08, training accuracy is 0.125000, time 225.875 samples/sec
+epoch 3, total_step 57940, total loss is 23.14 , inference loss is 13.06, weight deacy loss is 10.07, training accuracy is 0.218750, time 209.292 samples/sec
+epoch 3, total_step 57960, total loss is 22.07 , inference loss is 11.99, weight deacy loss is 10.07, training accuracy is 0.250000, time 292.740 samples/sec
+epoch 3, total_step 57980, total loss is 24.20 , inference loss is 14.13, weight deacy loss is 10.07, training accuracy is 0.250000, time 202.021 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.88652300000002
+best_threshold_index 101 0.9764814814814815
+best_threshold_index 101 0.9761111111111112
+best_threshold_index 98 0.9772222222222222
+best_threshold_index 98 0.9762962962962963
+best_threshold_index 101 0.9764814814814815
+best_threshold_index 98 0.9748148148148148
+best_threshold_index 101 0.9762962962962963
+best_threshold_index 107 0.9748148148148148
+best_threshold_index 101 0.9737037037037037
+best_threshold_index 100 0.975
+[lfw][58000]XNorm: 24.264329
+[lfw][58000]Accuracy-Flip: 0.97367+-0.00933
+test accuracy is: 0.9736666666666668
+epoch 3, total_step 58000, total loss is 21.43 , inference loss is 11.36, weight deacy loss is 10.07, training accuracy is 0.218750, time 18.185 samples/sec
+epoch 3, total_step 58020, total loss is 24.50 , inference loss is 14.43, weight deacy loss is 10.07, training accuracy is 0.187500, time 219.217 samples/sec
+epoch 3, total_step 58040, total loss is 26.23 , inference loss is 16.16, weight deacy loss is 10.07, training accuracy is 0.187500, time 216.040 samples/sec
+epoch 3, total_step 58060, total loss is 22.01 , inference loss is 11.94, weight deacy loss is 10.07, training accuracy is 0.093750, time 196.815 samples/sec
+epoch 3, total_step 58080, total loss is 25.00 , inference loss is 14.93, weight deacy loss is 10.07, training accuracy is 0.156250, time 296.931 samples/sec
+epoch 3, total_step 58100, total loss is 22.13 , inference loss is 12.07, weight deacy loss is 10.07, training accuracy is 0.187500, time 204.410 samples/sec
+epoch 3, total_step 58120, total loss is 23.03 , inference loss is 12.97, weight deacy loss is 10.06, training accuracy is 0.156250, time 213.103 samples/sec
+epoch 3, total_step 58140, total loss is 25.05 , inference loss is 14.99, weight deacy loss is 10.06, training accuracy is 0.250000, time 217.499 samples/sec
+epoch 3, total_step 58160, total loss is 24.69 , inference loss is 14.63, weight deacy loss is 10.06, training accuracy is 0.375000, time 223.215 samples/sec
+epoch 3, total_step 58180, total loss is 22.12 , inference loss is 12.06, weight deacy loss is 10.06, training accuracy is 0.187500, time 208.316 samples/sec
+epoch 3, total_step 58200, total loss is 22.49 , inference loss is 12.43, weight deacy loss is 10.06, training accuracy is 0.187500, time 12.403 samples/sec
+epoch 3, total_step 58220, total loss is 20.30 , inference loss is 10.25, weight deacy loss is 10.06, training accuracy is 0.218750, time 208.029 samples/sec
+epoch 3, total_step 58240, total loss is 19.47 , inference loss is 9.41, weight deacy loss is 10.06, training accuracy is 0.312500, time 214.750 samples/sec
+epoch 3, total_step 58260, total loss is 23.14 , inference loss is 13.09, weight deacy loss is 10.06, training accuracy is 0.093750, time 192.744 samples/sec
+epoch 3, total_step 58280, total loss is 22.73 , inference loss is 12.68, weight deacy loss is 10.05, training accuracy is 0.250000, time 214.261 samples/sec
+epoch 3, total_step 58300, total loss is 21.93 , inference loss is 11.88, weight deacy loss is 10.05, training accuracy is 0.187500, time 223.546 samples/sec
+epoch 3, total_step 58320, total loss is 24.05 , inference loss is 14.00, weight deacy loss is 10.05, training accuracy is 0.218750, time 213.409 samples/sec
+epoch 3, total_step 58340, total loss is 22.02 , inference loss is 11.97, weight deacy loss is 10.05, training accuracy is 0.218750, time 297.394 samples/sec
+epoch 3, total_step 58360, total loss is 21.68 , inference loss is 11.63, weight deacy loss is 10.05, training accuracy is 0.343750, time 216.114 samples/sec
+epoch 3, total_step 58380, total loss is 20.31 , inference loss is 10.26, weight deacy loss is 10.05, training accuracy is 0.343750, time 218.765 samples/sec
+epoch 3, total_step 58400, total loss is 19.28 , inference loss is 9.24, weight deacy loss is 10.05, training accuracy is 0.187500, time 211.388 samples/sec
+epoch 3, total_step 58420, total loss is 23.07 , inference loss is 13.02, weight deacy loss is 10.05, training accuracy is 0.125000, time 217.216 samples/sec
+epoch 3, total_step 58440, total loss is 22.86 , inference loss is 12.81, weight deacy loss is 10.04, training accuracy is 0.187500, time 210.149 samples/sec
+epoch 3, total_step 58460, total loss is 21.37 , inference loss is 11.32, weight deacy loss is 10.04, training accuracy is 0.156250, time 205.114 samples/sec
+epoch 3, total_step 58480, total loss is 21.35 , inference loss is 11.31, weight deacy loss is 10.04, training accuracy is 0.218750, time 208.983 samples/sec
+epoch 3, total_step 58500, total loss is 23.86 , inference loss is 13.81, weight deacy loss is 10.04, training accuracy is 0.218750, time 13.584 samples/sec
+epoch 3, total_step 58520, total loss is 24.37 , inference loss is 14.33, weight deacy loss is 10.04, training accuracy is 0.093750, time 211.713 samples/sec
+epoch 3, total_step 58540, total loss is 21.54 , inference loss is 11.50, weight deacy loss is 10.04, training accuracy is 0.156250, time 208.733 samples/sec
+epoch 3, total_step 58560, total loss is 28.31 , inference loss is 18.27, weight deacy loss is 10.04, training accuracy is 0.062500, time 212.841 samples/sec
+epoch 3, total_step 58580, total loss is 22.87 , inference loss is 12.84, weight deacy loss is 10.04, training accuracy is 0.156250, time 203.965 samples/sec
+epoch 3, total_step 58600, total loss is 20.58 , inference loss is 10.54, weight deacy loss is 10.04, training accuracy is 0.375000, time 295.719 samples/sec
+epoch 3, total_step 58620, total loss is 19.22 , inference loss is 9.18, weight deacy loss is 10.03, training accuracy is 0.250000, time 208.358 samples/sec
+epoch 3, total_step 58640, total loss is 23.54 , inference loss is 13.51, weight deacy loss is 10.03, training accuracy is 0.156250, time 208.438 samples/sec
+epoch 3, total_step 58660, total loss is 22.35 , inference loss is 12.32, weight deacy loss is 10.03, training accuracy is 0.250000, time 215.171 samples/sec
+epoch 3, total_step 58680, total loss is 23.26 , inference loss is 13.23, weight deacy loss is 10.03, training accuracy is 0.031250, time 226.096 samples/sec
+epoch 3, total_step 58700, total loss is 23.63 , inference loss is 13.60, weight deacy loss is 10.03, training accuracy is 0.250000, time 207.309 samples/sec
+epoch 3, total_step 58720, total loss is 27.03 , inference loss is 17.01, weight deacy loss is 10.03, training accuracy is 0.031250, time 215.984 samples/sec
+epoch 3, total_step 58740, total loss is 27.52 , inference loss is 17.49, weight deacy loss is 10.03, training accuracy is 0.187500, time 224.984 samples/sec
+epoch 3, total_step 58760, total loss is 23.88 , inference loss is 13.85, weight deacy loss is 10.03, training accuracy is 0.187500, time 202.449 samples/sec
+epoch 3, total_step 58780, total loss is 26.39 , inference loss is 16.37, weight deacy loss is 10.02, training accuracy is 0.125000, time 218.542 samples/sec
+epoch 3, total_step 58800, total loss is 27.53 , inference loss is 17.51, weight deacy loss is 10.02, training accuracy is 0.062500, time 13.525 samples/sec
+epoch 3, total_step 58820, total loss is 21.82 , inference loss is 11.79, weight deacy loss is 10.02, training accuracy is 0.187500, time 211.635 samples/sec
+epoch 3, total_step 58840, total loss is 27.14 , inference loss is 17.11, weight deacy loss is 10.02, training accuracy is 0.187500, time 213.533 samples/sec
+epoch 3, total_step 58860, total loss is 25.07 , inference loss is 15.05, weight deacy loss is 10.02, training accuracy is 0.125000, time 207.844 samples/sec
+epoch 3, total_step 58880, total loss is 23.14 , inference loss is 13.12, weight deacy loss is 10.02, training accuracy is 0.218750, time 297.244 samples/sec
+epoch 3, total_step 58900, total loss is 26.42 , inference loss is 16.40, weight deacy loss is 10.02, training accuracy is 0.156250, time 221.923 samples/sec
+epoch 3, total_step 58920, total loss is 21.95 , inference loss is 11.94, weight deacy loss is 10.02, training accuracy is 0.093750, time 220.176 samples/sec
+epoch 3, total_step 58940, total loss is 25.37 , inference loss is 15.36, weight deacy loss is 10.02, training accuracy is 0.125000, time 221.835 samples/sec
+epoch 3, total_step 58960, total loss is 27.88 , inference loss is 17.87, weight deacy loss is 10.01, training accuracy is 0.125000, time 210.843 samples/sec
+epoch 3, total_step 58980, total loss is 24.51 , inference loss is 14.50, weight deacy loss is 10.01, training accuracy is 0.125000, time 212.147 samples/sec
+epoch 3, total_step 59000, total loss is 22.53 , inference loss is 12.52, weight deacy loss is 10.01, training accuracy is 0.187500, time 212.048 samples/sec
+epoch 3, total_step 59020, total loss is 26.37 , inference loss is 16.36, weight deacy loss is 10.01, training accuracy is 0.031250, time 219.529 samples/sec
+epoch 3, total_step 59040, total loss is 18.48 , inference loss is 8.47, weight deacy loss is 10.01, training accuracy is 0.281250, time 297.170 samples/sec
+epoch 3, total_step 59060, total loss is 25.69 , inference loss is 15.68, weight deacy loss is 10.01, training accuracy is 0.156250, time 297.186 samples/sec
+epoch 3, total_step 59080, total loss is 26.56 , inference loss is 16.55, weight deacy loss is 10.01, training accuracy is 0.125000, time 213.157 samples/sec
+epoch 3, total_step 59100, total loss is 24.17 , inference loss is 14.17, weight deacy loss is 10.01, training accuracy is 0.187500, time 13.702 samples/sec
+epoch 3, total_step 59120, total loss is 24.22 , inference loss is 14.22, weight deacy loss is 10.00, training accuracy is 0.187500, time 210.930 samples/sec
+epoch 3, total_step 59140, total loss is 22.57 , inference loss is 12.57, weight deacy loss is 10.00, training accuracy is 0.218750, time 225.598 samples/sec
+epoch 3, total_step 59160, total loss is 23.33 , inference loss is 13.32, weight deacy loss is 10.00, training accuracy is 0.125000, time 212.647 samples/sec
+epoch 3, total_step 59180, total loss is 23.90 , inference loss is 13.90, weight deacy loss is 10.00, training accuracy is 0.156250, time 212.396 samples/sec
+epoch 3, total_step 59200, total loss is 28.88 , inference loss is 18.88, weight deacy loss is 10.00, training accuracy is 0.125000, time 214.644 samples/sec
+epoch 3, total_step 59220, total loss is 24.80 , inference loss is 14.80, weight deacy loss is 10.00, training accuracy is 0.187500, time 212.395 samples/sec
+epoch 3, total_step 59240, total loss is 26.97 , inference loss is 16.97, weight deacy loss is 10.00, training accuracy is 0.156250, time 211.570 samples/sec
+epoch 3, total_step 59260, total loss is 28.22 , inference loss is 18.23, weight deacy loss is 10.00, training accuracy is 0.156250, time 208.127 samples/sec
+epoch 3, total_step 59280, total loss is 29.47 , inference loss is 19.48, weight deacy loss is 10.00, training accuracy is 0.031250, time 214.782 samples/sec
+epoch 3, total_step 59300, total loss is 23.58 , inference loss is 13.58, weight deacy loss is 9.99, training accuracy is 0.156250, time 214.735 samples/sec
+epoch 3, total_step 59320, total loss is 22.80 , inference loss is 12.81, weight deacy loss is 9.99, training accuracy is 0.187500, time 300.110 samples/sec
+epoch 3, total_step 59340, total loss is 27.57 , inference loss is 17.58, weight deacy loss is 9.99, training accuracy is 0.218750, time 212.221 samples/sec
+epoch 3, total_step 59360, total loss is 26.91 , inference loss is 16.92, weight deacy loss is 9.99, training accuracy is 0.125000, time 215.707 samples/sec
+epoch 3, total_step 59380, total loss is 27.33 , inference loss is 17.35, weight deacy loss is 9.99, training accuracy is 0.156250, time 207.383 samples/sec
+epoch 3, total_step 59400, total loss is 26.35 , inference loss is 16.36, weight deacy loss is 9.99, training accuracy is 0.093750, time 12.541 samples/sec
+epoch 3, total_step 59420, total loss is 25.53 , inference loss is 15.54, weight deacy loss is 9.99, training accuracy is 0.062500, time 212.918 samples/sec
+epoch 3, total_step 59440, total loss is 29.70 , inference loss is 19.72, weight deacy loss is 9.99, training accuracy is 0.062500, time 208.809 samples/sec
+epoch 3, total_step 59460, total loss is 29.55 , inference loss is 19.57, weight deacy loss is 9.98, training accuracy is 0.156250, time 219.398 samples/sec
+epoch 3, total_step 59480, total loss is 22.57 , inference loss is 12.59, weight deacy loss is 9.98, training accuracy is 0.218750, time 218.531 samples/sec
+epoch 3, total_step 59500, total loss is 26.30 , inference loss is 16.32, weight deacy loss is 9.98, training accuracy is 0.156250, time 207.338 samples/sec
+epoch 3, total_step 59520, total loss is 26.37 , inference loss is 16.39, weight deacy loss is 9.98, training accuracy is 0.218750, time 233.549 samples/sec
+epoch 3, total_step 59540, total loss is 23.14 , inference loss is 13.16, weight deacy loss is 9.98, training accuracy is 0.156250, time 207.260 samples/sec
+epoch 3, total_step 59560, total loss is 23.69 , inference loss is 13.72, weight deacy loss is 9.98, training accuracy is 0.125000, time 213.683 samples/sec
+epoch 3, total_step 59580, total loss is 25.99 , inference loss is 16.02, weight deacy loss is 9.98, training accuracy is 0.093750, time 181.078 samples/sec
+epoch 3, total_step 59600, total loss is 24.41 , inference loss is 14.43, weight deacy loss is 9.98, training accuracy is 0.250000, time 174.045 samples/sec
+epoch 3, total_step 59620, total loss is 21.15 , inference loss is 11.17, weight deacy loss is 9.98, training accuracy is 0.093750, time 209.710 samples/sec
+epoch 3, total_step 59640, total loss is 22.96 , inference loss is 12.99, weight deacy loss is 9.97, training accuracy is 0.156250, time 236.937 samples/sec
+epoch 3, total_step 59660, total loss is 21.46 , inference loss is 11.49, weight deacy loss is 9.97, training accuracy is 0.125000, time 220.536 samples/sec
+epoch 3, total_step 59680, total loss is 25.02 , inference loss is 15.05, weight deacy loss is 9.97, training accuracy is 0.187500, time 207.932 samples/sec
+epoch 3, total_step 59700, total loss is 23.19 , inference loss is 13.22, weight deacy loss is 9.97, training accuracy is 0.187500, time 14.152 samples/sec
+epoch 3, total_step 59720, total loss is 21.29 , inference loss is 11.32, weight deacy loss is 9.97, training accuracy is 0.218750, time 224.930 samples/sec
+epoch 3, total_step 59740, total loss is 24.61 , inference loss is 14.64, weight deacy loss is 9.97, training accuracy is 0.156250, time 208.543 samples/sec
+epoch 3, total_step 59760, total loss is 23.08 , inference loss is 13.11, weight deacy loss is 9.97, training accuracy is 0.156250, time 212.350 samples/sec
+epoch 3, total_step 59780, total loss is 25.28 , inference loss is 15.32, weight deacy loss is 9.97, training accuracy is 0.156250, time 206.517 samples/sec
+epoch 3, total_step 59800, total loss is 21.92 , inference loss is 11.96, weight deacy loss is 9.96, training accuracy is 0.218750, time 215.792 samples/sec
+epoch 3, total_step 59820, total loss is 25.18 , inference loss is 15.22, weight deacy loss is 9.96, training accuracy is 0.062500, time 210.341 samples/sec
+epoch 3, total_step 59840, total loss is 22.72 , inference loss is 12.76, weight deacy loss is 9.96, training accuracy is 0.156250, time 206.708 samples/sec
+epoch 3, total_step 59860, total loss is 18.44 , inference loss is 8.48, weight deacy loss is 9.96, training accuracy is 0.375000, time 297.194 samples/sec
+epoch 3, total_step 59880, total loss is 25.97 , inference loss is 16.01, weight deacy loss is 9.96, training accuracy is 0.125000, time 208.156 samples/sec
+epoch 3, total_step 59900, total loss is 24.84 , inference loss is 14.88, weight deacy loss is 9.96, training accuracy is 0.093750, time 209.414 samples/sec
+epoch 3, total_step 59920, total loss is 27.42 , inference loss is 17.46, weight deacy loss is 9.96, training accuracy is 0.031250, time 205.892 samples/sec
+epoch 3, total_step 59940, total loss is 21.14 , inference loss is 11.18, weight deacy loss is 9.96, training accuracy is 0.187500, time 213.452 samples/sec
+epoch 3, total_step 59960, total loss is 25.08 , inference loss is 15.13, weight deacy loss is 9.96, training accuracy is 0.218750, time 200.084 samples/sec
+epoch 3, total_step 59980, total loss is 27.10 , inference loss is 17.14, weight deacy loss is 9.95, training accuracy is 0.156250, time 208.825 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.201575999999996
+best_threshold_index 106 0.9796296296296296
+best_threshold_index 106 0.9790740740740741
+best_threshold_index 105 0.9801851851851852
+best_threshold_index 106 0.9805555555555555
+best_threshold_index 106 0.9790740740740741
+best_threshold_index 106 0.9777777777777777
+best_threshold_index 105 0.9788888888888889
+best_threshold_index 106 0.9779629629629629
+best_threshold_index 106 0.9775925925925926
+best_threshold_index 105 0.9779629629629629
+[lfw][60000]XNorm: 23.403148
+[lfw][60000]Accuracy-Flip: 0.97800+-0.00897
+test accuracy is: 0.978
+epoch 3, total_step 60000, total loss is 19.63 , inference loss is 9.68, weight deacy loss is 9.95, training accuracy is 0.125000, time 14.662 samples/sec
+epoch 3, total_step 60020, total loss is 22.09 , inference loss is 12.14, weight deacy loss is 9.95, training accuracy is 0.156250, time 218.549 samples/sec
+epoch 3, total_step 60040, total loss is 26.43 , inference loss is 16.48, weight deacy loss is 9.95, training accuracy is 0.062500, time 217.476 samples/sec
+epoch 3, total_step 60060, total loss is 23.67 , inference loss is 13.72, weight deacy loss is 9.95, training accuracy is 0.156250, time 222.073 samples/sec
+epoch 3, total_step 60080, total loss is 21.87 , inference loss is 11.92, weight deacy loss is 9.95, training accuracy is 0.281250, time 225.580 samples/sec
+epoch 3, total_step 60100, total loss is 24.97 , inference loss is 15.03, weight deacy loss is 9.95, training accuracy is 0.062500, time 218.698 samples/sec
+epoch 3, total_step 60120, total loss is 24.72 , inference loss is 14.77, weight deacy loss is 9.95, training accuracy is 0.187500, time 209.054 samples/sec
+epoch 3, total_step 60140, total loss is 27.26 , inference loss is 17.31, weight deacy loss is 9.94, training accuracy is 0.125000, time 215.499 samples/sec
+epoch 3, total_step 60160, total loss is 24.51 , inference loss is 14.57, weight deacy loss is 9.94, training accuracy is 0.125000, time 219.537 samples/sec
+epoch 3, total_step 60180, total loss is 22.24 , inference loss is 12.30, weight deacy loss is 9.94, training accuracy is 0.125000, time 214.643 samples/sec
+epoch 3, total_step 60200, total loss is 24.88 , inference loss is 14.93, weight deacy loss is 9.94, training accuracy is 0.125000, time 294.520 samples/sec
+epoch 3, total_step 60220, total loss is 22.58 , inference loss is 12.64, weight deacy loss is 9.94, training accuracy is 0.156250, time 216.597 samples/sec
+epoch 3, total_step 60240, total loss is 25.54 , inference loss is 15.60, weight deacy loss is 9.94, training accuracy is 0.125000, time 212.929 samples/sec
+epoch 3, total_step 60260, total loss is 22.45 , inference loss is 12.51, weight deacy loss is 9.94, training accuracy is 0.218750, time 210.500 samples/sec
+epoch 3, total_step 60280, total loss is 23.91 , inference loss is 13.98, weight deacy loss is 9.94, training accuracy is 0.062500, time 218.369 samples/sec
+epoch 3, total_step 60300, total loss is 26.84 , inference loss is 16.91, weight deacy loss is 9.94, training accuracy is 0.250000, time 12.361 samples/sec
+epoch 3, total_step 60320, total loss is 23.79 , inference loss is 13.85, weight deacy loss is 9.93, training accuracy is 0.156250, time 214.502 samples/sec
+epoch 3, total_step 60340, total loss is 20.18 , inference loss is 10.24, weight deacy loss is 9.93, training accuracy is 0.156250, time 211.835 samples/sec
+epoch 3, total_step 60360, total loss is 21.75 , inference loss is 11.82, weight deacy loss is 9.93, training accuracy is 0.125000, time 222.775 samples/sec
+epoch 3, total_step 60380, total loss is 24.77 , inference loss is 14.84, weight deacy loss is 9.93, training accuracy is 0.218750, time 211.461 samples/sec
+epoch 3, total_step 60400, total loss is 25.85 , inference loss is 15.92, weight deacy loss is 9.93, training accuracy is 0.062500, time 216.611 samples/sec
+epoch 3, total_step 60420, total loss is 24.07 , inference loss is 14.14, weight deacy loss is 9.93, training accuracy is 0.218750, time 213.761 samples/sec
+epoch 3, total_step 60440, total loss is 25.14 , inference loss is 15.21, weight deacy loss is 9.93, training accuracy is 0.093750, time 217.798 samples/sec
+epoch 3, total_step 60460, total loss is 26.88 , inference loss is 16.95, weight deacy loss is 9.93, training accuracy is 0.125000, time 209.637 samples/sec2022-12-01 04:14:57.484367: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 8645 of 10000
+2022-12-01 04:14:58.921808: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 3, total_step 60480, total loss is 27.26 , inference loss is 17.34, weight deacy loss is 9.93, training accuracy is 0.125000, time 211.701 samples/sec
+epoch 3, total_step 60500, total loss is 25.39 , inference loss is 15.47, weight deacy loss is 9.92, training accuracy is 0.062500, time 218.411 samples/sec
+epoch 3, total_step 60520, total loss is 25.95 , inference loss is 16.03, weight deacy loss is 9.92, training accuracy is 0.250000, time 206.010 samples/sec
+epoch 3, total_step 60540, total loss is 20.44 , inference loss is 10.52, weight deacy loss is 9.92, training accuracy is 0.156250, time 210.299 samples/sec
+epoch 3, total_step 60560, total loss is 25.03 , inference loss is 15.11, weight deacy loss is 9.92, training accuracy is 0.187500, time 210.450 samples/sec
+epoch 3, total_step 60580, total loss is 23.63 , inference loss is 13.71, weight deacy loss is 9.92, training accuracy is 0.125000, time 194.757 samples/sec
+epoch 3, total_step 60600, total loss is 25.32 , inference loss is 15.40, weight deacy loss is 9.92, training accuracy is 0.093750, time 13.587 samples/sec
+epoch 3, total_step 60620, total loss is 23.69 , inference loss is 13.78, weight deacy loss is 9.92, training accuracy is 0.156250, time 206.951 samples/sec
+epoch 3, total_step 60640, total loss is 22.12 , inference loss is 12.21, weight deacy loss is 9.92, training accuracy is 0.093750, time 297.691 samples/sec
+epoch 3, total_step 60660, total loss is 25.07 , inference loss is 15.15, weight deacy loss is 9.91, training accuracy is 0.250000, time 216.992 samples/sec
+epoch 3, total_step 60680, total loss is 25.76 , inference loss is 15.85, weight deacy loss is 9.91, training accuracy is 0.125000, time 219.451 samples/sec
+epoch 3, total_step 60700, total loss is 26.05 , inference loss is 16.14, weight deacy loss is 9.91, training accuracy is 0.250000, time 219.532 samples/sec
+epoch 3, total_step 60720, total loss is 23.87 , inference loss is 13.96, weight deacy loss is 9.91, training accuracy is 0.093750, time 216.809 samples/sec
+epoch 3, total_step 60740, total loss is 22.85 , inference loss is 12.94, weight deacy loss is 9.91, training accuracy is 0.218750, time 217.202 samples/sec
+epoch 3, total_step 60760, total loss is 19.75 , inference loss is 9.84, weight deacy loss is 9.91, training accuracy is 0.187500, time 215.753 samples/sec
+epoch 3, total_step 60780, total loss is 23.76 , inference loss is 13.86, weight deacy loss is 9.91, training accuracy is 0.156250, time 216.767 samples/sec
+epoch 3, total_step 60800, total loss is 24.42 , inference loss is 14.51, weight deacy loss is 9.91, training accuracy is 0.125000, time 297.548 samples/sec
+epoch 3, total_step 60820, total loss is 26.76 , inference loss is 16.86, weight deacy loss is 9.91, training accuracy is 0.125000, time 298.276 samples/sec
+epoch 3, total_step 60840, total loss is 20.98 , inference loss is 11.08, weight deacy loss is 9.90, training accuracy is 0.281250, time 212.261 samples/sec
+epoch 3, total_step 60860, total loss is 26.71 , inference loss is 16.81, weight deacy loss is 9.90, training accuracy is 0.156250, time 208.701 samples/sec
+epoch 3, total_step 60880, total loss is 23.02 , inference loss is 13.12, weight deacy loss is 9.90, training accuracy is 0.093750, time 209.290 samples/sec
+epoch 3, total_step 60900, total loss is 26.98 , inference loss is 17.08, weight deacy loss is 9.90, training accuracy is 0.093750, time 13.764 samples/sec
+epoch 3, total_step 60920, total loss is 22.87 , inference loss is 12.97, weight deacy loss is 9.90, training accuracy is 0.093750, time 204.375 samples/sec
+epoch 3, total_step 60940, total loss is 29.50 , inference loss is 19.60, weight deacy loss is 9.90, training accuracy is 0.062500, time 214.858 samples/sec
+epoch 3, total_step 60960, total loss is 24.51 , inference loss is 14.61, weight deacy loss is 9.90, training accuracy is 0.093750, time 216.091 samples/sec
+epoch 3, total_step 60980, total loss is 25.64 , inference loss is 15.74, weight deacy loss is 9.90, training accuracy is 0.062500, time 212.235 samples/sec
+epoch 3, total_step 61000, total loss is 20.64 , inference loss is 10.74, weight deacy loss is 9.90, training accuracy is 0.125000, time 207.772 samples/sec
+epoch 3, total_step 61020, total loss is 25.14 , inference loss is 15.25, weight deacy loss is 9.89, training accuracy is 0.093750, time 234.162 samples/sec
+epoch 3, total_step 61040, total loss is 24.15 , inference loss is 14.26, weight deacy loss is 9.89, training accuracy is 0.062500, time 193.567 samples/sec
+epoch 3, total_step 61060, total loss is 21.68 , inference loss is 11.78, weight deacy loss is 9.89, training accuracy is 0.187500, time 201.981 samples/sec
+epoch 3, total_step 61080, total loss is 24.56 , inference loss is 14.67, weight deacy loss is 9.89, training accuracy is 0.093750, time 194.187 samples/sec
+epoch 3, total_step 61100, total loss is 26.55 , inference loss is 16.66, weight deacy loss is 9.89, training accuracy is 0.031250, time 235.216 samples/sec
+epoch 3, total_step 61120, total loss is 29.13 , inference loss is 19.24, weight deacy loss is 9.89, training accuracy is 0.093750, time 213.667 samples/sec
+epoch 3, total_step 61140, total loss is 25.41 , inference loss is 15.52, weight deacy loss is 9.89, training accuracy is 0.031250, time 192.868 samples/sec
+epoch 3, total_step 61160, total loss is 21.58 , inference loss is 11.69, weight deacy loss is 9.89, training accuracy is 0.218750, time 222.225 samples/sec
+epoch 3, total_step 61180, total loss is 23.53 , inference loss is 13.64, weight deacy loss is 9.89, training accuracy is 0.125000, time 207.349 samples/sec
+epoch 3, total_step 61200, total loss is 22.68 , inference loss is 12.80, weight deacy loss is 9.88, training accuracy is 0.125000, time 14.015 samples/sec
+epoch 3, total_step 61220, total loss is 23.48 , inference loss is 13.59, weight deacy loss is 9.88, training accuracy is 0.125000, time 208.185 samples/sec
+epoch 3, total_step 61240, total loss is 24.91 , inference loss is 15.03, weight deacy loss is 9.88, training accuracy is 0.156250, time 175.062 samples/sec
+epoch 3, total_step 61260, total loss is 21.19 , inference loss is 11.31, weight deacy loss is 9.88, training accuracy is 0.187500, time 212.831 samples/sec
+epoch 3, total_step 61280, total loss is 23.40 , inference loss is 13.52, weight deacy loss is 9.88, training accuracy is 0.187500, time 209.331 samples/sec
+epoch 3, total_step 61300, total loss is 22.64 , inference loss is 12.76, weight deacy loss is 9.88, training accuracy is 0.218750, time 176.777 samples/sec
+epoch 3, total_step 61320, total loss is 24.02 , inference loss is 14.15, weight deacy loss is 9.88, training accuracy is 0.156250, time 200.805 samples/sec
+End of epoch 3
+epoch 4, total_step 61340, total loss is 26.94 , inference loss is 17.06, weight deacy loss is 9.88, training accuracy is 0.093750, time 211.942 samples/sec
+epoch 4, total_step 61360, total loss is 17.45 , inference loss is 7.58, weight deacy loss is 9.88, training accuracy is 0.406250, time 212.576 samples/sec
+epoch 4, total_step 61380, total loss is 19.58 , inference loss is 9.70, weight deacy loss is 9.87, training accuracy is 0.343750, time 208.945 samples/sec
+epoch 4, total_step 61400, total loss is 19.48 , inference loss is 9.60, weight deacy loss is 9.87, training accuracy is 0.312500, time 208.441 samples/sec
+epoch 4, total_step 61420, total loss is 17.10 , inference loss is 7.23, weight deacy loss is 9.87, training accuracy is 0.500000, time 211.913 samples/sec
+epoch 4, total_step 61440, total loss is 17.26 , inference loss is 7.39, weight deacy loss is 9.87, training accuracy is 0.562500, time 221.969 samples/sec
+epoch 4, total_step 61460, total loss is 19.47 , inference loss is 9.60, weight deacy loss is 9.87, training accuracy is 0.468750, time 211.368 samples/sec
+epoch 4, total_step 61480, total loss is 15.25 , inference loss is 5.38, weight deacy loss is 9.87, training accuracy is 0.687500, time 177.057 samples/sec
+epoch 4, total_step 61500, total loss is 19.82 , inference loss is 9.95, weight deacy loss is 9.87, training accuracy is 0.437500, time 13.610 samples/sec
+epoch 4, total_step 61520, total loss is 20.48 , inference loss is 10.62, weight deacy loss is 9.87, training accuracy is 0.406250, time 217.896 samples/sec
+epoch 4, total_step 61540, total loss is 18.45 , inference loss is 8.59, weight deacy loss is 9.86, training accuracy is 0.250000, time 211.065 samples/sec
+epoch 4, total_step 61560, total loss is 18.01 , inference loss is 8.15, weight deacy loss is 9.86, training accuracy is 0.468750, time 212.976 samples/sec
+epoch 4, total_step 61580, total loss is 17.27 , inference loss is 7.41, weight deacy loss is 9.86, training accuracy is 0.281250, time 216.948 samples/sec
+epoch 4, total_step 61600, total loss is 19.39 , inference loss is 9.53, weight deacy loss is 9.86, training accuracy is 0.312500, time 230.759 samples/sec
+epoch 4, total_step 61620, total loss is 22.01 , inference loss is 12.15, weight deacy loss is 9.86, training accuracy is 0.312500, time 223.154 samples/sec
+epoch 4, total_step 61640, total loss is 24.24 , inference loss is 14.38, weight deacy loss is 9.86, training accuracy is 0.218750, time 207.491 samples/sec
+epoch 4, total_step 61660, total loss is 20.75 , inference loss is 10.90, weight deacy loss is 9.86, training accuracy is 0.343750, time 208.771 samples/sec
+epoch 4, total_step 61680, total loss is 16.34 , inference loss is 6.48, weight deacy loss is 9.86, training accuracy is 0.437500, time 208.968 samples/sec
+epoch 4, total_step 61700, total loss is 24.07 , inference loss is 14.22, weight deacy loss is 9.85, training accuracy is 0.312500, time 209.478 samples/sec
+epoch 4, total_step 61720, total loss is 18.29 , inference loss is 8.44, weight deacy loss is 9.85, training accuracy is 0.375000, time 198.770 samples/sec
+epoch 4, total_step 61740, total loss is 21.60 , inference loss is 11.74, weight deacy loss is 9.85, training accuracy is 0.312500, time 295.416 samples/sec
+epoch 4, total_step 61760, total loss is 19.59 , inference loss is 9.74, weight deacy loss is 9.85, training accuracy is 0.312500, time 212.860 samples/sec
+epoch 4, total_step 61780, total loss is 21.27 , inference loss is 11.42, weight deacy loss is 9.85, training accuracy is 0.312500, time 208.538 samples/sec
+epoch 4, total_step 61800, total loss is 19.49 , inference loss is 9.64, weight deacy loss is 9.85, training accuracy is 0.437500, time 13.624 samples/sec
+epoch 4, total_step 61820, total loss is 19.82 , inference loss is 9.97, weight deacy loss is 9.85, training accuracy is 0.343750, time 212.118 samples/sec
+epoch 4, total_step 61840, total loss is 21.63 , inference loss is 11.78, weight deacy loss is 9.85, training accuracy is 0.468750, time 208.278 samples/sec
+epoch 4, total_step 61860, total loss is 18.93 , inference loss is 9.09, weight deacy loss is 9.84, training accuracy is 0.312500, time 212.528 samples/sec
+epoch 4, total_step 61880, total loss is 21.54 , inference loss is 11.69, weight deacy loss is 9.84, training accuracy is 0.312500, time 204.672 samples/sec
+epoch 4, total_step 61900, total loss is 20.56 , inference loss is 10.71, weight deacy loss is 9.84, training accuracy is 0.218750, time 220.112 samples/sec
+epoch 4, total_step 61920, total loss is 21.88 , inference loss is 12.03, weight deacy loss is 9.84, training accuracy is 0.343750, time 207.086 samples/sec
+epoch 4, total_step 61940, total loss is 22.37 , inference loss is 12.53, weight deacy loss is 9.84, training accuracy is 0.343750, time 213.446 samples/sec
+epoch 4, total_step 61960, total loss is 18.34 , inference loss is 8.50, weight deacy loss is 9.84, training accuracy is 0.312500, time 210.664 samples/sec
+epoch 4, total_step 61980, total loss is 24.57 , inference loss is 14.73, weight deacy loss is 9.84, training accuracy is 0.281250, time 210.146 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.426797999999998
+best_threshold_index 114 0.9816666666666667
+best_threshold_index 114 0.9811111111111112
+best_threshold_index 112 0.9814814814814815
+best_threshold_index 112 0.982037037037037
+best_threshold_index 112 0.9814814814814815
+best_threshold_index 114 0.9803703703703703
+best_threshold_index 112 0.9811111111111112
+best_threshold_index 112 0.9805555555555555
+best_threshold_index 114 0.9796296296296296
+best_threshold_index 114 0.98
+[lfw][62000]XNorm: 22.299559
+[lfw][62000]Accuracy-Flip: 0.97983+-0.00664
+test accuracy is: 0.9798333333333333
+epoch 4, total_step 62000, total loss is 22.91 , inference loss is 13.07, weight deacy loss is 9.84, training accuracy is 0.218750, time 18.168 samples/sec
+epoch 4, total_step 62020, total loss is 25.42 , inference loss is 15.58, weight deacy loss is 9.84, training accuracy is 0.343750, time 215.061 samples/sec
+epoch 4, total_step 62040, total loss is 21.24 , inference loss is 11.41, weight deacy loss is 9.83, training accuracy is 0.281250, time 199.880 samples/sec
+epoch 4, total_step 62060, total loss is 20.98 , inference loss is 11.14, weight deacy loss is 9.83, training accuracy is 0.250000, time 219.052 samples/sec
+epoch 4, total_step 62080, total loss is 19.24 , inference loss is 9.41, weight deacy loss is 9.83, training accuracy is 0.218750, time 214.041 samples/sec
+epoch 4, total_step 62100, total loss is 21.08 , inference loss is 11.25, weight deacy loss is 9.83, training accuracy is 0.343750, time 14.419 samples/sec
+epoch 4, total_step 62120, total loss is 16.19 , inference loss is 6.36, weight deacy loss is 9.83, training accuracy is 0.468750, time 195.992 samples/sec
+epoch 4, total_step 62140, total loss is 18.25 , inference loss is 8.42, weight deacy loss is 9.83, training accuracy is 0.343750, time 208.346 samples/sec
+epoch 4, total_step 62160, total loss is 16.36 , inference loss is 6.53, weight deacy loss is 9.83, training accuracy is 0.375000, time 208.483 samples/sec
+epoch 4, total_step 62180, total loss is 23.03 , inference loss is 13.20, weight deacy loss is 9.83, training accuracy is 0.406250, time 210.223 samples/sec
+epoch 4, total_step 62200, total loss is 21.95 , inference loss is 12.12, weight deacy loss is 9.83, training accuracy is 0.250000, time 213.391 samples/sec
+epoch 4, total_step 62220, total loss is 22.77 , inference loss is 12.95, weight deacy loss is 9.83, training accuracy is 0.218750, time 216.533 samples/sec
+epoch 4, total_step 62240, total loss is 20.76 , inference loss is 10.94, weight deacy loss is 9.82, training accuracy is 0.343750, time 209.255 samples/sec
+epoch 4, total_step 62260, total loss is 19.43 , inference loss is 9.60, weight deacy loss is 9.82, training accuracy is 0.218750, time 238.118 samples/sec
+epoch 4, total_step 62280, total loss is 16.85 , inference loss is 7.03, weight deacy loss is 9.82, training accuracy is 0.343750, time 208.142 samples/sec
+epoch 4, total_step 62300, total loss is 20.39 , inference loss is 10.57, weight deacy loss is 9.82, training accuracy is 0.281250, time 203.816 samples/sec
+epoch 4, total_step 62320, total loss is 19.07 , inference loss is 9.25, weight deacy loss is 9.82, training accuracy is 0.281250, time 207.264 samples/sec
+epoch 4, total_step 62340, total loss is 24.80 , inference loss is 14.99, weight deacy loss is 9.82, training accuracy is 0.156250, time 210.366 samples/sec
+epoch 4, total_step 62360, total loss is 20.06 , inference loss is 10.24, weight deacy loss is 9.82, training accuracy is 0.406250, time 210.898 samples/sec
+epoch 4, total_step 62380, total loss is 22.77 , inference loss is 12.95, weight deacy loss is 9.82, training accuracy is 0.281250, time 191.677 samples/sec
+epoch 4, total_step 62400, total loss is 21.25 , inference loss is 11.44, weight deacy loss is 9.82, training accuracy is 0.312500, time 13.272 samples/sec
+epoch 4, total_step 62420, total loss is 23.08 , inference loss is 13.26, weight deacy loss is 9.81, training accuracy is 0.156250, time 207.215 samples/sec
+epoch 4, total_step 62440, total loss is 25.15 , inference loss is 15.34, weight deacy loss is 9.81, training accuracy is 0.281250, time 216.377 samples/sec
+epoch 4, total_step 62460, total loss is 19.20 , inference loss is 9.39, weight deacy loss is 9.81, training accuracy is 0.406250, time 207.351 samples/sec
+epoch 4, total_step 62480, total loss is 19.88 , inference loss is 10.07, weight deacy loss is 9.81, training accuracy is 0.250000, time 220.584 samples/sec
+epoch 4, total_step 62500, total loss is 22.18 , inference loss is 12.37, weight deacy loss is 9.81, training accuracy is 0.250000, time 212.540 samples/sec
+epoch 4, total_step 62520, total loss is 26.61 , inference loss is 16.80, weight deacy loss is 9.81, training accuracy is 0.187500, time 214.416 samples/sec
+epoch 4, total_step 62540, total loss is 21.57 , inference loss is 11.77, weight deacy loss is 9.81, training accuracy is 0.250000, time 215.465 samples/sec
+epoch 4, total_step 62560, total loss is 19.47 , inference loss is 9.67, weight deacy loss is 9.81, training accuracy is 0.343750, time 214.283 samples/sec
+epoch 4, total_step 62580, total loss is 18.73 , inference loss is 8.92, weight deacy loss is 9.80, training accuracy is 0.437500, time 226.147 samples/sec
+epoch 4, total_step 62600, total loss is 22.79 , inference loss is 12.99, weight deacy loss is 9.80, training accuracy is 0.312500, time 298.949 samples/sec
+epoch 4, total_step 62620, total loss is 22.12 , inference loss is 12.32, weight deacy loss is 9.80, training accuracy is 0.312500, time 299.411 samples/sec
+epoch 4, total_step 62640, total loss is 22.53 , inference loss is 12.73, weight deacy loss is 9.80, training accuracy is 0.312500, time 209.284 samples/sec
+epoch 4, total_step 62660, total loss is 22.40 , inference loss is 12.60, weight deacy loss is 9.80, training accuracy is 0.281250, time 216.232 samples/sec
+epoch 4, total_step 62680, total loss is 18.68 , inference loss is 8.88, weight deacy loss is 9.80, training accuracy is 0.375000, time 212.060 samples/sec
+epoch 4, total_step 62700, total loss is 21.39 , inference loss is 11.60, weight deacy loss is 9.80, training accuracy is 0.343750, time 13.396 samples/sec
+epoch 4, total_step 62720, total loss is 23.03 , inference loss is 13.23, weight deacy loss is 9.80, training accuracy is 0.218750, time 214.281 samples/sec
+epoch 4, total_step 62740, total loss is 21.66 , inference loss is 11.86, weight deacy loss is 9.80, training accuracy is 0.250000, time 206.505 samples/sec
+epoch 4, total_step 62760, total loss is 23.10 , inference loss is 13.30, weight deacy loss is 9.79, training accuracy is 0.156250, time 209.957 samples/sec
+epoch 4, total_step 62780, total loss is 22.61 , inference loss is 12.81, weight deacy loss is 9.79, training accuracy is 0.156250, time 296.496 samples/sec
+epoch 4, total_step 62800, total loss is 29.28 , inference loss is 19.48, weight deacy loss is 9.79, training accuracy is 0.156250, time 211.169 samples/sec
+epoch 4, total_step 62820, total loss is 19.53 , inference loss is 9.73, weight deacy loss is 9.79, training accuracy is 0.343750, time 211.416 samples/sec
+epoch 4, total_step 62840, total loss is 24.59 , inference loss is 14.80, weight deacy loss is 9.79, training accuracy is 0.375000, time 217.032 samples/sec
+epoch 4, total_step 62860, total loss is 19.80 , inference loss is 10.01, weight deacy loss is 9.79, training accuracy is 0.218750, time 210.745 samples/sec
+epoch 4, total_step 62880, total loss is 20.05 , inference loss is 10.26, weight deacy loss is 9.79, training accuracy is 0.437500, time 215.418 samples/sec
+epoch 4, total_step 62900, total loss is 23.51 , inference loss is 13.72, weight deacy loss is 9.79, training accuracy is 0.250000, time 214.212 samples/sec
+epoch 4, total_step 62920, total loss is 18.34 , inference loss is 8.55, weight deacy loss is 9.79, training accuracy is 0.343750, time 216.227 samples/sec
+epoch 4, total_step 62940, total loss is 19.99 , inference loss is 10.21, weight deacy loss is 9.78, training accuracy is 0.343750, time 215.909 samples/sec
+epoch 4, total_step 62960, total loss is 21.78 , inference loss is 11.99, weight deacy loss is 9.78, training accuracy is 0.343750, time 210.301 samples/sec
+epoch 4, total_step 62980, total loss is 22.29 , inference loss is 12.51, weight deacy loss is 9.78, training accuracy is 0.375000, time 215.985 samples/sec
+epoch 4, total_step 63000, total loss is 18.80 , inference loss is 9.02, weight deacy loss is 9.78, training accuracy is 0.312500, time 13.682 samples/sec
+epoch 4, total_step 63020, total loss is 19.56 , inference loss is 9.78, weight deacy loss is 9.78, training accuracy is 0.343750, time 203.004 samples/sec
+epoch 4, total_step 63040, total loss is 23.76 , inference loss is 13.98, weight deacy loss is 9.78, training accuracy is 0.312500, time 213.533 samples/sec
+epoch 4, total_step 63060, total loss is 22.90 , inference loss is 13.12, weight deacy loss is 9.78, training accuracy is 0.218750, time 297.409 samples/sec
+epoch 4, total_step 63080, total loss is 23.50 , inference loss is 13.72, weight deacy loss is 9.78, training accuracy is 0.218750, time 211.851 samples/sec
+epoch 4, total_step 63100, total loss is 17.99 , inference loss is 8.21, weight deacy loss is 9.78, training accuracy is 0.468750, time 214.172 samples/sec
+epoch 4, total_step 63120, total loss is 25.32 , inference loss is 15.55, weight deacy loss is 9.77, training accuracy is 0.250000, time 202.374 samples/sec
+epoch 4, total_step 63140, total loss is 24.27 , inference loss is 14.50, weight deacy loss is 9.77, training accuracy is 0.250000, time 213.433 samples/sec
+epoch 4, total_step 63160, total loss is 23.98 , inference loss is 14.21, weight deacy loss is 9.77, training accuracy is 0.218750, time 201.255 samples/sec
+epoch 4, total_step 63180, total loss is 18.56 , inference loss is 8.79, weight deacy loss is 9.77, training accuracy is 0.406250, time 214.597 samples/sec
+epoch 4, total_step 63200, total loss is 20.93 , inference loss is 11.16, weight deacy loss is 9.77, training accuracy is 0.312500, time 200.699 samples/sec
+epoch 4, total_step 63220, total loss is 22.89 , inference loss is 13.12, weight deacy loss is 9.77, training accuracy is 0.375000, time 209.871 samples/sec
+epoch 4, total_step 63240, total loss is 20.05 , inference loss is 10.28, weight deacy loss is 9.77, training accuracy is 0.375000, time 226.224 samples/sec
+epoch 4, total_step 63260, total loss is 24.05 , inference loss is 14.29, weight deacy loss is 9.77, training accuracy is 0.218750, time 205.928 samples/sec
+epoch 4, total_step 63280, total loss is 17.85 , inference loss is 8.09, weight deacy loss is 9.77, training accuracy is 0.312500, time 210.297 samples/sec
+epoch 4, total_step 63300, total loss is 21.24 , inference loss is 11.47, weight deacy loss is 9.76, training accuracy is 0.343750, time 14.283 samples/sec
+epoch 4, total_step 63320, total loss is 17.94 , inference loss is 8.18, weight deacy loss is 9.76, training accuracy is 0.343750, time 298.546 samples/sec
+epoch 4, total_step 63340, total loss is 19.24 , inference loss is 9.48, weight deacy loss is 9.76, training accuracy is 0.406250, time 224.510 samples/sec
+epoch 4, total_step 63360, total loss is 16.39 , inference loss is 6.63, weight deacy loss is 9.76, training accuracy is 0.406250, time 201.139 samples/sec
+epoch 4, total_step 63380, total loss is 16.73 , inference loss is 6.98, weight deacy loss is 9.76, training accuracy is 0.531250, time 199.449 samples/sec
+epoch 4, total_step 63400, total loss is 22.16 , inference loss is 12.40, weight deacy loss is 9.76, training accuracy is 0.406250, time 206.745 samples/sec
+epoch 4, total_step 63420, total loss is 18.43 , inference loss is 8.67, weight deacy loss is 9.76, training accuracy is 0.375000, time 205.672 samples/sec
+epoch 4, total_step 63440, total loss is 20.83 , inference loss is 11.08, weight deacy loss is 9.76, training accuracy is 0.281250, time 209.712 samples/sec
+epoch 4, total_step 63460, total loss is 23.16 , inference loss is 13.41, weight deacy loss is 9.75, training accuracy is 0.281250, time 217.700 samples/sec
+epoch 4, total_step 63480, total loss is 17.19 , inference loss is 7.44, weight deacy loss is 9.75, training accuracy is 0.406250, time 207.068 samples/sec
+epoch 4, total_step 63500, total loss is 20.97 , inference loss is 11.22, weight deacy loss is 9.75, training accuracy is 0.437500, time 213.919 samples/sec
+epoch 4, total_step 63520, total loss is 22.71 , inference loss is 12.96, weight deacy loss is 9.75, training accuracy is 0.343750, time 241.451 samples/sec
+epoch 4, total_step 63540, total loss is 21.26 , inference loss is 11.51, weight deacy loss is 9.75, training accuracy is 0.312500, time 209.458 samples/sec
+epoch 4, total_step 63560, total loss is 19.11 , inference loss is 9.37, weight deacy loss is 9.75, training accuracy is 0.312500, time 198.110 samples/sec
+epoch 4, total_step 63580, total loss is 17.94 , inference loss is 8.19, weight deacy loss is 9.75, training accuracy is 0.406250, time 228.866 samples/sec
+epoch 4, total_step 63600, total loss is 19.04 , inference loss is 9.29, weight deacy loss is 9.75, training accuracy is 0.281250, time 14.920 samples/sec
+epoch 4, total_step 63620, total loss is 19.01 , inference loss is 9.27, weight deacy loss is 9.74, training accuracy is 0.281250, time 215.408 samples/sec
+epoch 4, total_step 63640, total loss is 17.48 , inference loss is 7.74, weight deacy loss is 9.74, training accuracy is 0.375000, time 214.050 samples/sec
+epoch 4, total_step 63660, total loss is 18.42 , inference loss is 8.68, weight deacy loss is 9.74, training accuracy is 0.468750, time 212.566 samples/sec
+epoch 4, total_step 63680, total loss is 18.73 , inference loss is 8.99, weight deacy loss is 9.74, training accuracy is 0.375000, time 234.473 samples/sec
+epoch 4, total_step 63700, total loss is 22.47 , inference loss is 12.73, weight deacy loss is 9.74, training accuracy is 0.312500, time 208.767 samples/sec
+epoch 4, total_step 63720, total loss is 20.76 , inference loss is 11.02, weight deacy loss is 9.74, training accuracy is 0.375000, time 210.225 samples/sec
+epoch 4, total_step 63740, total loss is 17.75 , inference loss is 8.01, weight deacy loss is 9.74, training accuracy is 0.343750, time 210.540 samples/sec
+epoch 4, total_step 63760, total loss is 18.28 , inference loss is 8.54, weight deacy loss is 9.74, training accuracy is 0.312500, time 209.554 samples/sec
+epoch 4, total_step 63780, total loss is 20.61 , inference loss is 10.88, weight deacy loss is 9.74, training accuracy is 0.343750, time 203.396 samples/sec
+epoch 4, total_step 63800, total loss is 20.32 , inference loss is 10.59, weight deacy loss is 9.73, training accuracy is 0.312500, time 197.393 samples/sec
+epoch 4, total_step 63820, total loss is 22.41 , inference loss is 12.67, weight deacy loss is 9.73, training accuracy is 0.281250, time 213.056 samples/sec
+epoch 4, total_step 63840, total loss is 20.41 , inference loss is 10.67, weight deacy loss is 9.73, training accuracy is 0.406250, time 216.938 samples/sec
+epoch 4, total_step 63860, total loss is 17.85 , inference loss is 8.12, weight deacy loss is 9.73, training accuracy is 0.500000, time 211.206 samples/sec
+epoch 4, total_step 63880, total loss is 16.16 , inference loss is 6.43, weight deacy loss is 9.73, training accuracy is 0.500000, time 233.617 samples/sec
+epoch 4, total_step 63900, total loss is 23.58 , inference loss is 13.85, weight deacy loss is 9.73, training accuracy is 0.187500, time 12.742 samples/sec
+epoch 4, total_step 63920, total loss is 21.01 , inference loss is 11.28, weight deacy loss is 9.73, training accuracy is 0.281250, time 213.175 samples/sec
+epoch 4, total_step 63940, total loss is 17.64 , inference loss is 7.92, weight deacy loss is 9.73, training accuracy is 0.250000, time 213.201 samples/sec
+epoch 4, total_step 63960, total loss is 22.47 , inference loss is 12.75, weight deacy loss is 9.72, training accuracy is 0.250000, time 215.920 samples/sec
+epoch 4, total_step 63980, total loss is 21.78 , inference loss is 12.06, weight deacy loss is 9.72, training accuracy is 0.281250, time 209.224 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.74967400000001
+best_threshold_index 115 0.9838888888888889
+best_threshold_index 115 0.9840740740740741
+best_threshold_index 115 0.9838888888888889
+best_threshold_index 115 0.9838888888888889
+best_threshold_index 115 0.9844444444444445
+best_threshold_index 115 0.9827777777777778
+best_threshold_index 115 0.9831481481481481
+best_threshold_index 115 0.9827777777777778
+best_threshold_index 115 0.9825925925925926
+best_threshold_index 115 0.9835185185185186
+[lfw][64000]XNorm: 23.180064
+[lfw][64000]Accuracy-Flip: 0.98350+-0.00545
+test accuracy is: 0.9834999999999999
+epoch 4, total_step 64000, total loss is 15.52 , inference loss is 5.80, weight deacy loss is 9.72, training accuracy is 0.437500, time 17.871 samples/sec
+epoch 4, total_step 64020, total loss is 18.47 , inference loss is 8.75, weight deacy loss is 9.72, training accuracy is 0.343750, time 209.192 samples/sec
+epoch 4, total_step 64040, total loss is 17.31 , inference loss is 7.59, weight deacy loss is 9.72, training accuracy is 0.406250, time 217.173 samples/sec
+epoch 4, total_step 64060, total loss is 16.99 , inference loss is 7.27, weight deacy loss is 9.72, training accuracy is 0.437500, time 228.496 samples/sec
+epoch 4, total_step 64080, total loss is 18.47 , inference loss is 8.75, weight deacy loss is 9.72, training accuracy is 0.375000, time 223.244 samples/sec
+epoch 4, total_step 64100, total loss is 17.82 , inference loss is 8.10, weight deacy loss is 9.72, training accuracy is 0.437500, time 205.767 samples/sec
+epoch 4, total_step 64120, total loss is 21.79 , inference loss is 12.08, weight deacy loss is 9.72, training accuracy is 0.375000, time 221.285 samples/sec
+epoch 4, total_step 64140, total loss is 16.88 , inference loss is 7.16, weight deacy loss is 9.71, training accuracy is 0.218750, time 222.483 samples/sec
+epoch 4, total_step 64160, total loss is 21.46 , inference loss is 11.75, weight deacy loss is 9.71, training accuracy is 0.312500, time 216.536 samples/sec
+epoch 4, total_step 64180, total loss is 17.83 , inference loss is 8.12, weight deacy loss is 9.71, training accuracy is 0.312500, time 211.042 samples/sec
+epoch 4, total_step 64200, total loss is 20.17 , inference loss is 10.46, weight deacy loss is 9.71, training accuracy is 0.343750, time 13.868 samples/sec
+epoch 4, total_step 64220, total loss is 18.71 , inference loss is 9.00, weight deacy loss is 9.71, training accuracy is 0.312500, time 215.556 samples/sec
+epoch 4, total_step 64240, total loss is 17.57 , inference loss is 7.86, weight deacy loss is 9.71, training accuracy is 0.312500, time 214.313 samples/sec
+epoch 4, total_step 64260, total loss is 21.62 , inference loss is 11.91, weight deacy loss is 9.71, training accuracy is 0.343750, time 217.384 samples/sec
+epoch 4, total_step 64280, total loss is 17.78 , inference loss is 8.08, weight deacy loss is 9.71, training accuracy is 0.437500, time 217.594 samples/sec
+epoch 4, total_step 64300, total loss is 20.66 , inference loss is 10.96, weight deacy loss is 9.70, training accuracy is 0.343750, time 217.395 samples/sec
+epoch 4, total_step 64320, total loss is 21.42 , inference loss is 11.72, weight deacy loss is 9.70, training accuracy is 0.343750, time 213.830 samples/sec
+epoch 4, total_step 64340, total loss is 24.58 , inference loss is 14.88, weight deacy loss is 9.70, training accuracy is 0.187500, time 226.881 samples/sec
+epoch 4, total_step 64360, total loss is 19.12 , inference loss is 9.42, weight deacy loss is 9.70, training accuracy is 0.406250, time 297.941 samples/sec
+epoch 4, total_step 64380, total loss is 23.29 , inference loss is 13.59, weight deacy loss is 9.70, training accuracy is 0.187500, time 295.467 samples/sec
+epoch 4, total_step 64400, total loss is 18.82 , inference loss is 9.12, weight deacy loss is 9.70, training accuracy is 0.468750, time 211.728 samples/sec
+epoch 4, total_step 64420, total loss is 19.10 , inference loss is 9.41, weight deacy loss is 9.70, training accuracy is 0.406250, time 214.025 samples/sec
+epoch 4, total_step 64440, total loss is 17.83 , inference loss is 8.14, weight deacy loss is 9.70, training accuracy is 0.468750, time 207.810 samples/sec
+epoch 4, total_step 64460, total loss is 18.84 , inference loss is 9.14, weight deacy loss is 9.70, training accuracy is 0.468750, time 217.373 samples/sec
+epoch 4, total_step 64480, total loss is 19.43 , inference loss is 9.73, weight deacy loss is 9.69, training accuracy is 0.312500, time 212.129 samples/sec
+epoch 4, total_step 64500, total loss is 19.81 , inference loss is 10.12, weight deacy loss is 9.69, training accuracy is 0.343750, time 14.144 samples/sec
+epoch 4, total_step 64520, total loss is 19.57 , inference loss is 9.88, weight deacy loss is 9.69, training accuracy is 0.312500, time 209.497 samples/sec
+epoch 4, total_step 64540, total loss is 18.92 , inference loss is 9.23, weight deacy loss is 9.69, training accuracy is 0.312500, time 220.062 samples/sec
+epoch 4, total_step 64560, total loss is 21.42 , inference loss is 11.73, weight deacy loss is 9.69, training accuracy is 0.218750, time 204.415 samples/sec
+epoch 4, total_step 64580, total loss is 20.71 , inference loss is 11.02, weight deacy loss is 9.69, training accuracy is 0.468750, time 202.261 samples/sec
+epoch 4, total_step 64600, total loss is 19.90 , inference loss is 10.21, weight deacy loss is 9.69, training accuracy is 0.375000, time 206.869 samples/sec
+epoch 4, total_step 64620, total loss is 21.35 , inference loss is 11.66, weight deacy loss is 9.69, training accuracy is 0.375000, time 217.862 samples/sec
+epoch 4, total_step 64640, total loss is 22.80 , inference loss is 13.12, weight deacy loss is 9.69, training accuracy is 0.250000, time 298.240 samples/sec
+epoch 4, total_step 64660, total loss is 21.20 , inference loss is 11.52, weight deacy loss is 9.68, training accuracy is 0.281250, time 203.639 samples/sec
+epoch 4, total_step 64680, total loss is 19.90 , inference loss is 10.22, weight deacy loss is 9.68, training accuracy is 0.343750, time 223.729 samples/sec
+epoch 4, total_step 64700, total loss is 20.02 , inference loss is 10.34, weight deacy loss is 9.68, training accuracy is 0.125000, time 214.131 samples/sec
+epoch 4, total_step 64720, total loss is 24.48 , inference loss is 14.80, weight deacy loss is 9.68, training accuracy is 0.125000, time 206.579 samples/sec
+epoch 4, total_step 64740, total loss is 20.25 , inference loss is 10.57, weight deacy loss is 9.68, training accuracy is 0.281250, time 207.031 samples/sec
+epoch 4, total_step 64760, total loss is 17.11 , inference loss is 7.43, weight deacy loss is 9.68, training accuracy is 0.375000, time 217.377 samples/sec
+epoch 4, total_step 64780, total loss is 22.46 , inference loss is 12.79, weight deacy loss is 9.68, training accuracy is 0.281250, time 209.874 samples/sec
+epoch 4, total_step 64800, total loss is 20.65 , inference loss is 10.97, weight deacy loss is 9.68, training accuracy is 0.218750, time 13.322 samples/sec
+epoch 4, total_step 64820, total loss is 22.92 , inference loss is 13.24, weight deacy loss is 9.68, training accuracy is 0.281250, time 221.055 samples/sec
+epoch 4, total_step 64840, total loss is 17.57 , inference loss is 7.90, weight deacy loss is 9.67, training accuracy is 0.500000, time 210.182 samples/sec
+epoch 4, total_step 64860, total loss is 21.35 , inference loss is 11.68, weight deacy loss is 9.67, training accuracy is 0.312500, time 206.424 samples/sec
+epoch 4, total_step 64880, total loss is 20.21 , inference loss is 10.53, weight deacy loss is 9.67, training accuracy is 0.218750, time 210.318 samples/sec
+epoch 4, total_step 64900, total loss is 18.93 , inference loss is 9.26, weight deacy loss is 9.67, training accuracy is 0.281250, time 297.657 samples/sec
+epoch 4, total_step 64920, total loss is 19.90 , inference loss is 10.23, weight deacy loss is 9.67, training accuracy is 0.343750, time 198.442 samples/sec
+epoch 4, total_step 64940, total loss is 25.05 , inference loss is 15.38, weight deacy loss is 9.67, training accuracy is 0.281250, time 211.070 samples/sec
+epoch 4, total_step 64960, total loss is 25.48 , inference loss is 15.82, weight deacy loss is 9.67, training accuracy is 0.218750, time 211.709 samples/sec
+epoch 4, total_step 64980, total loss is 23.34 , inference loss is 13.68, weight deacy loss is 9.67, training accuracy is 0.281250, time 213.810 samples/sec
+epoch 4, total_step 65000, total loss is 21.48 , inference loss is 11.81, weight deacy loss is 9.67, training accuracy is 0.343750, time 203.343 samples/sec
+epoch 4, total_step 65020, total loss is 26.05 , inference loss is 16.38, weight deacy loss is 9.66, training accuracy is 0.187500, time 209.669 samples/sec
+epoch 4, total_step 65040, total loss is 22.23 , inference loss is 12.56, weight deacy loss is 9.66, training accuracy is 0.343750, time 204.480 samples/sec
+epoch 4, total_step 65060, total loss is 22.43 , inference loss is 12.77, weight deacy loss is 9.66, training accuracy is 0.343750, time 208.159 samples/sec
+epoch 4, total_step 65080, total loss is 26.69 , inference loss is 17.03, weight deacy loss is 9.66, training accuracy is 0.187500, time 215.392 samples/sec
+epoch 4, total_step 65100, total loss is 22.07 , inference loss is 12.41, weight deacy loss is 9.66, training accuracy is 0.343750, time 13.782 samples/sec
+epoch 4, total_step 65120, total loss is 24.73 , inference loss is 15.08, weight deacy loss is 9.66, training accuracy is 0.250000, time 209.552 samples/sec
+epoch 4, total_step 65140, total loss is 20.71 , inference loss is 11.05, weight deacy loss is 9.66, training accuracy is 0.250000, time 213.509 samples/sec
+epoch 4, total_step 65160, total loss is 20.26 , inference loss is 10.60, weight deacy loss is 9.66, training accuracy is 0.250000, time 214.346 samples/sec
+epoch 4, total_step 65180, total loss is 21.12 , inference loss is 11.46, weight deacy loss is 9.66, training accuracy is 0.281250, time 298.059 samples/sec
+epoch 4, total_step 65200, total loss is 20.88 , inference loss is 11.22, weight deacy loss is 9.65, training accuracy is 0.218750, time 224.429 samples/sec
+epoch 4, total_step 65220, total loss is 19.28 , inference loss is 9.62, weight deacy loss is 9.65, training accuracy is 0.437500, time 210.830 samples/sec
+epoch 4, total_step 65240, total loss is 20.39 , inference loss is 10.74, weight deacy loss is 9.65, training accuracy is 0.250000, time 208.732 samples/sec
+epoch 4, total_step 65260, total loss is 21.97 , inference loss is 12.32, weight deacy loss is 9.65, training accuracy is 0.218750, time 212.777 samples/sec
+epoch 4, total_step 65280, total loss is 31.48 , inference loss is 21.83, weight deacy loss is 9.65, training accuracy is 0.125000, time 205.903 samples/sec
+epoch 4, total_step 65300, total loss is 25.09 , inference loss is 15.44, weight deacy loss is 9.65, training accuracy is 0.187500, time 197.701 samples/sec
+epoch 4, total_step 65320, total loss is 25.54 , inference loss is 15.90, weight deacy loss is 9.65, training accuracy is 0.125000, time 213.033 samples/sec
+epoch 4, total_step 65340, total loss is 24.54 , inference loss is 14.89, weight deacy loss is 9.65, training accuracy is 0.250000, time 214.175 samples/sec
+epoch 4, total_step 65360, total loss is 24.52 , inference loss is 14.88, weight deacy loss is 9.65, training accuracy is 0.093750, time 202.081 samples/sec
+epoch 4, total_step 65380, total loss is 22.18 , inference loss is 12.53, weight deacy loss is 9.64, training accuracy is 0.312500, time 209.836 samples/sec
+epoch 4, total_step 65400, total loss is 22.87 , inference loss is 13.23, weight deacy loss is 9.64, training accuracy is 0.250000, time 13.961 samples/sec
+epoch 4, total_step 65420, total loss is 22.22 , inference loss is 12.58, weight deacy loss is 9.64, training accuracy is 0.281250, time 211.097 samples/sec
+epoch 4, total_step 65440, total loss is 21.23 , inference loss is 11.59, weight deacy loss is 9.64, training accuracy is 0.218750, time 301.172 samples/sec
+epoch 4, total_step 65460, total loss is 20.53 , inference loss is 10.89, weight deacy loss is 9.64, training accuracy is 0.281250, time 204.257 samples/sec
+epoch 4, total_step 65480, total loss is 20.07 , inference loss is 10.43, weight deacy loss is 9.64, training accuracy is 0.250000, time 203.935 samples/sec
+epoch 4, total_step 65500, total loss is 21.21 , inference loss is 11.57, weight deacy loss is 9.64, training accuracy is 0.218750, time 204.830 samples/sec
+epoch 4, total_step 65520, total loss is 24.94 , inference loss is 15.31, weight deacy loss is 9.64, training accuracy is 0.187500, time 211.599 samples/sec
+epoch 4, total_step 65540, total loss is 22.84 , inference loss is 13.20, weight deacy loss is 9.64, training accuracy is 0.250000, time 218.844 samples/sec
+epoch 4, total_step 65560, total loss is 23.43 , inference loss is 13.80, weight deacy loss is 9.63, training accuracy is 0.187500, time 213.366 samples/sec
+epoch 4, total_step 65580, total loss is 22.42 , inference loss is 12.79, weight deacy loss is 9.63, training accuracy is 0.406250, time 210.610 samples/sec
+epoch 4, total_step 65600, total loss is 21.19 , inference loss is 11.56, weight deacy loss is 9.63, training accuracy is 0.343750, time 210.618 samples/sec
+epoch 4, total_step 65620, total loss is 24.80 , inference loss is 15.17, weight deacy loss is 9.63, training accuracy is 0.187500, time 220.479 samples/sec
+epoch 4, total_step 65640, total loss is 18.94 , inference loss is 9.31, weight deacy loss is 9.63, training accuracy is 0.375000, time 212.361 samples/sec
+epoch 4, total_step 65660, total loss is 24.50 , inference loss is 14.87, weight deacy loss is 9.63, training accuracy is 0.312500, time 212.825 samples/sec
+epoch 4, total_step 65680, total loss is 20.93 , inference loss is 11.30, weight deacy loss is 9.63, training accuracy is 0.218750, time 224.621 samples/sec
+epoch 4, total_step 65700, total loss is 22.33 , inference loss is 12.70, weight deacy loss is 9.63, training accuracy is 0.281250, time 12.032 samples/sec
+epoch 4, total_step 65720, total loss is 22.19 , inference loss is 12.57, weight deacy loss is 9.63, training accuracy is 0.250000, time 216.515 samples/sec
+epoch 4, total_step 65740, total loss is 20.99 , inference loss is 11.37, weight deacy loss is 9.62, training accuracy is 0.218750, time 214.576 samples/sec
+epoch 4, total_step 65760, total loss is 19.65 , inference loss is 10.03, weight deacy loss is 9.62, training accuracy is 0.312500, time 209.790 samples/sec
+epoch 4, total_step 65780, total loss is 26.06 , inference loss is 16.43, weight deacy loss is 9.62, training accuracy is 0.437500, time 205.714 samples/sec
+epoch 4, total_step 65800, total loss is 21.41 , inference loss is 11.79, weight deacy loss is 9.62, training accuracy is 0.187500, time 206.472 samples/sec
+epoch 4, total_step 65820, total loss is 19.25 , inference loss is 9.63, weight deacy loss is 9.62, training accuracy is 0.281250, time 206.591 samples/sec
+epoch 4, total_step 65840, total loss is 20.76 , inference loss is 11.14, weight deacy loss is 9.62, training accuracy is 0.406250, time 214.440 samples/sec
+epoch 4, total_step 65860, total loss is 19.63 , inference loss is 10.01, weight deacy loss is 9.62, training accuracy is 0.218750, time 218.141 samples/sec
+epoch 4, total_step 65880, total loss is 20.26 , inference loss is 10.65, weight deacy loss is 9.62, training accuracy is 0.125000, time 205.322 samples/sec
+epoch 4, total_step 65900, total loss is 23.70 , inference loss is 14.09, weight deacy loss is 9.62, training accuracy is 0.187500, time 193.403 samples/sec
+epoch 4, total_step 65920, total loss is 25.67 , inference loss is 16.06, weight deacy loss is 9.61, training accuracy is 0.218750, time 208.318 samples/sec
+epoch 4, total_step 65940, total loss is 23.69 , inference loss is 14.08, weight deacy loss is 9.61, training accuracy is 0.343750, time 211.919 samples/sec
+epoch 4, total_step 65960, total loss is 17.48 , inference loss is 7.87, weight deacy loss is 9.61, training accuracy is 0.312500, time 219.255 samples/sec
+epoch 4, total_step 65980, total loss is 18.34 , inference loss is 8.73, weight deacy loss is 9.61, training accuracy is 0.375000, time 212.667 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.817127999999986
+best_threshold_index 110 0.9848148148148148
+best_threshold_index 110 0.9842592592592593
+best_threshold_index 110 0.9859259259259259
+best_threshold_index 110 0.9861111111111112
+best_threshold_index 110 0.9857407407407407
+best_threshold_index 110 0.9837037037037037
+best_threshold_index 110 0.985
+best_threshold_index 110 0.9840740740740741
+best_threshold_index 110 0.9833333333333333
+best_threshold_index 110 0.9837037037037037
+[lfw][66000]XNorm: 20.801256
+[lfw][66000]Accuracy-Flip: 0.98467+-0.00859
+test accuracy is: 0.9846666666666668
+epoch 4, total_step 66000, total loss is 24.92 , inference loss is 15.31, weight deacy loss is 9.61, training accuracy is 0.281250, time 18.289 samples/sec
+epoch 4, total_step 66020, total loss is 20.19 , inference loss is 10.58, weight deacy loss is 9.61, training accuracy is 0.312500, time 218.089 samples/sec
+epoch 4, total_step 66040, total loss is 24.39 , inference loss is 14.79, weight deacy loss is 9.61, training accuracy is 0.312500, time 215.650 samples/sec
+epoch 4, total_step 66060, total loss is 18.72 , inference loss is 9.11, weight deacy loss is 9.61, training accuracy is 0.312500, time 216.895 samples/sec
+epoch 4, total_step 66080, total loss is 19.84 , inference loss is 10.24, weight deacy loss is 9.61, training accuracy is 0.343750, time 214.046 samples/sec
+epoch 4, total_step 66100, total loss is 23.10 , inference loss is 13.50, weight deacy loss is 9.61, training accuracy is 0.250000, time 214.949 samples/sec
+epoch 4, total_step 66120, total loss is 20.00 , inference loss is 10.40, weight deacy loss is 9.60, training accuracy is 0.281250, time 223.670 samples/sec
+epoch 4, total_step 66140, total loss is 21.96 , inference loss is 12.36, weight deacy loss is 9.60, training accuracy is 0.281250, time 213.649 samples/sec
+epoch 4, total_step 66160, total loss is 21.05 , inference loss is 11.45, weight deacy loss is 9.60, training accuracy is 0.250000, time 299.458 samples/sec
+epoch 4, total_step 66180, total loss is 20.95 , inference loss is 11.35, weight deacy loss is 9.60, training accuracy is 0.343750, time 298.761 samples/sec
+epoch 4, total_step 66200, total loss is 19.56 , inference loss is 9.96, weight deacy loss is 9.60, training accuracy is 0.343750, time 207.984 samples/sec
+epoch 4, total_step 66220, total loss is 25.18 , inference loss is 15.58, weight deacy loss is 9.60, training accuracy is 0.187500, time 215.986 samples/sec
+epoch 4, total_step 66240, total loss is 19.60 , inference loss is 10.00, weight deacy loss is 9.60, training accuracy is 0.156250, time 215.453 samples/sec
+epoch 4, total_step 66260, total loss is 24.25 , inference loss is 14.66, weight deacy loss is 9.60, training accuracy is 0.250000, time 211.777 samples/sec
+epoch 4, total_step 66280, total loss is 23.95 , inference loss is 14.36, weight deacy loss is 9.60, training accuracy is 0.250000, time 217.843 samples/sec
+epoch 4, total_step 66300, total loss is 21.63 , inference loss is 12.04, weight deacy loss is 9.59, training accuracy is 0.312500, time 11.010 samples/sec
+epoch 4, total_step 66320, total loss is 21.89 , inference loss is 12.30, weight deacy loss is 9.59, training accuracy is 0.312500, time 208.129 samples/sec
+epoch 4, total_step 66340, total loss is 22.12 , inference loss is 12.52, weight deacy loss is 9.59, training accuracy is 0.312500, time 211.854 samples/sec
+epoch 4, total_step 66360, total loss is 24.37 , inference loss is 14.78, weight deacy loss is 9.59, training accuracy is 0.250000, time 210.332 samples/sec
+epoch 4, total_step 66380, total loss is 21.22 , inference loss is 11.63, weight deacy loss is 9.59, training accuracy is 0.312500, time 217.219 samples/sec
+epoch 4, total_step 66400, total loss is 20.92 , inference loss is 11.33, weight deacy loss is 9.59, training accuracy is 0.343750, time 208.964 samples/sec
+epoch 4, total_step 66420, total loss is 23.00 , inference loss is 13.41, weight deacy loss is 9.59, training accuracy is 0.156250, time 206.615 samples/sec
+epoch 4, total_step 66440, total loss is 15.33 , inference loss is 5.74, weight deacy loss is 9.59, training accuracy is 0.468750, time 210.986 samples/sec
+epoch 4, total_step 66460, total loss is 19.95 , inference loss is 10.37, weight deacy loss is 9.59, training accuracy is 0.343750, time 216.344 samples/sec
+epoch 4, total_step 66480, total loss is 18.10 , inference loss is 8.51, weight deacy loss is 9.58, training accuracy is 0.375000, time 219.002 samples/sec
+epoch 4, total_step 66500, total loss is 22.54 , inference loss is 12.96, weight deacy loss is 9.58, training accuracy is 0.375000, time 224.296 samples/sec
+epoch 4, total_step 66520, total loss is 18.60 , inference loss is 9.02, weight deacy loss is 9.58, training accuracy is 0.343750, time 211.223 samples/sec
+epoch 4, total_step 66540, total loss is 23.00 , inference loss is 13.42, weight deacy loss is 9.58, training accuracy is 0.281250, time 209.783 samples/sec
+epoch 4, total_step 66560, total loss is 22.85 , inference loss is 13.27, weight deacy loss is 9.58, training accuracy is 0.187500, time 298.035 samples/sec
+epoch 4, total_step 66580, total loss is 18.33 , inference loss is 8.75, weight deacy loss is 9.58, training accuracy is 0.281250, time 216.008 samples/sec
+epoch 4, total_step 66600, total loss is 21.14 , inference loss is 11.56, weight deacy loss is 9.58, training accuracy is 0.312500, time 14.625 samples/sec
+epoch 4, total_step 66620, total loss is 26.09 , inference loss is 16.51, weight deacy loss is 9.58, training accuracy is 0.312500, time 213.740 samples/sec
+epoch 4, total_step 66640, total loss is 19.67 , inference loss is 10.09, weight deacy loss is 9.58, training accuracy is 0.281250, time 217.154 samples/sec
+epoch 4, total_step 66660, total loss is 20.59 , inference loss is 11.01, weight deacy loss is 9.57, training accuracy is 0.250000, time 207.055 samples/sec
+epoch 4, total_step 66680, total loss is 18.73 , inference loss is 9.15, weight deacy loss is 9.57, training accuracy is 0.437500, time 213.911 samples/sec
+epoch 4, total_step 66700, total loss is 22.34 , inference loss is 12.77, weight deacy loss is 9.57, training accuracy is 0.250000, time 201.597 samples/sec
+epoch 4, total_step 66720, total loss is 21.17 , inference loss is 11.59, weight deacy loss is 9.57, training accuracy is 0.343750, time 217.538 samples/sec
+epoch 4, total_step 66740, total loss is 19.11 , inference loss is 9.54, weight deacy loss is 9.57, training accuracy is 0.312500, time 211.024 samples/sec
+epoch 4, total_step 66760, total loss is 20.96 , inference loss is 11.39, weight deacy loss is 9.57, training accuracy is 0.250000, time 214.061 samples/sec
+epoch 4, total_step 66780, total loss is 21.57 , inference loss is 12.00, weight deacy loss is 9.57, training accuracy is 0.281250, time 207.103 samples/sec
+epoch 4, total_step 66800, total loss is 21.35 , inference loss is 11.79, weight deacy loss is 9.57, training accuracy is 0.250000, time 208.066 samples/sec
+epoch 4, total_step 66820, total loss is 23.67 , inference loss is 14.10, weight deacy loss is 9.57, training accuracy is 0.062500, time 299.980 samples/sec
+epoch 4, total_step 66840, total loss is 19.43 , inference loss is 9.87, weight deacy loss is 9.56, training accuracy is 0.250000, time 213.090 samples/sec
+epoch 4, total_step 66860, total loss is 17.87 , inference loss is 8.31, weight deacy loss is 9.56, training accuracy is 0.281250, time 206.578 samples/sec
+epoch 4, total_step 66880, total loss is 20.05 , inference loss is 10.48, weight deacy loss is 9.56, training accuracy is 0.281250, time 211.163 samples/sec
+epoch 4, total_step 66900, total loss is 22.62 , inference loss is 13.06, weight deacy loss is 9.56, training accuracy is 0.250000, time 13.327 samples/sec
+epoch 4, total_step 66920, total loss is 22.77 , inference loss is 13.21, weight deacy loss is 9.56, training accuracy is 0.218750, time 216.593 samples/sec
+epoch 4, total_step 66940, total loss is 18.89 , inference loss is 9.33, weight deacy loss is 9.56, training accuracy is 0.375000, time 210.963 samples/sec
+epoch 4, total_step 66960, total loss is 19.75 , inference loss is 10.20, weight deacy loss is 9.56, training accuracy is 0.281250, time 225.575 samples/sec
+epoch 4, total_step 66980, total loss is 19.30 , inference loss is 9.75, weight deacy loss is 9.56, training accuracy is 0.250000, time 210.624 samples/sec
+epoch 4, total_step 67000, total loss is 18.79 , inference loss is 9.23, weight deacy loss is 9.56, training accuracy is 0.375000, time 207.802 samples/sec
+epoch 4, total_step 67020, total loss is 17.44 , inference loss is 7.88, weight deacy loss is 9.55, training accuracy is 0.343750, time 219.355 samples/sec
+epoch 4, total_step 67040, total loss is 19.16 , inference loss is 9.61, weight deacy loss is 9.55, training accuracy is 0.312500, time 226.017 samples/sec
+epoch 4, total_step 67060, total loss is 17.88 , inference loss is 8.33, weight deacy loss is 9.55, training accuracy is 0.406250, time 212.565 samples/sec
+epoch 4, total_step 67080, total loss is 20.32 , inference loss is 10.77, weight deacy loss is 9.55, training accuracy is 0.343750, time 213.476 samples/sec
+epoch 4, total_step 67100, total loss is 22.60 , inference loss is 13.05, weight deacy loss is 9.55, training accuracy is 0.250000, time 225.216 samples/sec
+epoch 4, total_step 67120, total loss is 24.74 , inference loss is 15.19, weight deacy loss is 9.55, training accuracy is 0.218750, time 216.582 samples/sec
+epoch 4, total_step 67140, total loss is 21.98 , inference loss is 12.43, weight deacy loss is 9.55, training accuracy is 0.281250, time 204.172 samples/sec
+epoch 4, total_step 67160, total loss is 23.31 , inference loss is 13.76, weight deacy loss is 9.55, training accuracy is 0.281250, time 220.088 samples/sec
+epoch 4, total_step 67180, total loss is 16.46 , inference loss is 6.92, weight deacy loss is 9.55, training accuracy is 0.312500, time 212.008 samples/sec
+epoch 4, total_step 67200, total loss is 26.04 , inference loss is 16.50, weight deacy loss is 9.54, training accuracy is 0.156250, time 13.695 samples/sec
+epoch 4, total_step 67220, total loss is 22.10 , inference loss is 12.56, weight deacy loss is 9.54, training accuracy is 0.187500, time 214.320 samples/sec
+epoch 4, total_step 67240, total loss is 19.23 , inference loss is 9.69, weight deacy loss is 9.54, training accuracy is 0.343750, time 212.145 samples/sec
+epoch 4, total_step 67260, total loss is 20.59 , inference loss is 11.05, weight deacy loss is 9.54, training accuracy is 0.187500, time 216.617 samples/sec
+epoch 4, total_step 67280, total loss is 16.23 , inference loss is 6.69, weight deacy loss is 9.54, training accuracy is 0.343750, time 221.107 samples/sec
+epoch 4, total_step 67300, total loss is 20.60 , inference loss is 11.06, weight deacy loss is 9.54, training accuracy is 0.250000, time 212.263 samples/sec
+epoch 4, total_step 67320, total loss is 19.59 , inference loss is 10.05, weight deacy loss is 9.54, training accuracy is 0.343750, time 214.218 samples/sec
+epoch 4, total_step 67340, total loss is 21.25 , inference loss is 11.72, weight deacy loss is 9.54, training accuracy is 0.281250, time 207.036 samples/sec
+epoch 4, total_step 67360, total loss is 21.96 , inference loss is 12.42, weight deacy loss is 9.54, training accuracy is 0.218750, time 227.241 samples/sec
+epoch 4, total_step 67380, total loss is 28.26 , inference loss is 18.72, weight deacy loss is 9.53, training accuracy is 0.093750, time 209.816 samples/sec
+epoch 4, total_step 67400, total loss is 17.27 , inference loss is 7.73, weight deacy loss is 9.53, training accuracy is 0.312500, time 210.384 samples/sec
+epoch 4, total_step 67420, total loss is 20.56 , inference loss is 11.03, weight deacy loss is 9.53, training accuracy is 0.343750, time 208.135 samples/sec
+epoch 4, total_step 67440, total loss is 21.70 , inference loss is 12.17, weight deacy loss is 9.53, training accuracy is 0.218750, time 213.308 samples/sec
+epoch 4, total_step 67460, total loss is 19.66 , inference loss is 10.13, weight deacy loss is 9.53, training accuracy is 0.406250, time 211.433 samples/sec
+epoch 4, total_step 67480, total loss is 24.39 , inference loss is 14.86, weight deacy loss is 9.53, training accuracy is 0.218750, time 209.781 samples/sec
+epoch 4, total_step 67500, total loss is 18.19 , inference loss is 8.66, weight deacy loss is 9.53, training accuracy is 0.312500, time 13.985 samples/sec
+epoch 4, total_step 67520, total loss is 20.02 , inference loss is 10.49, weight deacy loss is 9.53, training accuracy is 0.250000, time 222.690 samples/sec
+epoch 4, total_step 67540, total loss is 21.29 , inference loss is 11.77, weight deacy loss is 9.53, training accuracy is 0.218750, time 205.866 samples/sec
+epoch 4, total_step 67560, total loss is 19.17 , inference loss is 9.64, weight deacy loss is 9.52, training accuracy is 0.343750, time 220.824 samples/sec
+epoch 4, total_step 67580, total loss is 24.83 , inference loss is 15.31, weight deacy loss is 9.52, training accuracy is 0.281250, time 220.843 samples/sec
+epoch 4, total_step 67600, total loss is 18.81 , inference loss is 9.28, weight deacy loss is 9.52, training accuracy is 0.187500, time 214.023 samples/sec
+epoch 4, total_step 67620, total loss is 19.66 , inference loss is 10.14, weight deacy loss is 9.52, training accuracy is 0.250000, time 298.471 samples/sec
+epoch 4, total_step 67640, total loss is 21.28 , inference loss is 11.76, weight deacy loss is 9.52, training accuracy is 0.250000, time 210.660 samples/sec
+epoch 4, total_step 67660, total loss is 20.79 , inference loss is 11.27, weight deacy loss is 9.52, training accuracy is 0.250000, time 207.378 samples/sec
+epoch 4, total_step 67680, total loss is 23.10 , inference loss is 13.58, weight deacy loss is 9.52, training accuracy is 0.218750, time 218.402 samples/sec
+epoch 4, total_step 67700, total loss is 22.15 , inference loss is 12.63, weight deacy loss is 9.52, training accuracy is 0.250000, time 206.654 samples/sec
+epoch 4, total_step 67720, total loss is 19.55 , inference loss is 10.03, weight deacy loss is 9.52, training accuracy is 0.312500, time 218.524 samples/sec
+epoch 4, total_step 67740, total loss is 23.01 , inference loss is 13.49, weight deacy loss is 9.52, training accuracy is 0.281250, time 211.363 samples/sec
+epoch 4, total_step 67760, total loss is 19.04 , inference loss is 9.53, weight deacy loss is 9.51, training accuracy is 0.343750, time 206.364 samples/sec
+epoch 4, total_step 67780, total loss is 21.00 , inference loss is 11.49, weight deacy loss is 9.51, training accuracy is 0.250000, time 213.229 samples/sec
+epoch 4, total_step 67800, total loss is 19.90 , inference loss is 10.39, weight deacy loss is 9.51, training accuracy is 0.281250, time 13.999 samples/sec
+epoch 4, total_step 67820, total loss is 22.31 , inference loss is 12.80, weight deacy loss is 9.51, training accuracy is 0.218750, time 216.333 samples/sec
+epoch 4, total_step 67840, total loss is 21.67 , inference loss is 12.16, weight deacy loss is 9.51, training accuracy is 0.250000, time 217.460 samples/sec
+epoch 4, total_step 67860, total loss is 20.23 , inference loss is 10.72, weight deacy loss is 9.51, training accuracy is 0.343750, time 214.523 samples/sec
+epoch 4, total_step 67880, total loss is 20.74 , inference loss is 11.23, weight deacy loss is 9.51, training accuracy is 0.281250, time 217.042 samples/sec
+epoch 4, total_step 67900, total loss is 21.37 , inference loss is 11.87, weight deacy loss is 9.51, training accuracy is 0.156250, time 296.317 samples/sec
+epoch 4, total_step 67920, total loss is 21.86 , inference loss is 12.35, weight deacy loss is 9.51, training accuracy is 0.281250, time 217.107 samples/sec
+epoch 4, total_step 67940, total loss is 22.93 , inference loss is 13.42, weight deacy loss is 9.50, training accuracy is 0.156250, time 211.862 samples/sec
+epoch 4, total_step 67960, total loss is 23.91 , inference loss is 14.40, weight deacy loss is 9.50, training accuracy is 0.281250, time 222.601 samples/sec
+epoch 4, total_step 67980, total loss is 22.37 , inference loss is 12.87, weight deacy loss is 9.50, training accuracy is 0.250000, time 214.216 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.55919499999997
+best_threshold_index 107 0.9838888888888889
+best_threshold_index 107 0.9827777777777778
+best_threshold_index 107 0.9831481481481481
+best_threshold_index 107 0.9846296296296296
+best_threshold_index 107 0.9838888888888889
+best_threshold_index 107 0.9827777777777778
+best_threshold_index 107 0.9837037037037037
+best_threshold_index 107 0.9829629629629629
+best_threshold_index 107 0.9816666666666667
+best_threshold_index 107 0.9822222222222222
+[lfw][68000]XNorm: 23.280288
+[lfw][68000]Accuracy-Flip: 0.98317+-0.00751
+test accuracy is: 0.9831666666666667
+epoch 4, total_step 68000, total loss is 24.83 , inference loss is 15.33, weight deacy loss is 9.50, training accuracy is 0.312500, time 18.906 samples/sec
+epoch 4, total_step 68020, total loss is 20.88 , inference loss is 11.38, weight deacy loss is 9.50, training accuracy is 0.187500, time 208.376 samples/sec
+epoch 4, total_step 68040, total loss is 18.06 , inference loss is 8.56, weight deacy loss is 9.50, training accuracy is 0.406250, time 214.096 samples/sec
+epoch 4, total_step 68060, total loss is 19.51 , inference loss is 10.01, weight deacy loss is 9.50, training accuracy is 0.343750, time 210.874 samples/sec
+epoch 4, total_step 68080, total loss is 21.21 , inference loss is 11.72, weight deacy loss is 9.50, training accuracy is 0.250000, time 213.080 samples/sec
+epoch 4, total_step 68100, total loss is 21.81 , inference loss is 12.31, weight deacy loss is 9.50, training accuracy is 0.156250, time 13.400 samples/sec
+epoch 4, total_step 68120, total loss is 19.40 , inference loss is 9.91, weight deacy loss is 9.49, training accuracy is 0.281250, time 218.103 samples/sec
+epoch 4, total_step 68140, total loss is 19.60 , inference loss is 10.11, weight deacy loss is 9.49, training accuracy is 0.156250, time 296.209 samples/sec
+epoch 4, total_step 68160, total loss is 20.82 , inference loss is 11.33, weight deacy loss is 9.49, training accuracy is 0.156250, time 234.463 samples/sec
+epoch 4, total_step 68180, total loss is 25.89 , inference loss is 16.40, weight deacy loss is 9.49, training accuracy is 0.156250, time 221.688 samples/sec
+epoch 4, total_step 68200, total loss is 24.27 , inference loss is 14.78, weight deacy loss is 9.49, training accuracy is 0.218750, time 209.112 samples/sec
+epoch 4, total_step 68220, total loss is 24.14 , inference loss is 14.65, weight deacy loss is 9.49, training accuracy is 0.187500, time 207.678 samples/sec
+epoch 4, total_step 68240, total loss is 25.55 , inference loss is 16.06, weight deacy loss is 9.49, training accuracy is 0.187500, time 210.708 samples/sec
+epoch 4, total_step 68260, total loss is 19.64 , inference loss is 10.15, weight deacy loss is 9.49, training accuracy is 0.343750, time 210.303 samples/sec
+epoch 4, total_step 68280, total loss is 15.20 , inference loss is 5.72, weight deacy loss is 9.49, training accuracy is 0.437500, time 206.054 samples/sec
+epoch 4, total_step 68300, total loss is 18.70 , inference loss is 9.22, weight deacy loss is 9.48, training accuracy is 0.156250, time 207.788 samples/sec
+epoch 4, total_step 68320, total loss is 25.17 , inference loss is 15.69, weight deacy loss is 9.48, training accuracy is 0.218750, time 217.179 samples/sec
+epoch 4, total_step 68340, total loss is 23.65 , inference loss is 14.17, weight deacy loss is 9.48, training accuracy is 0.218750, time 225.421 samples/sec
+epoch 4, total_step 68360, total loss is 22.55 , inference loss is 13.07, weight deacy loss is 9.48, training accuracy is 0.187500, time 215.536 samples/sec
+epoch 4, total_step 68380, total loss is 21.70 , inference loss is 12.22, weight deacy loss is 9.48, training accuracy is 0.218750, time 199.487 samples/sec
+epoch 4, total_step 68400, total loss is 18.52 , inference loss is 9.04, weight deacy loss is 9.48, training accuracy is 0.406250, time 10.637 samples/sec
+epoch 4, total_step 68420, total loss is 21.89 , inference loss is 12.41, weight deacy loss is 9.48, training accuracy is 0.218750, time 216.365 samples/sec
+epoch 4, total_step 68440, total loss is 20.16 , inference loss is 10.69, weight deacy loss is 9.48, training accuracy is 0.281250, time 204.874 samples/sec
+epoch 4, total_step 68460, total loss is 17.71 , inference loss is 8.23, weight deacy loss is 9.48, training accuracy is 0.312500, time 210.307 samples/sec
+epoch 4, total_step 68480, total loss is 22.49 , inference loss is 13.01, weight deacy loss is 9.48, training accuracy is 0.218750, time 229.081 samples/sec
+epoch 4, total_step 68500, total loss is 23.44 , inference loss is 13.96, weight deacy loss is 9.47, training accuracy is 0.250000, time 208.093 samples/sec
+epoch 4, total_step 68520, total loss is 19.43 , inference loss is 9.96, weight deacy loss is 9.47, training accuracy is 0.312500, time 212.361 samples/sec
+epoch 4, total_step 68540, total loss is 19.02 , inference loss is 9.55, weight deacy loss is 9.47, training accuracy is 0.375000, time 216.556 samples/sec
+epoch 4, total_step 68560, total loss is 22.33 , inference loss is 12.86, weight deacy loss is 9.47, training accuracy is 0.312500, time 213.406 samples/sec
+epoch 4, total_step 68580, total loss is 27.21 , inference loss is 17.74, weight deacy loss is 9.47, training accuracy is 0.250000, time 227.118 samples/sec
+epoch 4, total_step 68600, total loss is 20.80 , inference loss is 11.33, weight deacy loss is 9.47, training accuracy is 0.187500, time 206.145 samples/sec
+epoch 4, total_step 68620, total loss is 25.41 , inference loss is 15.95, weight deacy loss is 9.47, training accuracy is 0.187500, time 215.865 samples/sec
+epoch 4, total_step 68640, total loss is 20.83 , inference loss is 11.37, weight deacy loss is 9.47, training accuracy is 0.281250, time 208.943 samples/sec
+epoch 4, total_step 68660, total loss is 23.00 , inference loss is 13.53, weight deacy loss is 9.47, training accuracy is 0.218750, time 215.140 samples/sec
+epoch 4, total_step 68680, total loss is 18.03 , inference loss is 8.57, weight deacy loss is 9.46, training accuracy is 0.375000, time 214.263 samples/sec
+epoch 4, total_step 68700, total loss is 22.63 , inference loss is 13.17, weight deacy loss is 9.46, training accuracy is 0.218750, time 13.017 samples/sec
+epoch 4, total_step 68720, total loss is 20.95 , inference loss is 11.49, weight deacy loss is 9.46, training accuracy is 0.375000, time 201.731 samples/sec
+epoch 4, total_step 68740, total loss is 22.72 , inference loss is 13.26, weight deacy loss is 9.46, training accuracy is 0.218750, time 211.166 samples/sec
+epoch 4, total_step 68760, total loss is 20.21 , inference loss is 10.75, weight deacy loss is 9.46, training accuracy is 0.281250, time 210.959 samples/sec
+epoch 4, total_step 68780, total loss is 21.86 , inference loss is 12.40, weight deacy loss is 9.46, training accuracy is 0.250000, time 213.976 samples/sec
+epoch 4, total_step 68800, total loss is 18.03 , inference loss is 8.57, weight deacy loss is 9.46, training accuracy is 0.343750, time 207.133 samples/sec
+epoch 4, total_step 68820, total loss is 20.81 , inference loss is 11.35, weight deacy loss is 9.46, training accuracy is 0.281250, time 222.394 samples/sec
+epoch 4, total_step 68840, total loss is 17.95 , inference loss is 8.49, weight deacy loss is 9.46, training accuracy is 0.343750, time 215.273 samples/sec
+epoch 4, total_step 68860, total loss is 20.46 , inference loss is 11.00, weight deacy loss is 9.45, training accuracy is 0.343750, time 201.817 samples/sec
+epoch 4, total_step 68880, total loss is 20.06 , inference loss is 10.61, weight deacy loss is 9.45, training accuracy is 0.156250, time 210.943 samples/sec
+epoch 4, total_step 68900, total loss is 29.80 , inference loss is 20.35, weight deacy loss is 9.45, training accuracy is 0.250000, time 219.399 samples/sec
+epoch 4, total_step 68920, total loss is 20.98 , inference loss is 11.52, weight deacy loss is 9.45, training accuracy is 0.312500, time 216.238 samples/sec
+epoch 4, total_step 68940, total loss is 21.02 , inference loss is 11.56, weight deacy loss is 9.45, training accuracy is 0.375000, time 220.339 samples/sec
+epoch 4, total_step 68960, total loss is 16.99 , inference loss is 7.54, weight deacy loss is 9.45, training accuracy is 0.406250, time 212.493 samples/sec
+epoch 4, total_step 68980, total loss is 22.94 , inference loss is 13.49, weight deacy loss is 9.45, training accuracy is 0.156250, time 298.662 samples/sec
+epoch 4, total_step 69000, total loss is 17.72 , inference loss is 8.27, weight deacy loss is 9.45, training accuracy is 0.406250, time 13.190 samples/sec
+epoch 4, total_step 69020, total loss is 20.42 , inference loss is 10.98, weight deacy loss is 9.45, training accuracy is 0.250000, time 210.305 samples/sec
+epoch 4, total_step 69040, total loss is 18.79 , inference loss is 9.34, weight deacy loss is 9.44, training accuracy is 0.437500, time 207.429 samples/sec
+epoch 4, total_step 69060, total loss is 18.22 , inference loss is 8.78, weight deacy loss is 9.44, training accuracy is 0.375000, time 227.528 samples/sec
+epoch 4, total_step 69080, total loss is 22.18 , inference loss is 12.74, weight deacy loss is 9.44, training accuracy is 0.281250, time 209.251 samples/sec
+epoch 4, total_step 69100, total loss is 21.09 , inference loss is 11.65, weight deacy loss is 9.44, training accuracy is 0.281250, time 212.609 samples/sec
+epoch 4, total_step 69120, total loss is 17.71 , inference loss is 8.27, weight deacy loss is 9.44, training accuracy is 0.281250, time 223.953 samples/sec
+epoch 4, total_step 69140, total loss is 18.09 , inference loss is 8.65, weight deacy loss is 9.44, training accuracy is 0.375000, time 212.148 samples/sec
+epoch 4, total_step 69160, total loss is 24.43 , inference loss is 14.99, weight deacy loss is 9.44, training accuracy is 0.218750, time 211.139 samples/sec
+epoch 4, total_step 69180, total loss is 20.13 , inference loss is 10.69, weight deacy loss is 9.44, training accuracy is 0.375000, time 205.743 samples/sec
+epoch 4, total_step 69200, total loss is 21.29 , inference loss is 11.85, weight deacy loss is 9.44, training accuracy is 0.187500, time 209.077 samples/sec
+epoch 4, total_step 69220, total loss is 25.49 , inference loss is 16.06, weight deacy loss is 9.43, training accuracy is 0.156250, time 212.500 samples/sec
+epoch 4, total_step 69240, total loss is 19.06 , inference loss is 9.63, weight deacy loss is 9.43, training accuracy is 0.343750, time 210.409 samples/sec
+epoch 4, total_step 69260, total loss is 21.18 , inference loss is 11.75, weight deacy loss is 9.43, training accuracy is 0.375000, time 298.583 samples/sec
+epoch 4, total_step 69280, total loss is 22.54 , inference loss is 13.11, weight deacy loss is 9.43, training accuracy is 0.312500, time 210.443 samples/sec
+epoch 4, total_step 69300, total loss is 20.54 , inference loss is 11.11, weight deacy loss is 9.43, training accuracy is 0.281250, time 11.693 samples/sec
+epoch 4, total_step 69320, total loss is 20.23 , inference loss is 10.80, weight deacy loss is 9.43, training accuracy is 0.406250, time 208.998 samples/sec
+epoch 4, total_step 69340, total loss is 20.43 , inference loss is 11.00, weight deacy loss is 9.43, training accuracy is 0.218750, time 225.290 samples/sec
+epoch 4, total_step 69360, total loss is 23.56 , inference loss is 14.14, weight deacy loss is 9.43, training accuracy is 0.281250, time 208.457 samples/sec
+epoch 4, total_step 69380, total loss is 21.75 , inference loss is 12.32, weight deacy loss is 9.43, training accuracy is 0.312500, time 213.143 samples/sec
+epoch 4, total_step 69400, total loss is 25.67 , inference loss is 16.24, weight deacy loss is 9.42, training accuracy is 0.218750, time 206.078 samples/sec
+epoch 4, total_step 69420, total loss is 22.98 , inference loss is 13.55, weight deacy loss is 9.42, training accuracy is 0.281250, time 212.807 samples/sec
+epoch 4, total_step 69440, total loss is 20.59 , inference loss is 11.17, weight deacy loss is 9.42, training accuracy is 0.312500, time 216.466 samples/sec
+epoch 4, total_step 69460, total loss is 20.87 , inference loss is 11.44, weight deacy loss is 9.42, training accuracy is 0.343750, time 217.576 samples/sec
+epoch 4, total_step 69480, total loss is 22.59 , inference loss is 13.17, weight deacy loss is 9.42, training accuracy is 0.343750, time 224.762 samples/sec
+epoch 4, total_step 69500, total loss is 21.98 , inference loss is 12.56, weight deacy loss is 9.42, training accuracy is 0.312500, time 214.490 samples/sec
+epoch 4, total_step 69520, total loss is 20.73 , inference loss is 11.31, weight deacy loss is 9.42, training accuracy is 0.156250, time 300.628 samples/sec
+epoch 4, total_step 69540, total loss is 21.45 , inference loss is 12.03, weight deacy loss is 9.42, training accuracy is 0.187500, time 208.056 samples/sec
+epoch 4, total_step 69560, total loss is 23.62 , inference loss is 14.20, weight deacy loss is 9.42, training accuracy is 0.125000, time 207.201 samples/sec
+epoch 4, total_step 69580, total loss is 20.51 , inference loss is 11.10, weight deacy loss is 9.41, training accuracy is 0.218750, time 209.491 samples/sec
+epoch 4, total_step 69600, total loss is 22.23 , inference loss is 12.82, weight deacy loss is 9.41, training accuracy is 0.093750, time 14.045 samples/sec
+epoch 4, total_step 69620, total loss is 24.53 , inference loss is 15.12, weight deacy loss is 9.41, training accuracy is 0.125000, time 209.506 samples/sec
+epoch 4, total_step 69640, total loss is 23.20 , inference loss is 13.79, weight deacy loss is 9.41, training accuracy is 0.250000, time 205.831 samples/sec
+epoch 4, total_step 69660, total loss is 21.72 , inference loss is 12.31, weight deacy loss is 9.41, training accuracy is 0.125000, time 204.619 samples/sec
+epoch 4, total_step 69680, total loss is 22.28 , inference loss is 12.87, weight deacy loss is 9.41, training accuracy is 0.250000, time 217.206 samples/sec
+epoch 4, total_step 69700, total loss is 18.77 , inference loss is 9.36, weight deacy loss is 9.41, training accuracy is 0.281250, time 215.468 samples/sec
+epoch 4, total_step 69720, total loss is 23.94 , inference loss is 14.53, weight deacy loss is 9.41, training accuracy is 0.156250, time 211.841 samples/sec
+epoch 4, total_step 69740, total loss is 20.06 , inference loss is 10.65, weight deacy loss is 9.41, training accuracy is 0.250000, time 189.727 samples/sec
+epoch 4, total_step 69760, total loss is 24.28 , inference loss is 14.87, weight deacy loss is 9.41, training accuracy is 0.250000, time 234.376 samples/sec
+epoch 4, total_step 69780, total loss is 20.88 , inference loss is 11.48, weight deacy loss is 9.40, training accuracy is 0.187500, time 299.730 samples/sec
+epoch 4, total_step 69800, total loss is 22.51 , inference loss is 13.10, weight deacy loss is 9.40, training accuracy is 0.218750, time 213.963 samples/sec
+epoch 4, total_step 69820, total loss is 21.02 , inference loss is 11.62, weight deacy loss is 9.40, training accuracy is 0.187500, time 210.573 samples/sec
+epoch 4, total_step 69840, total loss is 21.31 , inference loss is 11.91, weight deacy loss is 9.40, training accuracy is 0.218750, time 216.237 samples/sec
+epoch 4, total_step 69860, total loss is 22.14 , inference loss is 12.74, weight deacy loss is 9.40, training accuracy is 0.250000, time 216.189 samples/sec
+epoch 4, total_step 69880, total loss is 18.80 , inference loss is 9.40, weight deacy loss is 9.40, training accuracy is 0.281250, time 217.223 samples/sec
+epoch 4, total_step 69900, total loss is 22.71 , inference loss is 13.31, weight deacy loss is 9.40, training accuracy is 0.343750, time 13.950 samples/sec
+epoch 4, total_step 69920, total loss is 22.91 , inference loss is 13.51, weight deacy loss is 9.40, training accuracy is 0.312500, time 293.712 samples/sec
+epoch 4, total_step 69940, total loss is 27.50 , inference loss is 18.10, weight deacy loss is 9.40, training accuracy is 0.187500, time 214.210 samples/sec
+epoch 4, total_step 69960, total loss is 21.08 , inference loss is 11.69, weight deacy loss is 9.39, training accuracy is 0.187500, time 221.511 samples/sec
+epoch 4, total_step 69980, total loss is 20.06 , inference loss is 10.67, weight deacy loss is 9.39, training accuracy is 0.250000, time 208.815 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.47982799999998
+best_threshold_index 113 0.9872222222222222
+best_threshold_index 113 0.9862962962962963
+best_threshold_index 113 0.9881481481481481
+best_threshold_index 113 0.9881481481481481
+best_threshold_index 113 0.9868518518518519
+best_threshold_index 113 0.9862962962962963
+best_threshold_index 113 0.9868518518518519
+best_threshold_index 113 0.9859259259259259
+best_threshold_index 113 0.9853703703703703
+best_threshold_index 113 0.9855555555555555
+[lfw][70000]XNorm: 22.954492
+[lfw][70000]Accuracy-Flip: 0.98667+-0.00830
+test accuracy is: 0.9866666666666666
+epoch 4, total_step 70000, total loss is 23.48 , inference loss is 14.09, weight deacy loss is 9.39, training accuracy is 0.156250, time 18.445 samples/sec
+epoch 4, total_step 70020, total loss is 22.88 , inference loss is 13.49, weight deacy loss is 9.39, training accuracy is 0.250000, time 207.251 samples/sec
+epoch 4, total_step 70040, total loss is 23.02 , inference loss is 13.63, weight deacy loss is 9.39, training accuracy is 0.187500, time 208.450 samples/sec
+epoch 4, total_step 70060, total loss is 20.64 , inference loss is 11.25, weight deacy loss is 9.39, training accuracy is 0.187500, time 206.154 samples/sec
+epoch 4, total_step 70080, total loss is 20.50 , inference loss is 11.11, weight deacy loss is 9.39, training accuracy is 0.281250, time 218.808 samples/sec
+epoch 4, total_step 70100, total loss is 21.82 , inference loss is 12.44, weight deacy loss is 9.39, training accuracy is 0.218750, time 212.668 samples/sec
+epoch 4, total_step 70120, total loss is 19.82 , inference loss is 10.43, weight deacy loss is 9.39, training accuracy is 0.312500, time 210.755 samples/sec
+epoch 4, total_step 70140, total loss is 22.77 , inference loss is 13.39, weight deacy loss is 9.39, training accuracy is 0.218750, time 227.107 samples/sec
+epoch 4, total_step 70160, total loss is 22.38 , inference loss is 12.99, weight deacy loss is 9.38, training accuracy is 0.218750, time 214.607 samples/sec
+epoch 4, total_step 70180, total loss is 20.91 , inference loss is 11.53, weight deacy loss is 9.38, training accuracy is 0.343750, time 206.831 samples/sec
+epoch 4, total_step 70200, total loss is 24.87 , inference loss is 15.48, weight deacy loss is 9.38, training accuracy is 0.343750, time 14.450 samples/sec
+epoch 4, total_step 70220, total loss is 27.64 , inference loss is 18.26, weight deacy loss is 9.38, training accuracy is 0.187500, time 217.684 samples/sec
+epoch 4, total_step 70240, total loss is 19.11 , inference loss is 9.73, weight deacy loss is 9.38, training accuracy is 0.250000, time 218.466 samples/sec
+epoch 4, total_step 70260, total loss is 22.00 , inference loss is 12.62, weight deacy loss is 9.38, training accuracy is 0.281250, time 209.408 samples/sec
+epoch 4, total_step 70280, total loss is 20.72 , inference loss is 11.34, weight deacy loss is 9.38, training accuracy is 0.250000, time 214.703 samples/sec
+epoch 4, total_step 70300, total loss is 21.11 , inference loss is 11.73, weight deacy loss is 9.38, training accuracy is 0.125000, time 227.669 samples/sec
+epoch 4, total_step 70320, total loss is 25.45 , inference loss is 16.07, weight deacy loss is 9.38, training accuracy is 0.218750, time 209.224 samples/sec
+epoch 4, total_step 70340, total loss is 20.39 , inference loss is 11.02, weight deacy loss is 9.37, training accuracy is 0.312500, time 206.521 samples/sec
+epoch 4, total_step 70360, total loss is 23.12 , inference loss is 13.75, weight deacy loss is 9.37, training accuracy is 0.218750, time 215.167 samples/sec
+epoch 4, total_step 70380, total loss is 20.97 , inference loss is 11.60, weight deacy loss is 9.37, training accuracy is 0.156250, time 197.490 samples/sec
+epoch 4, total_step 70400, total loss is 20.19 , inference loss is 10.82, weight deacy loss is 9.37, training accuracy is 0.406250, time 207.984 samples/sec
+epoch 4, total_step 70420, total loss is 25.93 , inference loss is 16.55, weight deacy loss is 9.37, training accuracy is 0.187500, time 213.260 samples/sec
+epoch 4, total_step 70440, total loss is 20.87 , inference loss is 11.50, weight deacy loss is 9.37, training accuracy is 0.218750, time 222.234 samples/sec
+epoch 4, total_step 70460, total loss is 20.28 , inference loss is 10.91, weight deacy loss is 9.37, training accuracy is 0.281250, time 212.963 samples/sec
+epoch 4, total_step 70480, total loss is 20.01 , inference loss is 10.64, weight deacy loss is 9.37, training accuracy is 0.468750, time 211.486 samples/sec
+epoch 4, total_step 70500, total loss is 21.29 , inference loss is 11.92, weight deacy loss is 9.37, training accuracy is 0.218750, time 14.338 samples/sec
+epoch 4, total_step 70520, total loss is 20.08 , inference loss is 10.72, weight deacy loss is 9.37, training accuracy is 0.312500, time 219.654 samples/sec
+epoch 4, total_step 70540, total loss is 21.19 , inference loss is 11.83, weight deacy loss is 9.36, training accuracy is 0.218750, time 223.567 samples/sec
+epoch 4, total_step 70560, total loss is 23.24 , inference loss is 13.87, weight deacy loss is 9.36, training accuracy is 0.218750, time 211.875 samples/sec
+epoch 4, total_step 70580, total loss is 19.29 , inference loss is 9.93, weight deacy loss is 9.36, training accuracy is 0.375000, time 222.469 samples/sec
+epoch 4, total_step 70600, total loss is 27.18 , inference loss is 17.82, weight deacy loss is 9.36, training accuracy is 0.031250, time 204.046 samples/sec
+epoch 4, total_step 70620, total loss is 23.33 , inference loss is 13.97, weight deacy loss is 9.36, training accuracy is 0.218750, time 220.729 samples/sec
+epoch 4, total_step 70640, total loss is 22.86 , inference loss is 13.50, weight deacy loss is 9.36, training accuracy is 0.187500, time 216.603 samples/sec
+epoch 4, total_step 70660, total loss is 21.80 , inference loss is 12.44, weight deacy loss is 9.36, training accuracy is 0.312500, time 211.964 samples/sec
+epoch 4, total_step 70680, total loss is 18.42 , inference loss is 9.06, weight deacy loss is 9.36, training accuracy is 0.375000, time 204.763 samples/sec
+epoch 4, total_step 70700, total loss is 21.40 , inference loss is 12.05, weight deacy loss is 9.36, training accuracy is 0.250000, time 231.559 samples/sec
+epoch 4, total_step 70720, total loss is 21.97 , inference loss is 12.61, weight deacy loss is 9.36, training accuracy is 0.218750, time 210.380 samples/sec
+epoch 4, total_step 70740, total loss is 20.71 , inference loss is 11.35, weight deacy loss is 9.35, training accuracy is 0.343750, time 215.567 samples/sec
+epoch 4, total_step 70760, total loss is 22.81 , inference loss is 13.45, weight deacy loss is 9.35, training accuracy is 0.156250, time 218.649 samples/sec
+epoch 4, total_step 70780, total loss is 25.41 , inference loss is 16.06, weight deacy loss is 9.35, training accuracy is 0.250000, time 201.811 samples/sec
+epoch 4, total_step 70800, total loss is 21.39 , inference loss is 12.04, weight deacy loss is 9.35, training accuracy is 0.250000, time 14.103 samples/sec
+epoch 4, total_step 70820, total loss is 22.23 , inference loss is 12.88, weight deacy loss is 9.35, training accuracy is 0.156250, time 213.915 samples/sec
+epoch 4, total_step 70840, total loss is 19.70 , inference loss is 10.35, weight deacy loss is 9.35, training accuracy is 0.250000, time 208.790 samples/sec
+epoch 4, total_step 70860, total loss is 20.10 , inference loss is 10.75, weight deacy loss is 9.35, training accuracy is 0.312500, time 200.008 samples/sec
+epoch 4, total_step 70880, total loss is 21.63 , inference loss is 12.28, weight deacy loss is 9.35, training accuracy is 0.218750, time 198.966 samples/sec
+epoch 4, total_step 70900, total loss is 22.33 , inference loss is 12.99, weight deacy loss is 9.35, training accuracy is 0.312500, time 210.808 samples/sec
+epoch 4, total_step 70920, total loss is 19.83 , inference loss is 10.48, weight deacy loss is 9.34, training accuracy is 0.375000, time 217.384 samples/sec
+epoch 4, total_step 70940, total loss is 22.05 , inference loss is 12.70, weight deacy loss is 9.34, training accuracy is 0.218750, time 212.616 samples/sec
+epoch 4, total_step 70960, total loss is 23.32 , inference loss is 13.98, weight deacy loss is 9.34, training accuracy is 0.281250, time 218.438 samples/sec
+epoch 4, total_step 70980, total loss is 22.09 , inference loss is 12.75, weight deacy loss is 9.34, training accuracy is 0.250000, time 214.412 samples/sec
+epoch 4, total_step 71000, total loss is 21.51 , inference loss is 12.17, weight deacy loss is 9.34, training accuracy is 0.218750, time 201.766 samples/sec
+epoch 4, total_step 71020, total loss is 19.14 , inference loss is 9.80, weight deacy loss is 9.34, training accuracy is 0.281250, time 203.698 samples/sec
+epoch 4, total_step 71040, total loss is 22.08 , inference loss is 12.74, weight deacy loss is 9.34, training accuracy is 0.250000, time 297.826 samples/sec
+epoch 4, total_step 71060, total loss is 20.81 , inference loss is 11.48, weight deacy loss is 9.34, training accuracy is 0.281250, time 212.487 samples/sec
+epoch 4, total_step 71080, total loss is 23.18 , inference loss is 13.85, weight deacy loss is 9.34, training accuracy is 0.218750, time 207.208 samples/sec
+epoch 4, total_step 71100, total loss is 23.14 , inference loss is 13.81, weight deacy loss is 9.34, training accuracy is 0.281250, time 11.428 samples/sec
+epoch 4, total_step 71120, total loss is 19.26 , inference loss is 9.92, weight deacy loss is 9.33, training accuracy is 0.375000, time 212.212 samples/sec
+epoch 4, total_step 71140, total loss is 17.68 , inference loss is 8.35, weight deacy loss is 9.33, training accuracy is 0.312500, time 206.674 samples/sec
+epoch 4, total_step 71160, total loss is 18.23 , inference loss is 8.90, weight deacy loss is 9.33, training accuracy is 0.281250, time 219.135 samples/sec
+epoch 4, total_step 71180, total loss is 17.61 , inference loss is 8.27, weight deacy loss is 9.33, training accuracy is 0.312500, time 221.047 samples/sec
+epoch 4, total_step 71200, total loss is 21.09 , inference loss is 11.76, weight deacy loss is 9.33, training accuracy is 0.312500, time 217.419 samples/sec
+epoch 4, total_step 71220, total loss is 19.56 , inference loss is 10.23, weight deacy loss is 9.33, training accuracy is 0.406250, time 206.348 samples/sec
+epoch 4, total_step 71240, total loss is 18.74 , inference loss is 9.41, weight deacy loss is 9.33, training accuracy is 0.250000, time 205.591 samples/sec
+epoch 4, total_step 71260, total loss is 15.71 , inference loss is 6.38, weight deacy loss is 9.33, training accuracy is 0.343750, time 205.393 samples/sec
+epoch 4, total_step 71280, total loss is 19.94 , inference loss is 10.61, weight deacy loss is 9.33, training accuracy is 0.218750, time 212.756 samples/sec
+epoch 4, total_step 71300, total loss is 20.44 , inference loss is 11.11, weight deacy loss is 9.32, training accuracy is 0.343750, time 297.154 samples/sec
+epoch 4, total_step 71320, total loss is 20.76 , inference loss is 11.44, weight deacy loss is 9.32, training accuracy is 0.375000, time 210.640 samples/sec
+epoch 4, total_step 71340, total loss is 18.77 , inference loss is 9.45, weight deacy loss is 9.32, training accuracy is 0.406250, time 236.261 samples/sec
+epoch 4, total_step 71360, total loss is 21.24 , inference loss is 11.92, weight deacy loss is 9.32, training accuracy is 0.281250, time 211.918 samples/sec
+epoch 4, total_step 71380, total loss is 20.52 , inference loss is 11.19, weight deacy loss is 9.32, training accuracy is 0.218750, time 211.764 samples/sec
+epoch 4, total_step 71400, total loss is 16.97 , inference loss is 7.65, weight deacy loss is 9.32, training accuracy is 0.250000, time 14.032 samples/sec
+epoch 4, total_step 71420, total loss is 19.86 , inference loss is 10.54, weight deacy loss is 9.32, training accuracy is 0.156250, time 209.269 samples/sec
+epoch 4, total_step 71440, total loss is 21.39 , inference loss is 12.07, weight deacy loss is 9.32, training accuracy is 0.156250, time 210.913 samples/sec
+epoch 4, total_step 71460, total loss is 24.79 , inference loss is 15.47, weight deacy loss is 9.32, training accuracy is 0.218750, time 212.732 samples/sec
+epoch 4, total_step 71480, total loss is 18.87 , inference loss is 9.56, weight deacy loss is 9.32, training accuracy is 0.250000, time 219.150 samples/sec
+epoch 4, total_step 71500, total loss is 19.20 , inference loss is 9.89, weight deacy loss is 9.31, training accuracy is 0.250000, time 216.124 samples/sec
+epoch 4, total_step 71520, total loss is 19.31 , inference loss is 10.00, weight deacy loss is 9.31, training accuracy is 0.312500, time 224.775 samples/sec
+epoch 4, total_step 71540, total loss is 20.19 , inference loss is 10.87, weight deacy loss is 9.31, training accuracy is 0.250000, time 214.241 samples/sec
+epoch 4, total_step 71560, total loss is 23.02 , inference loss is 13.70, weight deacy loss is 9.31, training accuracy is 0.250000, time 211.642 samples/sec
+epoch 4, total_step 71580, total loss is 17.54 , inference loss is 8.23, weight deacy loss is 9.31, training accuracy is 0.406250, time 208.061 samples/sec
+epoch 4, total_step 71600, total loss is 20.19 , inference loss is 10.88, weight deacy loss is 9.31, training accuracy is 0.187500, time 225.695 samples/sec
+epoch 4, total_step 71620, total loss is 19.11 , inference loss is 9.80, weight deacy loss is 9.31, training accuracy is 0.187500, time 296.753 samples/sec
+epoch 4, total_step 71640, total loss is 17.45 , inference loss is 8.14, weight deacy loss is 9.31, training accuracy is 0.437500, time 246.845 samples/sec
+epoch 4, total_step 71660, total loss is 21.25 , inference loss is 11.94, weight deacy loss is 9.31, training accuracy is 0.250000, time 205.058 samples/sec
+epoch 4, total_step 71680, total loss is 19.22 , inference loss is 9.91, weight deacy loss is 9.30, training accuracy is 0.250000, time 215.529 samples/sec
+epoch 4, total_step 71700, total loss is 18.70 , inference loss is 9.40, weight deacy loss is 9.30, training accuracy is 0.281250, time 12.608 samples/sec
+epoch 4, total_step 71720, total loss is 17.81 , inference loss is 8.50, weight deacy loss is 9.30, training accuracy is 0.250000, time 179.080 samples/sec
+epoch 4, total_step 71740, total loss is 19.98 , inference loss is 10.68, weight deacy loss is 9.30, training accuracy is 0.281250, time 200.396 samples/sec
+epoch 4, total_step 71760, total loss is 23.71 , inference loss is 14.41, weight deacy loss is 9.30, training accuracy is 0.312500, time 215.293 samples/sec
+epoch 4, total_step 71780, total loss is 18.43 , inference loss is 9.13, weight deacy loss is 9.30, training accuracy is 0.281250, time 210.383 samples/sec
+epoch 4, total_step 71800, total loss is 16.99 , inference loss is 7.69, weight deacy loss is 9.30, training accuracy is 0.343750, time 208.780 samples/sec
+epoch 4, total_step 71820, total loss is 20.00 , inference loss is 10.71, weight deacy loss is 9.30, training accuracy is 0.281250, time 223.624 samples/sec
+epoch 4, total_step 71840, total loss is 20.26 , inference loss is 10.97, weight deacy loss is 9.30, training accuracy is 0.156250, time 205.325 samples/sec
+epoch 4, total_step 71860, total loss is 26.53 , inference loss is 17.24, weight deacy loss is 9.30, training accuracy is 0.187500, time 217.389 samples/sec
+epoch 4, total_step 71880, total loss is 20.05 , inference loss is 10.75, weight deacy loss is 9.29, training accuracy is 0.406250, time 216.471 samples/sec
+epoch 4, total_step 71900, total loss is 21.41 , inference loss is 12.12, weight deacy loss is 9.29, training accuracy is 0.250000, time 224.560 samples/sec
+epoch 4, total_step 71920, total loss is 14.46 , inference loss is 5.17, weight deacy loss is 9.29, training accuracy is 0.562500, time 188.618 samples/sec
+epoch 4, total_step 71940, total loss is 23.06 , inference loss is 13.77, weight deacy loss is 9.29, training accuracy is 0.281250, time 210.167 samples/sec
+epoch 4, total_step 71960, total loss is 19.49 , inference loss is 10.20, weight deacy loss is 9.29, training accuracy is 0.218750, time 219.522 samples/sec
+epoch 4, total_step 71980, total loss is 21.51 , inference loss is 12.22, weight deacy loss is 9.29, training accuracy is 0.218750, time 206.709 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.629170999999975
+best_threshold_index 106 0.9829629629629629
+best_threshold_index 106 0.982037037037037
+best_threshold_index 106 0.9831481481481481
+best_threshold_index 106 0.9829629629629629
+best_threshold_index 107 0.9824074074074074
+best_threshold_index 107 0.9811111111111112
+best_threshold_index 106 0.9822222222222222
+best_threshold_index 106 0.9812962962962963
+best_threshold_index 107 0.9811111111111112
+best_threshold_index 106 0.9812962962962963
+[lfw][72000]XNorm: 22.500911
+[lfw][72000]Accuracy-Flip: 0.98150+-0.00693
+test accuracy is: 0.9814999999999999
+epoch 4, total_step 72000, total loss is 18.54 , inference loss is 9.25, weight deacy loss is 9.29, training accuracy is 0.312500, time 19.281 samples/sec
+epoch 4, total_step 72020, total loss is 16.71 , inference loss is 7.43, weight deacy loss is 9.29, training accuracy is 0.437500, time 205.703 samples/sec
+epoch 4, total_step 72040, total loss is 17.49 , inference loss is 8.21, weight deacy loss is 9.29, training accuracy is 0.343750, time 211.337 samples/sec
+epoch 4, total_step 72060, total loss is 24.00 , inference loss is 14.72, weight deacy loss is 9.29, training accuracy is 0.218750, time 208.717 samples/sec
+epoch 4, total_step 72080, total loss is 25.47 , inference loss is 16.19, weight deacy loss is 9.28, training accuracy is 0.281250, time 218.133 samples/sec
+epoch 4, total_step 72100, total loss is 22.25 , inference loss is 12.96, weight deacy loss is 9.28, training accuracy is 0.312500, time 224.586 samples/sec
+epoch 4, total_step 72120, total loss is 22.42 , inference loss is 13.14, weight deacy loss is 9.28, training accuracy is 0.218750, time 213.001 samples/sec
+epoch 4, total_step 72140, total loss is 18.61 , inference loss is 9.33, weight deacy loss is 9.28, training accuracy is 0.375000, time 235.765 samples/sec
+epoch 4, total_step 72160, total loss is 22.73 , inference loss is 13.45, weight deacy loss is 9.28, training accuracy is 0.156250, time 204.663 samples/sec
+epoch 4, total_step 72180, total loss is 22.69 , inference loss is 13.41, weight deacy loss is 9.28, training accuracy is 0.281250, time 213.475 samples/sec
+epoch 4, total_step 72200, total loss is 20.66 , inference loss is 11.39, weight deacy loss is 9.28, training accuracy is 0.250000, time 212.458 samples/sec
+epoch 4, total_step 72220, total loss is 21.76 , inference loss is 12.48, weight deacy loss is 9.28, training accuracy is 0.250000, time 209.109 samples/sec
+epoch 4, total_step 72240, total loss is 20.16 , inference loss is 10.88, weight deacy loss is 9.28, training accuracy is 0.312500, time 211.314 samples/sec
+epoch 4, total_step 72260, total loss is 23.54 , inference loss is 14.27, weight deacy loss is 9.27, training accuracy is 0.187500, time 210.637 samples/sec
+epoch 4, total_step 72280, total loss is 21.40 , inference loss is 12.12, weight deacy loss is 9.27, training accuracy is 0.250000, time 205.919 samples/sec
+epoch 4, total_step 72300, total loss is 23.45 , inference loss is 14.17, weight deacy loss is 9.27, training accuracy is 0.218750, time 13.350 samples/sec
+epoch 4, total_step 72320, total loss is 22.10 , inference loss is 12.83, weight deacy loss is 9.27, training accuracy is 0.343750, time 221.310 samples/sec
+epoch 4, total_step 72340, total loss is 17.81 , inference loss is 8.54, weight deacy loss is 9.27, training accuracy is 0.250000, time 212.715 samples/sec
+epoch 4, total_step 72360, total loss is 23.77 , inference loss is 14.50, weight deacy loss is 9.27, training accuracy is 0.312500, time 299.506 samples/sec
+epoch 4, total_step 72380, total loss is 24.41 , inference loss is 15.14, weight deacy loss is 9.27, training accuracy is 0.218750, time 222.350 samples/sec
+epoch 4, total_step 72400, total loss is 19.44 , inference loss is 10.17, weight deacy loss is 9.27, training accuracy is 0.250000, time 212.057 samples/sec
+epoch 4, total_step 72420, total loss is 20.54 , inference loss is 11.28, weight deacy loss is 9.27, training accuracy is 0.218750, time 212.617 samples/sec
+epoch 4, total_step 72440, total loss is 21.17 , inference loss is 11.90, weight deacy loss is 9.27, training accuracy is 0.125000, time 211.150 samples/sec
+epoch 4, total_step 72460, total loss is 21.30 , inference loss is 12.03, weight deacy loss is 9.26, training accuracy is 0.250000, time 201.142 samples/sec
+epoch 4, total_step 72480, total loss is 21.80 , inference loss is 12.54, weight deacy loss is 9.26, training accuracy is 0.250000, time 209.980 samples/sec
+epoch 4, total_step 72500, total loss is 16.78 , inference loss is 7.52, weight deacy loss is 9.26, training accuracy is 0.468750, time 206.521 samples/sec
+epoch 4, total_step 72520, total loss is 20.99 , inference loss is 11.73, weight deacy loss is 9.26, training accuracy is 0.156250, time 222.020 samples/sec
+epoch 4, total_step 72540, total loss is 25.02 , inference loss is 15.76, weight deacy loss is 9.26, training accuracy is 0.093750, time 213.159 samples/sec
+epoch 4, total_step 72560, total loss is 23.43 , inference loss is 14.18, weight deacy loss is 9.26, training accuracy is 0.187500, time 214.218 samples/sec
+epoch 4, total_step 72580, total loss is 20.52 , inference loss is 11.27, weight deacy loss is 9.26, training accuracy is 0.281250, time 210.582 samples/sec
+epoch 4, total_step 72600, total loss is 22.96 , inference loss is 13.70, weight deacy loss is 9.26, training accuracy is 0.218750, time 13.208 samples/sec
+epoch 4, total_step 72620, total loss is 23.65 , inference loss is 14.40, weight deacy loss is 9.26, training accuracy is 0.218750, time 299.849 samples/sec
+epoch 4, total_step 72640, total loss is 20.68 , inference loss is 11.43, weight deacy loss is 9.26, training accuracy is 0.250000, time 203.469 samples/sec
+epoch 4, total_step 72660, total loss is 18.06 , inference loss is 8.81, weight deacy loss is 9.25, training accuracy is 0.312500, time 237.929 samples/sec
+epoch 4, total_step 72680, total loss is 17.08 , inference loss is 7.82, weight deacy loss is 9.25, training accuracy is 0.375000, time 212.943 samples/sec
+epoch 4, total_step 72700, total loss is 19.51 , inference loss is 10.26, weight deacy loss is 9.25, training accuracy is 0.281250, time 212.134 samples/sec
+epoch 4, total_step 72720, total loss is 20.20 , inference loss is 10.95, weight deacy loss is 9.25, training accuracy is 0.125000, time 218.141 samples/sec
+epoch 4, total_step 72740, total loss is 21.46 , inference loss is 12.21, weight deacy loss is 9.25, training accuracy is 0.187500, time 211.270 samples/sec
+epoch 4, total_step 72760, total loss is 18.76 , inference loss is 9.51, weight deacy loss is 9.25, training accuracy is 0.250000, time 209.026 samples/sec
+epoch 4, total_step 72780, total loss is 18.03 , inference loss is 8.78, weight deacy loss is 9.25, training accuracy is 0.312500, time 223.582 samples/sec
+epoch 4, total_step 72800, total loss is 20.12 , inference loss is 10.87, weight deacy loss is 9.25, training accuracy is 0.156250, time 213.560 samples/sec
+epoch 4, total_step 72820, total loss is 20.78 , inference loss is 11.54, weight deacy loss is 9.25, training accuracy is 0.156250, time 210.688 samples/sec
+epoch 4, total_step 72840, total loss is 21.51 , inference loss is 12.26, weight deacy loss is 9.25, training accuracy is 0.156250, time 207.868 samples/sec
+epoch 4, total_step 72860, total loss is 22.70 , inference loss is 13.46, weight deacy loss is 9.24, training accuracy is 0.218750, time 206.366 samples/sec
+epoch 4, total_step 72880, total loss is 19.60 , inference loss is 10.36, weight deacy loss is 9.24, training accuracy is 0.343750, time 213.004 samples/sec
+epoch 4, total_step 72900, total loss is 27.12 , inference loss is 17.87, weight deacy loss is 9.24, training accuracy is 0.218750, time 15.042 samples/sec
+epoch 4, total_step 72920, total loss is 18.00 , inference loss is 8.76, weight deacy loss is 9.24, training accuracy is 0.187500, time 211.629 samples/sec
+epoch 4, total_step 72940, total loss is 25.68 , inference loss is 16.44, weight deacy loss is 9.24, training accuracy is 0.156250, time 206.305 samples/sec
+epoch 4, total_step 72960, total loss is 20.31 , inference loss is 11.07, weight deacy loss is 9.24, training accuracy is 0.250000, time 189.330 samples/sec
+epoch 4, total_step 72980, total loss is 21.72 , inference loss is 12.49, weight deacy loss is 9.24, training accuracy is 0.250000, time 209.272 samples/sec
+epoch 4, total_step 73000, total loss is 22.61 , inference loss is 13.38, weight deacy loss is 9.24, training accuracy is 0.156250, time 213.055 samples/sec
+epoch 4, total_step 73020, total loss is 17.73 , inference loss is 8.49, weight deacy loss is 9.24, training accuracy is 0.250000, time 208.671 samples/sec
+epoch 4, total_step 73040, total loss is 17.48 , inference loss is 8.24, weight deacy loss is 9.24, training accuracy is 0.250000, time 215.414 samples/sec
+epoch 4, total_step 73060, total loss is 17.46 , inference loss is 8.22, weight deacy loss is 9.23, training accuracy is 0.250000, time 207.749 samples/sec
+epoch 4, total_step 73080, total loss is 20.99 , inference loss is 11.76, weight deacy loss is 9.23, training accuracy is 0.250000, time 216.214 samples/sec
+epoch 4, total_step 73100, total loss is 19.62 , inference loss is 10.39, weight deacy loss is 9.23, training accuracy is 0.312500, time 202.212 samples/sec
+epoch 4, total_step 73120, total loss is 18.88 , inference loss is 9.65, weight deacy loss is 9.23, training accuracy is 0.281250, time 233.507 samples/sec
+epoch 4, total_step 73140, total loss is 20.67 , inference loss is 11.44, weight deacy loss is 9.23, training accuracy is 0.281250, time 207.012 samples/sec
+epoch 4, total_step 73160, total loss is 20.54 , inference loss is 11.31, weight deacy loss is 9.23, training accuracy is 0.281250, time 209.615 samples/sec
+epoch 4, total_step 73180, total loss is 21.42 , inference loss is 12.19, weight deacy loss is 9.23, training accuracy is 0.218750, time 212.371 samples/sec
+epoch 4, total_step 73200, total loss is 20.19 , inference loss is 10.96, weight deacy loss is 9.23, training accuracy is 0.218750, time 13.925 samples/sec
+epoch 4, total_step 73220, total loss is 20.65 , inference loss is 11.42, weight deacy loss is 9.23, training accuracy is 0.250000, time 215.173 samples/sec
+epoch 4, total_step 73240, total loss is 18.77 , inference loss is 9.55, weight deacy loss is 9.22, training accuracy is 0.406250, time 213.156 samples/sec
+epoch 4, total_step 73260, total loss is 23.15 , inference loss is 13.93, weight deacy loss is 9.22, training accuracy is 0.218750, time 211.008 samples/sec
+epoch 4, total_step 73280, total loss is 20.32 , inference loss is 11.10, weight deacy loss is 9.22, training accuracy is 0.187500, time 210.689 samples/sec
+epoch 4, total_step 73300, total loss is 22.91 , inference loss is 13.69, weight deacy loss is 9.22, training accuracy is 0.062500, time 211.542 samples/sec
+epoch 4, total_step 73320, total loss is 24.52 , inference loss is 15.30, weight deacy loss is 9.22, training accuracy is 0.093750, time 217.097 samples/sec
+epoch 4, total_step 73340, total loss is 23.42 , inference loss is 14.20, weight deacy loss is 9.22, training accuracy is 0.156250, time 215.257 samples/sec
+epoch 4, total_step 73360, total loss is 23.76 , inference loss is 14.55, weight deacy loss is 9.22, training accuracy is 0.187500, time 210.614 samples/sec
+epoch 4, total_step 73380, total loss is 22.51 , inference loss is 13.29, weight deacy loss is 9.22, training accuracy is 0.187500, time 217.416 samples/sec
+epoch 4, total_step 73400, total loss is 19.93 , inference loss is 10.71, weight deacy loss is 9.22, training accuracy is 0.281250, time 216.592 samples/sec
+epoch 4, total_step 73420, total loss is 24.49 , inference loss is 15.27, weight deacy loss is 9.22, training accuracy is 0.125000, time 297.487 samples/sec
+epoch 4, total_step 73440, total loss is 24.11 , inference loss is 14.89, weight deacy loss is 9.21, training accuracy is 0.093750, time 296.088 samples/sec
+epoch 4, total_step 73460, total loss is 21.99 , inference loss is 12.78, weight deacy loss is 9.21, training accuracy is 0.218750, time 213.002 samples/sec
+epoch 4, total_step 73480, total loss is 21.89 , inference loss is 12.68, weight deacy loss is 9.21, training accuracy is 0.250000, time 205.286 samples/sec
+epoch 4, total_step 73500, total loss is 24.23 , inference loss is 15.02, weight deacy loss is 9.21, training accuracy is 0.187500, time 13.831 samples/sec
+epoch 4, total_step 73520, total loss is 19.51 , inference loss is 10.30, weight deacy loss is 9.21, training accuracy is 0.281250, time 215.745 samples/sec
+epoch 4, total_step 73540, total loss is 20.59 , inference loss is 11.38, weight deacy loss is 9.21, training accuracy is 0.250000, time 216.562 samples/sec
+epoch 4, total_step 73560, total loss is 19.47 , inference loss is 10.26, weight deacy loss is 9.21, training accuracy is 0.093750, time 214.406 samples/sec
+epoch 4, total_step 73580, total loss is 24.99 , inference loss is 15.78, weight deacy loss is 9.21, training accuracy is 0.156250, time 215.568 samples/sec
+epoch 4, total_step 73600, total loss is 19.41 , inference loss is 10.21, weight deacy loss is 9.21, training accuracy is 0.281250, time 209.562 samples/sec
+epoch 4, total_step 73620, total loss is 21.55 , inference loss is 12.34, weight deacy loss is 9.21, training accuracy is 0.156250, time 211.233 samples/sec
+epoch 4, total_step 73640, total loss is 25.21 , inference loss is 16.00, weight deacy loss is 9.20, training accuracy is 0.156250, time 233.446 samples/sec
+epoch 4, total_step 73660, total loss is 19.80 , inference loss is 10.60, weight deacy loss is 9.20, training accuracy is 0.343750, time 206.550 samples/sec
+epoch 4, total_step 73680, total loss is 26.54 , inference loss is 17.34, weight deacy loss is 9.20, training accuracy is 0.156250, time 212.025 samples/sec
+epoch 4, total_step 73700, total loss is 25.29 , inference loss is 16.08, weight deacy loss is 9.20, training accuracy is 0.187500, time 207.792 samples/sec
+epoch 4, total_step 73720, total loss is 19.68 , inference loss is 10.48, weight deacy loss is 9.20, training accuracy is 0.218750, time 213.262 samples/sec
+epoch 4, total_step 73740, total loss is 21.91 , inference loss is 12.71, weight deacy loss is 9.20, training accuracy is 0.312500, time 214.081 samples/sec
+epoch 4, total_step 73760, total loss is 21.17 , inference loss is 11.97, weight deacy loss is 9.20, training accuracy is 0.218750, time 212.561 samples/sec
+epoch 4, total_step 73780, total loss is 20.75 , inference loss is 11.55, weight deacy loss is 9.20, training accuracy is 0.156250, time 208.873 samples/sec
+epoch 4, total_step 73800, total loss is 23.25 , inference loss is 14.06, weight deacy loss is 9.20, training accuracy is 0.156250, time 13.826 samples/sec
+epoch 4, total_step 73820, total loss is 21.36 , inference loss is 12.17, weight deacy loss is 9.20, training accuracy is 0.187500, time 208.734 samples/sec
+epoch 4, total_step 73840, total loss is 18.96 , inference loss is 9.76, weight deacy loss is 9.19, training accuracy is 0.218750, time 204.483 samples/sec
+epoch 4, total_step 73860, total loss is 20.04 , inference loss is 10.84, weight deacy loss is 9.19, training accuracy is 0.218750, time 211.633 samples/sec
+epoch 4, total_step 73880, total loss is 23.19 , inference loss is 14.00, weight deacy loss is 9.19, training accuracy is 0.156250, time 294.458 samples/sec
+epoch 4, total_step 73900, total loss is 19.17 , inference loss is 9.98, weight deacy loss is 9.19, training accuracy is 0.343750, time 206.546 samples/sec
+epoch 4, total_step 73920, total loss is 19.71 , inference loss is 10.52, weight deacy loss is 9.19, training accuracy is 0.218750, time 206.505 samples/sec
+epoch 4, total_step 73940, total loss is 22.96 , inference loss is 13.77, weight deacy loss is 9.19, training accuracy is 0.250000, time 207.811 samples/sec
+epoch 4, total_step 73960, total loss is 21.41 , inference loss is 12.22, weight deacy loss is 9.19, training accuracy is 0.281250, time 209.360 samples/sec
+epoch 4, total_step 73980, total loss is 21.99 , inference loss is 12.80, weight deacy loss is 9.19, training accuracy is 0.093750, time 207.820 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.833625999999995
+best_threshold_index 104 0.9835185185185186
+best_threshold_index 114 0.9825925925925926
+best_threshold_index 104 0.9831481481481481
+best_threshold_index 104 0.9851851851851852
+best_threshold_index 104 0.9835185185185186
+best_threshold_index 104 0.9827777777777778
+best_threshold_index 104 0.9833333333333333
+best_threshold_index 104 0.9825925925925926
+best_threshold_index 104 0.9816666666666667
+best_threshold_index 104 0.9822222222222222
+[lfw][74000]XNorm: 23.233927
+[lfw][74000]Accuracy-Flip: 0.98200+-0.00802
+test accuracy is: 0.982
+epoch 4, total_step 74000, total loss is 25.05 , inference loss is 15.87, weight deacy loss is 9.19, training accuracy is 0.125000, time 18.548 samples/sec
+epoch 4, total_step 74020, total loss is 20.23 , inference loss is 11.05, weight deacy loss is 9.19, training accuracy is 0.250000, time 225.506 samples/sec
+epoch 4, total_step 74040, total loss is 21.58 , inference loss is 12.39, weight deacy loss is 9.18, training accuracy is 0.437500, time 207.400 samples/sec
+epoch 4, total_step 74060, total loss is 19.72 , inference loss is 10.54, weight deacy loss is 9.18, training accuracy is 0.281250, time 218.647 samples/sec
+epoch 4, total_step 74080, total loss is 21.51 , inference loss is 12.32, weight deacy loss is 9.18, training accuracy is 0.312500, time 211.216 samples/sec
+epoch 4, total_step 74100, total loss is 22.55 , inference loss is 13.37, weight deacy loss is 9.18, training accuracy is 0.187500, time 13.559 samples/sec
+epoch 4, total_step 74120, total loss is 20.38 , inference loss is 11.20, weight deacy loss is 9.18, training accuracy is 0.281250, time 211.077 samples/sec
+epoch 4, total_step 74140, total loss is 20.38 , inference loss is 11.20, weight deacy loss is 9.18, training accuracy is 0.281250, time 218.833 samples/sec
+epoch 4, total_step 74160, total loss is 22.40 , inference loss is 13.22, weight deacy loss is 9.18, training accuracy is 0.093750, time 223.190 samples/sec
+epoch 4, total_step 74180, total loss is 18.66 , inference loss is 9.48, weight deacy loss is 9.18, training accuracy is 0.125000, time 215.175 samples/sec
+epoch 4, total_step 74200, total loss is 18.16 , inference loss is 8.98, weight deacy loss is 9.18, training accuracy is 0.281250, time 210.607 samples/sec
+epoch 4, total_step 74220, total loss is 18.53 , inference loss is 9.36, weight deacy loss is 9.18, training accuracy is 0.281250, time 212.435 samples/sec
+epoch 4, total_step 74240, total loss is 19.85 , inference loss is 10.68, weight deacy loss is 9.17, training accuracy is 0.218750, time 210.978 samples/sec
+epoch 4, total_step 74260, total loss is 22.83 , inference loss is 13.65, weight deacy loss is 9.17, training accuracy is 0.187500, time 296.952 samples/sec
+epoch 4, total_step 74280, total loss is 18.82 , inference loss is 9.65, weight deacy loss is 9.17, training accuracy is 0.281250, time 211.140 samples/sec
+epoch 4, total_step 74300, total loss is 22.65 , inference loss is 13.48, weight deacy loss is 9.17, training accuracy is 0.093750, time 209.134 samples/sec
+epoch 4, total_step 74320, total loss is 24.16 , inference loss is 14.99, weight deacy loss is 9.17, training accuracy is 0.031250, time 200.452 samples/sec
+epoch 4, total_step 74340, total loss is 24.03 , inference loss is 14.86, weight deacy loss is 9.17, training accuracy is 0.156250, time 209.097 samples/sec
+epoch 4, total_step 74360, total loss is 19.17 , inference loss is 10.00, weight deacy loss is 9.17, training accuracy is 0.218750, time 210.043 samples/sec
+epoch 4, total_step 74380, total loss is 20.22 , inference loss is 11.05, weight deacy loss is 9.17, training accuracy is 0.156250, time 214.223 samples/sec
+epoch 4, total_step 74400, total loss is 21.44 , inference loss is 12.28, weight deacy loss is 9.17, training accuracy is 0.156250, time 13.980 samples/sec
+epoch 4, total_step 74420, total loss is 22.12 , inference loss is 12.96, weight deacy loss is 9.17, training accuracy is 0.187500, time 207.806 samples/sec
+epoch 4, total_step 74440, total loss is 17.81 , inference loss is 8.65, weight deacy loss is 9.16, training accuracy is 0.375000, time 212.169 samples/sec
+epoch 4, total_step 74460, total loss is 21.79 , inference loss is 12.63, weight deacy loss is 9.16, training accuracy is 0.062500, time 198.780 samples/sec
+epoch 4, total_step 74480, total loss is 21.29 , inference loss is 12.13, weight deacy loss is 9.16, training accuracy is 0.250000, time 216.532 samples/sec
+epoch 4, total_step 74500, total loss is 27.45 , inference loss is 18.29, weight deacy loss is 9.16, training accuracy is 0.156250, time 207.463 samples/sec
+epoch 4, total_step 74520, total loss is 18.55 , inference loss is 9.39, weight deacy loss is 9.16, training accuracy is 0.343750, time 297.675 samples/sec
+epoch 4, total_step 74540, total loss is 21.38 , inference loss is 12.22, weight deacy loss is 9.16, training accuracy is 0.218750, time 205.570 samples/sec
+epoch 4, total_step 74560, total loss is 21.10 , inference loss is 11.94, weight deacy loss is 9.16, training accuracy is 0.187500, time 206.558 samples/sec
+epoch 4, total_step 74580, total loss is 24.29 , inference loss is 15.13, weight deacy loss is 9.16, training accuracy is 0.250000, time 203.736 samples/sec
+epoch 4, total_step 74600, total loss is 18.38 , inference loss is 9.22, weight deacy loss is 9.16, training accuracy is 0.281250, time 213.820 samples/sec
+epoch 4, total_step 74620, total loss is 17.71 , inference loss is 8.56, weight deacy loss is 9.16, training accuracy is 0.281250, time 216.518 samples/sec
+epoch 4, total_step 74640, total loss is 22.45 , inference loss is 13.30, weight deacy loss is 9.15, training accuracy is 0.187500, time 216.304 samples/sec
+epoch 4, total_step 74660, total loss is 19.83 , inference loss is 10.68, weight deacy loss is 9.15, training accuracy is 0.250000, time 210.658 samples/sec
+epoch 4, total_step 74680, total loss is 21.54 , inference loss is 12.39, weight deacy loss is 9.15, training accuracy is 0.062500, time 228.861 samples/sec
+epoch 4, total_step 74700, total loss is 21.06 , inference loss is 11.91, weight deacy loss is 9.15, training accuracy is 0.187500, time 12.864 samples/sec
+epoch 4, total_step 74720, total loss is 17.57 , inference loss is 8.42, weight deacy loss is 9.15, training accuracy is 0.312500, time 222.817 samples/sec
+epoch 4, total_step 74740, total loss is 22.29 , inference loss is 13.14, weight deacy loss is 9.15, training accuracy is 0.187500, time 210.034 samples/sec
+epoch 4, total_step 74760, total loss is 21.11 , inference loss is 11.96, weight deacy loss is 9.15, training accuracy is 0.156250, time 205.784 samples/sec
+epoch 4, total_step 74780, total loss is 25.04 , inference loss is 15.90, weight deacy loss is 9.15, training accuracy is 0.156250, time 221.617 samples/sec
+epoch 4, total_step 74800, total loss is 21.43 , inference loss is 12.28, weight deacy loss is 9.15, training accuracy is 0.125000, time 298.773 samples/sec
+epoch 4, total_step 74820, total loss is 17.55 , inference loss is 8.41, weight deacy loss is 9.15, training accuracy is 0.093750, time 219.205 samples/sec
+epoch 4, total_step 74840, total loss is 20.99 , inference loss is 11.84, weight deacy loss is 9.14, training accuracy is 0.218750, time 215.181 samples/sec
+epoch 4, total_step 74860, total loss is 20.42 , inference loss is 11.28, weight deacy loss is 9.14, training accuracy is 0.187500, time 219.538 samples/sec
+epoch 4, total_step 74880, total loss is 21.34 , inference loss is 12.20, weight deacy loss is 9.14, training accuracy is 0.156250, time 202.702 samples/sec
+epoch 4, total_step 74900, total loss is 20.13 , inference loss is 10.99, weight deacy loss is 9.14, training accuracy is 0.250000, time 203.175 samples/sec
+epoch 4, total_step 74920, total loss is 18.38 , inference loss is 9.23, weight deacy loss is 9.14, training accuracy is 0.281250, time 219.077 samples/sec
+epoch 4, total_step 74940, total loss is 19.73 , inference loss is 10.60, weight deacy loss is 9.14, training accuracy is 0.156250, time 213.739 samples/sec
+epoch 4, total_step 74960, total loss is 20.94 , inference loss is 11.80, weight deacy loss is 9.14, training accuracy is 0.187500, time 232.194 samples/sec
+epoch 4, total_step 74980, total loss is 24.84 , inference loss is 15.70, weight deacy loss is 9.14, training accuracy is 0.156250, time 203.687 samples/sec
+epoch 4, total_step 75000, total loss is 23.03 , inference loss is 13.89, weight deacy loss is 9.14, training accuracy is 0.250000, time 14.034 samples/sec
+epoch 4, total_step 75020, total loss is 23.05 , inference loss is 13.91, weight deacy loss is 9.14, training accuracy is 0.093750, time 215.906 samples/sec
+epoch 4, total_step 75040, total loss is 21.01 , inference loss is 11.88, weight deacy loss is 9.13, training accuracy is 0.250000, time 212.555 samples/sec
+epoch 4, total_step 75060, total loss is 19.54 , inference loss is 10.41, weight deacy loss is 9.13, training accuracy is 0.218750, time 298.040 samples/sec
+epoch 4, total_step 75080, total loss is 18.22 , inference loss is 9.09, weight deacy loss is 9.13, training accuracy is 0.218750, time 299.341 samples/sec
+epoch 4, total_step 75100, total loss is 22.19 , inference loss is 13.06, weight deacy loss is 9.13, training accuracy is 0.156250, time 212.094 samples/sec
+epoch 4, total_step 75120, total loss is 17.70 , inference loss is 8.57, weight deacy loss is 9.13, training accuracy is 0.250000, time 214.616 samples/sec
+epoch 4, total_step 75140, total loss is 21.25 , inference loss is 12.12, weight deacy loss is 9.13, training accuracy is 0.250000, time 212.474 samples/sec
+epoch 4, total_step 75160, total loss is 17.93 , inference loss is 8.81, weight deacy loss is 9.13, training accuracy is 0.187500, time 211.314 samples/sec
+epoch 4, total_step 75180, total loss is 21.78 , inference loss is 12.65, weight deacy loss is 9.13, training accuracy is 0.125000, time 213.487 samples/sec
+epoch 4, total_step 75200, total loss is 23.24 , inference loss is 14.11, weight deacy loss is 9.13, training accuracy is 0.281250, time 212.497 samples/sec
+epoch 4, total_step 75220, total loss is 22.89 , inference loss is 13.76, weight deacy loss is 9.13, training accuracy is 0.250000, time 213.831 samples/sec
+epoch 4, total_step 75240, total loss is 23.24 , inference loss is 14.11, weight deacy loss is 9.12, training accuracy is 0.218750, time 298.915 samples/sec
+epoch 4, total_step 75260, total loss is 23.48 , inference loss is 14.35, weight deacy loss is 9.12, training accuracy is 0.187500, time 297.021 samples/sec
+epoch 4, total_step 75280, total loss is 17.62 , inference loss is 8.50, weight deacy loss is 9.12, training accuracy is 0.281250, time 213.529 samples/sec
+epoch 4, total_step 75300, total loss is 22.35 , inference loss is 13.23, weight deacy loss is 9.12, training accuracy is 0.125000, time 13.142 samples/sec
+epoch 4, total_step 75320, total loss is 24.10 , inference loss is 14.98, weight deacy loss is 9.12, training accuracy is 0.250000, time 213.985 samples/sec
+epoch 4, total_step 75340, total loss is 22.51 , inference loss is 13.39, weight deacy loss is 9.12, training accuracy is 0.187500, time 211.981 samples/sec
+epoch 4, total_step 75360, total loss is 20.14 , inference loss is 11.03, weight deacy loss is 9.12, training accuracy is 0.187500, time 215.625 samples/sec
+epoch 4, total_step 75380, total loss is 19.38 , inference loss is 10.27, weight deacy loss is 9.12, training accuracy is 0.187500, time 211.139 samples/sec
+epoch 4, total_step 75400, total loss is 22.92 , inference loss is 13.81, weight deacy loss is 9.12, training accuracy is 0.125000, time 227.394 samples/sec
+epoch 4, total_step 75420, total loss is 26.09 , inference loss is 16.98, weight deacy loss is 9.12, training accuracy is 0.093750, time 201.094 samples/sec
+epoch 4, total_step 75440, total loss is 21.14 , inference loss is 12.03, weight deacy loss is 9.11, training accuracy is 0.125000, time 218.003 samples/sec
+epoch 4, total_step 75460, total loss is 25.40 , inference loss is 16.29, weight deacy loss is 9.11, training accuracy is 0.187500, time 213.011 samples/sec
+epoch 4, total_step 75480, total loss is 21.79 , inference loss is 12.68, weight deacy loss is 9.11, training accuracy is 0.375000, time 219.034 samples/sec
+epoch 4, total_step 75500, total loss is 20.72 , inference loss is 11.61, weight deacy loss is 9.11, training accuracy is 0.218750, time 207.162 samples/sec
+epoch 4, total_step 75520, total loss is 22.77 , inference loss is 13.66, weight deacy loss is 9.11, training accuracy is 0.093750, time 297.670 samples/sec
+epoch 4, total_step 75540, total loss is 23.31 , inference loss is 14.20, weight deacy loss is 9.11, training accuracy is 0.093750, time 218.260 samples/sec
+epoch 4, total_step 75560, total loss is 24.50 , inference loss is 15.39, weight deacy loss is 9.11, training accuracy is 0.062500, time 203.191 samples/sec
+epoch 4, total_step 75580, total loss is 24.88 , inference loss is 15.77, weight deacy loss is 9.11, training accuracy is 0.156250, time 201.400 samples/sec
+epoch 4, total_step 75600, total loss is 25.40 , inference loss is 16.29, weight deacy loss is 9.11, training accuracy is 0.156250, time 13.933 samples/sec
+epoch 4, total_step 75620, total loss is 24.56 , inference loss is 15.45, weight deacy loss is 9.11, training accuracy is 0.031250, time 221.167 samples/sec
+epoch 4, total_step 75640, total loss is 24.77 , inference loss is 15.67, weight deacy loss is 9.10, training accuracy is 0.125000, time 217.517 samples/sec
+epoch 4, total_step 75660, total loss is 24.34 , inference loss is 15.24, weight deacy loss is 9.10, training accuracy is 0.156250, time 209.809 samples/sec
+epoch 4, total_step 75680, total loss is 20.44 , inference loss is 11.34, weight deacy loss is 9.10, training accuracy is 0.218750, time 212.703 samples/sec
+epoch 4, total_step 75700, total loss is 19.55 , inference loss is 10.45, weight deacy loss is 9.10, training accuracy is 0.250000, time 208.519 samples/sec
+epoch 4, total_step 75720, total loss is 20.25 , inference loss is 11.15, weight deacy loss is 9.10, training accuracy is 0.281250, time 199.502 samples/sec
+epoch 4, total_step 75740, total loss is 22.55 , inference loss is 13.45, weight deacy loss is 9.10, training accuracy is 0.218750, time 215.827 samples/sec
+epoch 4, total_step 75760, total loss is 23.86 , inference loss is 14.76, weight deacy loss is 9.10, training accuracy is 0.156250, time 216.010 samples/sec
+epoch 4, total_step 75780, total loss is 24.21 , inference loss is 15.11, weight deacy loss is 9.10, training accuracy is 0.187500, time 297.853 samples/sec
+epoch 4, total_step 75800, total loss is 23.31 , inference loss is 14.22, weight deacy loss is 9.10, training accuracy is 0.125000, time 210.612 samples/sec2022-12-01 05:16:12.338689: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 8334 of 10000
+2022-12-01 05:16:14.286529: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 4, total_step 75820, total loss is 22.91 , inference loss is 13.81, weight deacy loss is 9.10, training accuracy is 0.187500, time 218.536 samples/sec
+epoch 4, total_step 75840, total loss is 24.86 , inference loss is 15.77, weight deacy loss is 9.09, training accuracy is 0.062500, time 217.114 samples/sec
+epoch 4, total_step 75860, total loss is 24.45 , inference loss is 15.36, weight deacy loss is 9.09, training accuracy is 0.093750, time 210.516 samples/sec
+epoch 4, total_step 75880, total loss is 19.48 , inference loss is 10.38, weight deacy loss is 9.09, training accuracy is 0.125000, time 213.915 samples/sec
+epoch 4, total_step 75900, total loss is 25.35 , inference loss is 16.26, weight deacy loss is 9.09, training accuracy is 0.156250, time 14.373 samples/sec
+epoch 4, total_step 75920, total loss is 22.99 , inference loss is 13.90, weight deacy loss is 9.09, training accuracy is 0.187500, time 213.573 samples/sec
+epoch 4, total_step 75940, total loss is 20.15 , inference loss is 11.06, weight deacy loss is 9.09, training accuracy is 0.125000, time 206.626 samples/sec
+epoch 4, total_step 75960, total loss is 23.70 , inference loss is 14.61, weight deacy loss is 9.09, training accuracy is 0.156250, time 213.960 samples/sec
+epoch 4, total_step 75980, total loss is 19.49 , inference loss is 10.40, weight deacy loss is 9.09, training accuracy is 0.156250, time 212.016 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.674079000000017
+best_threshold_index 105 0.9838888888888889
+best_threshold_index 109 0.9816666666666667
+best_threshold_index 109 0.9833333333333333
+best_threshold_index 109 0.9833333333333333
+best_threshold_index 109 0.9827777777777778
+best_threshold_index 109 0.9812962962962963
+best_threshold_index 109 0.9827777777777778
+best_threshold_index 105 0.9812962962962963
+best_threshold_index 109 0.9807407407407407
+best_threshold_index 109 0.9812962962962963
+[lfw][76000]XNorm: 22.789207
+[lfw][76000]Accuracy-Flip: 0.98083+-0.00973
+test accuracy is: 0.9808333333333333
+epoch 4, total_step 76000, total loss is 26.15 , inference loss is 17.06, weight deacy loss is 9.09, training accuracy is 0.218750, time 19.461 samples/sec
+epoch 4, total_step 76020, total loss is 21.64 , inference loss is 12.56, weight deacy loss is 9.09, training accuracy is 0.156250, time 206.684 samples/sec
+epoch 4, total_step 76040, total loss is 20.74 , inference loss is 11.65, weight deacy loss is 9.09, training accuracy is 0.156250, time 214.403 samples/sec
+epoch 4, total_step 76060, total loss is 20.12 , inference loss is 11.04, weight deacy loss is 9.08, training accuracy is 0.125000, time 207.830 samples/sec
+epoch 4, total_step 76080, total loss is 20.12 , inference loss is 11.04, weight deacy loss is 9.08, training accuracy is 0.187500, time 209.602 samples/sec
+epoch 4, total_step 76100, total loss is 21.76 , inference loss is 12.68, weight deacy loss is 9.08, training accuracy is 0.218750, time 209.028 samples/sec
+epoch 4, total_step 76120, total loss is 21.82 , inference loss is 12.74, weight deacy loss is 9.08, training accuracy is 0.281250, time 219.603 samples/sec
+epoch 4, total_step 76140, total loss is 22.58 , inference loss is 13.50, weight deacy loss is 9.08, training accuracy is 0.187500, time 211.031 samples/sec
+epoch 4, total_step 76160, total loss is 19.73 , inference loss is 10.65, weight deacy loss is 9.08, training accuracy is 0.031250, time 299.074 samples/sec
+epoch 4, total_step 76180, total loss is 20.38 , inference loss is 11.30, weight deacy loss is 9.08, training accuracy is 0.187500, time 211.886 samples/sec
+epoch 4, total_step 76200, total loss is 22.23 , inference loss is 13.15, weight deacy loss is 9.08, training accuracy is 0.156250, time 11.990 samples/sec
+epoch 4, total_step 76220, total loss is 26.17 , inference loss is 17.09, weight deacy loss is 9.08, training accuracy is 0.218750, time 201.807 samples/sec
+epoch 4, total_step 76240, total loss is 18.99 , inference loss is 9.91, weight deacy loss is 9.08, training accuracy is 0.250000, time 215.058 samples/sec
+epoch 4, total_step 76260, total loss is 21.56 , inference loss is 12.49, weight deacy loss is 9.07, training accuracy is 0.156250, time 211.277 samples/sec
+epoch 4, total_step 76280, total loss is 23.69 , inference loss is 14.61, weight deacy loss is 9.07, training accuracy is 0.125000, time 212.535 samples/sec
+epoch 4, total_step 76300, total loss is 22.13 , inference loss is 13.05, weight deacy loss is 9.07, training accuracy is 0.093750, time 212.155 samples/sec
+epoch 4, total_step 76320, total loss is 20.45 , inference loss is 11.38, weight deacy loss is 9.07, training accuracy is 0.250000, time 209.592 samples/sec
+epoch 4, total_step 76340, total loss is 23.28 , inference loss is 14.21, weight deacy loss is 9.07, training accuracy is 0.093750, time 231.456 samples/sec
+epoch 4, total_step 76360, total loss is 21.38 , inference loss is 12.31, weight deacy loss is 9.07, training accuracy is 0.187500, time 206.610 samples/sec
+epoch 4, total_step 76380, total loss is 27.24 , inference loss is 18.17, weight deacy loss is 9.07, training accuracy is 0.125000, time 200.216 samples/sec
+epoch 4, total_step 76400, total loss is 19.13 , inference loss is 10.06, weight deacy loss is 9.07, training accuracy is 0.187500, time 197.393 samples/sec
+epoch 4, total_step 76420, total loss is 19.41 , inference loss is 10.34, weight deacy loss is 9.07, training accuracy is 0.250000, time 196.225 samples/sec
+epoch 4, total_step 76440, total loss is 19.60 , inference loss is 10.54, weight deacy loss is 9.07, training accuracy is 0.187500, time 226.434 samples/sec
+epoch 4, total_step 76460, total loss is 22.07 , inference loss is 13.00, weight deacy loss is 9.07, training accuracy is 0.250000, time 296.098 samples/sec
+epoch 4, total_step 76480, total loss is 20.17 , inference loss is 11.11, weight deacy loss is 9.06, training accuracy is 0.156250, time 202.849 samples/sec
+epoch 4, total_step 76500, total loss is 23.75 , inference loss is 14.68, weight deacy loss is 9.06, training accuracy is 0.156250, time 13.643 samples/sec
+epoch 4, total_step 76520, total loss is 24.01 , inference loss is 14.94, weight deacy loss is 9.06, training accuracy is 0.125000, time 207.371 samples/sec
+epoch 4, total_step 76540, total loss is 19.30 , inference loss is 10.24, weight deacy loss is 9.06, training accuracy is 0.312500, time 204.962 samples/sec
+epoch 4, total_step 76560, total loss is 23.74 , inference loss is 14.68, weight deacy loss is 9.06, training accuracy is 0.156250, time 195.098 samples/sec
+epoch 4, total_step 76580, total loss is 18.79 , inference loss is 9.73, weight deacy loss is 9.06, training accuracy is 0.250000, time 197.146 samples/sec
+epoch 4, total_step 76600, total loss is 21.62 , inference loss is 12.56, weight deacy loss is 9.06, training accuracy is 0.156250, time 232.408 samples/sec
+epoch 4, total_step 76620, total loss is 19.43 , inference loss is 10.37, weight deacy loss is 9.06, training accuracy is 0.187500, time 195.929 samples/sec
+epoch 4, total_step 76640, total loss is 20.58 , inference loss is 11.53, weight deacy loss is 9.06, training accuracy is 0.343750, time 207.733 samples/sec
+End of epoch 4
+epoch 5, total_step 76660, total loss is 31.67 , inference loss is 22.62, weight deacy loss is 9.06, training accuracy is 0.000000, time 213.622 samples/sec
+epoch 5, total_step 76680, total loss is 23.10 , inference loss is 14.05, weight deacy loss is 9.05, training accuracy is 0.218750, time 203.897 samples/sec
+epoch 5, total_step 76700, total loss is 18.12 , inference loss is 9.06, weight deacy loss is 9.05, training accuracy is 0.375000, time 298.183 samples/sec
+epoch 5, total_step 76720, total loss is 18.84 , inference loss is 9.79, weight deacy loss is 9.05, training accuracy is 0.375000, time 210.864 samples/sec
+epoch 5, total_step 76740, total loss is 16.49 , inference loss is 7.44, weight deacy loss is 9.05, training accuracy is 0.437500, time 212.479 samples/sec
+epoch 5, total_step 76760, total loss is 17.22 , inference loss is 8.17, weight deacy loss is 9.05, training accuracy is 0.531250, time 207.457 samples/sec
+epoch 5, total_step 76780, total loss is 21.53 , inference loss is 12.48, weight deacy loss is 9.05, training accuracy is 0.468750, time 215.807 samples/sec
+epoch 5, total_step 76800, total loss is 17.74 , inference loss is 8.70, weight deacy loss is 9.05, training accuracy is 0.375000, time 10.852 samples/sec
+epoch 5, total_step 76820, total loss is 17.56 , inference loss is 8.51, weight deacy loss is 9.05, training accuracy is 0.437500, time 208.987 samples/sec
+epoch 5, total_step 76840, total loss is 19.21 , inference loss is 10.17, weight deacy loss is 9.05, training accuracy is 0.406250, time 212.495 samples/sec
+epoch 5, total_step 76860, total loss is 17.16 , inference loss is 8.11, weight deacy loss is 9.05, training accuracy is 0.500000, time 217.900 samples/sec
+epoch 5, total_step 76880, total loss is 15.28 , inference loss is 6.23, weight deacy loss is 9.04, training accuracy is 0.468750, time 215.481 samples/sec
+epoch 5, total_step 76900, total loss is 19.68 , inference loss is 10.64, weight deacy loss is 9.04, training accuracy is 0.343750, time 217.722 samples/sec
+epoch 5, total_step 76920, total loss is 17.84 , inference loss is 8.80, weight deacy loss is 9.04, training accuracy is 0.406250, time 248.485 samples/sec
+epoch 5, total_step 76940, total loss is 15.16 , inference loss is 6.12, weight deacy loss is 9.04, training accuracy is 0.500000, time 219.597 samples/sec
+epoch 5, total_step 76960, total loss is 15.96 , inference loss is 6.92, weight deacy loss is 9.04, training accuracy is 0.437500, time 215.837 samples/sec
+epoch 5, total_step 76980, total loss is 20.31 , inference loss is 11.27, weight deacy loss is 9.04, training accuracy is 0.343750, time 225.624 samples/sec
+epoch 5, total_step 77000, total loss is 20.89 , inference loss is 11.85, weight deacy loss is 9.04, training accuracy is 0.312500, time 296.996 samples/sec
+epoch 5, total_step 77020, total loss is 19.68 , inference loss is 10.64, weight deacy loss is 9.04, training accuracy is 0.406250, time 296.996 samples/sec
+epoch 5, total_step 77040, total loss is 17.62 , inference loss is 8.59, weight deacy loss is 9.04, training accuracy is 0.375000, time 208.130 samples/sec
+epoch 5, total_step 77060, total loss is 21.78 , inference loss is 12.74, weight deacy loss is 9.04, training accuracy is 0.375000, time 202.823 samples/sec
+epoch 5, total_step 77080, total loss is 15.11 , inference loss is 6.08, weight deacy loss is 9.03, training accuracy is 0.593750, time 210.509 samples/sec
+epoch 5, total_step 77100, total loss is 16.54 , inference loss is 7.51, weight deacy loss is 9.03, training accuracy is 0.468750, time 14.546 samples/sec
+epoch 5, total_step 77120, total loss is 16.42 , inference loss is 7.39, weight deacy loss is 9.03, training accuracy is 0.375000, time 215.860 samples/sec
+epoch 5, total_step 77140, total loss is 23.42 , inference loss is 14.39, weight deacy loss is 9.03, training accuracy is 0.375000, time 212.919 samples/sec
+epoch 5, total_step 77160, total loss is 20.61 , inference loss is 11.58, weight deacy loss is 9.03, training accuracy is 0.343750, time 209.160 samples/sec
+epoch 5, total_step 77180, total loss is 21.29 , inference loss is 12.26, weight deacy loss is 9.03, training accuracy is 0.437500, time 209.839 samples/sec
+epoch 5, total_step 77200, total loss is 16.74 , inference loss is 7.71, weight deacy loss is 9.03, training accuracy is 0.562500, time 210.974 samples/sec
+epoch 5, total_step 77220, total loss is 18.99 , inference loss is 9.96, weight deacy loss is 9.03, training accuracy is 0.406250, time 204.344 samples/sec
+epoch 5, total_step 77240, total loss is 17.95 , inference loss is 8.92, weight deacy loss is 9.03, training accuracy is 0.562500, time 224.226 samples/sec
+epoch 5, total_step 77260, total loss is 20.28 , inference loss is 11.25, weight deacy loss is 9.03, training accuracy is 0.343750, time 218.705 samples/sec
+epoch 5, total_step 77280, total loss is 19.58 , inference loss is 10.55, weight deacy loss is 9.02, training accuracy is 0.343750, time 220.296 samples/sec
+epoch 5, total_step 77300, total loss is 18.25 , inference loss is 9.23, weight deacy loss is 9.02, training accuracy is 0.281250, time 206.817 samples/sec
+epoch 5, total_step 77320, total loss is 16.27 , inference loss is 7.25, weight deacy loss is 9.02, training accuracy is 0.468750, time 215.311 samples/sec
+epoch 5, total_step 77340, total loss is 20.14 , inference loss is 11.12, weight deacy loss is 9.02, training accuracy is 0.437500, time 216.381 samples/sec
+epoch 5, total_step 77360, total loss is 20.94 , inference loss is 11.92, weight deacy loss is 9.02, training accuracy is 0.312500, time 209.049 samples/sec
+epoch 5, total_step 77380, total loss is 18.78 , inference loss is 9.75, weight deacy loss is 9.02, training accuracy is 0.312500, time 213.756 samples/sec
+epoch 5, total_step 77400, total loss is 24.18 , inference loss is 15.16, weight deacy loss is 9.02, training accuracy is 0.250000, time 13.049 samples/sec
+epoch 5, total_step 77420, total loss is 15.72 , inference loss is 6.70, weight deacy loss is 9.02, training accuracy is 0.500000, time 208.155 samples/sec
+epoch 5, total_step 77440, total loss is 17.10 , inference loss is 8.08, weight deacy loss is 9.02, training accuracy is 0.531250, time 214.234 samples/sec
+epoch 5, total_step 77460, total loss is 19.82 , inference loss is 10.80, weight deacy loss is 9.02, training accuracy is 0.343750, time 216.362 samples/sec
+epoch 5, total_step 77480, total loss is 19.22 , inference loss is 10.21, weight deacy loss is 9.02, training accuracy is 0.437500, time 219.295 samples/sec
+epoch 5, total_step 77500, total loss is 17.47 , inference loss is 8.46, weight deacy loss is 9.01, training accuracy is 0.375000, time 215.626 samples/sec
+epoch 5, total_step 77520, total loss is 21.09 , inference loss is 12.07, weight deacy loss is 9.01, training accuracy is 0.375000, time 208.552 samples/sec
+epoch 5, total_step 77540, total loss is 17.03 , inference loss is 8.02, weight deacy loss is 9.01, training accuracy is 0.406250, time 218.608 samples/sec
+epoch 5, total_step 77560, total loss is 17.47 , inference loss is 8.46, weight deacy loss is 9.01, training accuracy is 0.375000, time 211.183 samples/sec
+epoch 5, total_step 77580, total loss is 16.46 , inference loss is 7.44, weight deacy loss is 9.01, training accuracy is 0.437500, time 222.090 samples/sec
+epoch 5, total_step 77600, total loss is 17.38 , inference loss is 8.37, weight deacy loss is 9.01, training accuracy is 0.375000, time 223.990 samples/sec
+epoch 5, total_step 77620, total loss is 18.16 , inference loss is 9.15, weight deacy loss is 9.01, training accuracy is 0.406250, time 212.098 samples/sec
+epoch 5, total_step 77640, total loss is 22.05 , inference loss is 13.05, weight deacy loss is 9.01, training accuracy is 0.250000, time 215.930 samples/sec
+epoch 5, total_step 77660, total loss is 20.27 , inference loss is 11.26, weight deacy loss is 9.01, training accuracy is 0.406250, time 213.789 samples/sec
+epoch 5, total_step 77680, total loss is 19.79 , inference loss is 10.78, weight deacy loss is 9.01, training accuracy is 0.437500, time 297.705 samples/sec
+epoch 5, total_step 77700, total loss is 19.16 , inference loss is 10.16, weight deacy loss is 9.01, training accuracy is 0.406250, time 12.984 samples/sec
+epoch 5, total_step 77720, total loss is 17.13 , inference loss is 8.13, weight deacy loss is 9.01, training accuracy is 0.312500, time 204.853 samples/sec
+epoch 5, total_step 77740, total loss is 18.14 , inference loss is 9.14, weight deacy loss is 9.00, training accuracy is 0.312500, time 203.791 samples/sec
+epoch 5, total_step 77760, total loss is 18.57 , inference loss is 9.57, weight deacy loss is 9.00, training accuracy is 0.406250, time 215.460 samples/sec
+epoch 5, total_step 77780, total loss is 18.59 , inference loss is 9.59, weight deacy loss is 9.00, training accuracy is 0.312500, time 219.025 samples/sec
+epoch 5, total_step 77800, total loss is 23.77 , inference loss is 14.77, weight deacy loss is 9.00, training accuracy is 0.406250, time 219.898 samples/sec
+epoch 5, total_step 77820, total loss is 18.39 , inference loss is 9.39, weight deacy loss is 9.00, training accuracy is 0.406250, time 225.943 samples/sec
+epoch 5, total_step 77840, total loss is 18.48 , inference loss is 9.48, weight deacy loss is 9.00, training accuracy is 0.343750, time 217.374 samples/sec
+epoch 5, total_step 77860, total loss is 17.69 , inference loss is 8.70, weight deacy loss is 9.00, training accuracy is 0.312500, time 222.483 samples/sec
+epoch 5, total_step 77880, total loss is 18.75 , inference loss is 9.75, weight deacy loss is 9.00, training accuracy is 0.312500, time 217.298 samples/sec
+epoch 5, total_step 77900, total loss is 25.37 , inference loss is 16.38, weight deacy loss is 9.00, training accuracy is 0.250000, time 210.807 samples/sec
+epoch 5, total_step 77920, total loss is 19.66 , inference loss is 10.66, weight deacy loss is 9.00, training accuracy is 0.250000, time 211.170 samples/sec
+epoch 5, total_step 77940, total loss is 21.31 , inference loss is 12.32, weight deacy loss is 9.00, training accuracy is 0.250000, time 187.322 samples/sec
+epoch 5, total_step 77960, total loss is 16.04 , inference loss is 7.05, weight deacy loss is 8.99, training accuracy is 0.500000, time 193.934 samples/sec
+epoch 5, total_step 77980, total loss is 21.69 , inference loss is 12.70, weight deacy loss is 8.99, training accuracy is 0.312500, time 214.822 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.307763999999995
+best_threshold_index 123 0.985
+best_threshold_index 121 0.9844444444444445
+best_threshold_index 114 0.9846296296296296
+best_threshold_index 121 0.9851851851851852
+best_threshold_index 121 0.9840740740740741
+best_threshold_index 121 0.9842592592592593
+best_threshold_index 114 0.9844444444444445
+best_threshold_index 121 0.9844444444444445
+best_threshold_index 121 0.9838888888888889
+best_threshold_index 121 0.9838888888888889
+[lfw][78000]XNorm: 22.340605
+[lfw][78000]Accuracy-Flip: 0.98250+-0.00455
+test accuracy is: 0.9824999999999999
+epoch 5, total_step 78000, total loss is 21.74 , inference loss is 12.75, weight deacy loss is 8.99, training accuracy is 0.187500, time 19.523 samples/sec
+epoch 5, total_step 78020, total loss is 19.12 , inference loss is 10.13, weight deacy loss is 8.99, training accuracy is 0.375000, time 210.381 samples/sec
+epoch 5, total_step 78040, total loss is 19.96 , inference loss is 10.97, weight deacy loss is 8.99, training accuracy is 0.406250, time 207.489 samples/sec
+epoch 5, total_step 78060, total loss is 19.60 , inference loss is 10.61, weight deacy loss is 8.99, training accuracy is 0.468750, time 200.268 samples/sec
+epoch 5, total_step 78080, total loss is 22.98 , inference loss is 14.00, weight deacy loss is 8.99, training accuracy is 0.156250, time 213.947 samples/sec
+epoch 5, total_step 78100, total loss is 19.19 , inference loss is 10.21, weight deacy loss is 8.99, training accuracy is 0.250000, time 205.675 samples/sec
+epoch 5, total_step 78120, total loss is 19.56 , inference loss is 10.58, weight deacy loss is 8.99, training accuracy is 0.281250, time 210.067 samples/sec
+epoch 5, total_step 78140, total loss is 22.65 , inference loss is 13.66, weight deacy loss is 8.99, training accuracy is 0.250000, time 210.496 samples/sec
+epoch 5, total_step 78160, total loss is 23.87 , inference loss is 14.89, weight deacy loss is 8.99, training accuracy is 0.343750, time 206.434 samples/sec
+epoch 5, total_step 78180, total loss is 18.96 , inference loss is 9.98, weight deacy loss is 8.98, training accuracy is 0.312500, time 213.220 samples/sec
+epoch 5, total_step 78200, total loss is 17.47 , inference loss is 8.49, weight deacy loss is 8.98, training accuracy is 0.375000, time 206.211 samples/sec
+epoch 5, total_step 78220, total loss is 19.83 , inference loss is 10.84, weight deacy loss is 8.98, training accuracy is 0.187500, time 211.665 samples/sec
+epoch 5, total_step 78240, total loss is 17.37 , inference loss is 8.39, weight deacy loss is 8.98, training accuracy is 0.375000, time 211.595 samples/sec
+epoch 5, total_step 78260, total loss is 20.93 , inference loss is 11.95, weight deacy loss is 8.98, training accuracy is 0.343750, time 217.109 samples/sec
+epoch 5, total_step 78280, total loss is 22.33 , inference loss is 13.35, weight deacy loss is 8.98, training accuracy is 0.218750, time 214.921 samples/sec
+epoch 5, total_step 78300, total loss is 16.52 , inference loss is 7.54, weight deacy loss is 8.98, training accuracy is 0.312500, time 14.682 samples/sec
+epoch 5, total_step 78320, total loss is 21.98 , inference loss is 13.00, weight deacy loss is 8.98, training accuracy is 0.250000, time 208.664 samples/sec
+epoch 5, total_step 78340, total loss is 21.37 , inference loss is 12.39, weight deacy loss is 8.98, training accuracy is 0.250000, time 218.106 samples/sec
+epoch 5, total_step 78360, total loss is 21.38 , inference loss is 12.40, weight deacy loss is 8.98, training accuracy is 0.312500, time 207.955 samples/sec
+epoch 5, total_step 78380, total loss is 19.86 , inference loss is 10.89, weight deacy loss is 8.98, training accuracy is 0.375000, time 209.823 samples/sec
+epoch 5, total_step 78400, total loss is 18.95 , inference loss is 9.98, weight deacy loss is 8.97, training accuracy is 0.312500, time 218.679 samples/sec
+epoch 5, total_step 78420, total loss is 13.97 , inference loss is 4.99, weight deacy loss is 8.97, training accuracy is 0.375000, time 219.884 samples/sec
+epoch 5, total_step 78440, total loss is 17.92 , inference loss is 8.94, weight deacy loss is 8.97, training accuracy is 0.312500, time 209.983 samples/sec
+epoch 5, total_step 78460, total loss is 18.77 , inference loss is 9.80, weight deacy loss is 8.97, training accuracy is 0.406250, time 216.026 samples/sec
+epoch 5, total_step 78480, total loss is 17.83 , inference loss is 8.86, weight deacy loss is 8.97, training accuracy is 0.343750, time 214.239 samples/sec
+epoch 5, total_step 78500, total loss is 21.51 , inference loss is 12.54, weight deacy loss is 8.97, training accuracy is 0.281250, time 204.766 samples/sec
+epoch 5, total_step 78520, total loss is 19.78 , inference loss is 10.81, weight deacy loss is 8.97, training accuracy is 0.375000, time 215.831 samples/sec
+epoch 5, total_step 78540, total loss is 19.30 , inference loss is 10.33, weight deacy loss is 8.97, training accuracy is 0.406250, time 206.078 samples/sec
+epoch 5, total_step 78560, total loss is 17.69 , inference loss is 8.72, weight deacy loss is 8.97, training accuracy is 0.343750, time 202.067 samples/sec
+epoch 5, total_step 78580, total loss is 16.73 , inference loss is 7.76, weight deacy loss is 8.97, training accuracy is 0.375000, time 209.276 samples/sec
+epoch 5, total_step 78600, total loss is 18.95 , inference loss is 9.99, weight deacy loss is 8.96, training accuracy is 0.218750, time 13.978 samples/sec
+epoch 5, total_step 78620, total loss is 17.32 , inference loss is 8.36, weight deacy loss is 8.96, training accuracy is 0.468750, time 206.758 samples/sec
+epoch 5, total_step 78640, total loss is 18.39 , inference loss is 9.43, weight deacy loss is 8.96, training accuracy is 0.312500, time 210.164 samples/sec
+epoch 5, total_step 78660, total loss is 15.43 , inference loss is 6.47, weight deacy loss is 8.96, training accuracy is 0.437500, time 205.440 samples/sec
+epoch 5, total_step 78680, total loss is 17.50 , inference loss is 8.53, weight deacy loss is 8.96, training accuracy is 0.437500, time 214.743 samples/sec
+epoch 5, total_step 78700, total loss is 19.33 , inference loss is 10.37, weight deacy loss is 8.96, training accuracy is 0.406250, time 215.551 samples/sec
+epoch 5, total_step 78720, total loss is 17.93 , inference loss is 8.97, weight deacy loss is 8.96, training accuracy is 0.343750, time 212.323 samples/sec
+epoch 5, total_step 78740, total loss is 17.77 , inference loss is 8.81, weight deacy loss is 8.96, training accuracy is 0.281250, time 211.824 samples/sec
+epoch 5, total_step 78760, total loss is 20.58 , inference loss is 11.63, weight deacy loss is 8.96, training accuracy is 0.343750, time 212.801 samples/sec
+epoch 5, total_step 78780, total loss is 20.49 , inference loss is 11.53, weight deacy loss is 8.96, training accuracy is 0.343750, time 221.300 samples/sec
+epoch 5, total_step 78800, total loss is 17.07 , inference loss is 8.12, weight deacy loss is 8.95, training accuracy is 0.375000, time 213.301 samples/sec
+epoch 5, total_step 78820, total loss is 20.07 , inference loss is 11.11, weight deacy loss is 8.95, training accuracy is 0.375000, time 299.104 samples/sec
+epoch 5, total_step 78840, total loss is 16.61 , inference loss is 7.66, weight deacy loss is 8.95, training accuracy is 0.375000, time 296.664 samples/sec
+epoch 5, total_step 78860, total loss is 19.69 , inference loss is 10.74, weight deacy loss is 8.95, training accuracy is 0.343750, time 209.511 samples/sec
+epoch 5, total_step 78880, total loss is 14.97 , inference loss is 6.02, weight deacy loss is 8.95, training accuracy is 0.500000, time 207.755 samples/sec
+epoch 5, total_step 78900, total loss is 16.56 , inference loss is 7.61, weight deacy loss is 8.95, training accuracy is 0.562500, time 10.922 samples/sec
+epoch 5, total_step 78920, total loss is 15.08 , inference loss is 6.13, weight deacy loss is 8.95, training accuracy is 0.375000, time 215.337 samples/sec
+epoch 5, total_step 78940, total loss is 17.84 , inference loss is 8.89, weight deacy loss is 8.95, training accuracy is 0.406250, time 219.294 samples/sec
+epoch 5, total_step 78960, total loss is 17.37 , inference loss is 8.42, weight deacy loss is 8.95, training accuracy is 0.531250, time 210.821 samples/sec
+epoch 5, total_step 78980, total loss is 17.27 , inference loss is 8.33, weight deacy loss is 8.95, training accuracy is 0.406250, time 211.261 samples/sec
+epoch 5, total_step 79000, total loss is 20.34 , inference loss is 11.40, weight deacy loss is 8.94, training accuracy is 0.281250, time 296.425 samples/sec
+epoch 5, total_step 79020, total loss is 15.47 , inference loss is 6.52, weight deacy loss is 8.94, training accuracy is 0.531250, time 236.218 samples/sec
+epoch 5, total_step 79040, total loss is 17.53 , inference loss is 8.59, weight deacy loss is 8.94, training accuracy is 0.312500, time 235.826 samples/sec
+epoch 5, total_step 79060, total loss is 15.43 , inference loss is 6.49, weight deacy loss is 8.94, training accuracy is 0.406250, time 202.508 samples/sec
+epoch 5, total_step 79080, total loss is 18.36 , inference loss is 9.42, weight deacy loss is 8.94, training accuracy is 0.406250, time 204.431 samples/sec
+epoch 5, total_step 79100, total loss is 16.72 , inference loss is 7.78, weight deacy loss is 8.94, training accuracy is 0.437500, time 202.012 samples/sec
+epoch 5, total_step 79120, total loss is 24.02 , inference loss is 15.09, weight deacy loss is 8.94, training accuracy is 0.218750, time 218.558 samples/sec
+epoch 5, total_step 79140, total loss is 16.97 , inference loss is 8.03, weight deacy loss is 8.94, training accuracy is 0.312500, time 227.007 samples/sec
+epoch 5, total_step 79160, total loss is 14.83 , inference loss is 5.90, weight deacy loss is 8.94, training accuracy is 0.468750, time 210.883 samples/sec
+epoch 5, total_step 79180, total loss is 17.28 , inference loss is 8.34, weight deacy loss is 8.94, training accuracy is 0.593750, time 207.026 samples/sec
+epoch 5, total_step 79200, total loss is 18.52 , inference loss is 9.59, weight deacy loss is 8.93, training accuracy is 0.375000, time 13.460 samples/sec
+epoch 5, total_step 79220, total loss is 15.74 , inference loss is 6.80, weight deacy loss is 8.93, training accuracy is 0.281250, time 212.331 samples/sec
+epoch 5, total_step 79240, total loss is 18.10 , inference loss is 9.16, weight deacy loss is 8.93, training accuracy is 0.500000, time 217.294 samples/sec
+epoch 5, total_step 79260, total loss is 16.09 , inference loss is 7.16, weight deacy loss is 8.93, training accuracy is 0.531250, time 211.911 samples/sec
+epoch 5, total_step 79280, total loss is 15.28 , inference loss is 6.35, weight deacy loss is 8.93, training accuracy is 0.406250, time 297.059 samples/sec
+epoch 5, total_step 79300, total loss is 14.43 , inference loss is 5.50, weight deacy loss is 8.93, training accuracy is 0.562500, time 227.276 samples/sec
+epoch 5, total_step 79320, total loss is 13.81 , inference loss is 4.88, weight deacy loss is 8.93, training accuracy is 0.468750, time 209.536 samples/sec
+epoch 5, total_step 79340, total loss is 17.35 , inference loss is 8.42, weight deacy loss is 8.93, training accuracy is 0.312500, time 210.131 samples/sec
+epoch 5, total_step 79360, total loss is 16.31 , inference loss is 7.38, weight deacy loss is 8.93, training accuracy is 0.375000, time 217.436 samples/sec
+epoch 5, total_step 79380, total loss is 16.43 , inference loss is 7.51, weight deacy loss is 8.93, training accuracy is 0.468750, time 250.200 samples/sec
+epoch 5, total_step 79400, total loss is 19.63 , inference loss is 10.70, weight deacy loss is 8.92, training accuracy is 0.375000, time 214.664 samples/sec
+epoch 5, total_step 79420, total loss is 21.84 , inference loss is 12.91, weight deacy loss is 8.92, training accuracy is 0.406250, time 212.466 samples/sec
+epoch 5, total_step 79440, total loss is 12.36 , inference loss is 3.44, weight deacy loss is 8.92, training accuracy is 0.562500, time 214.253 samples/sec
+epoch 5, total_step 79460, total loss is 15.45 , inference loss is 6.52, weight deacy loss is 8.92, training accuracy is 0.437500, time 225.843 samples/sec
+epoch 5, total_step 79480, total loss is 14.55 , inference loss is 5.63, weight deacy loss is 8.92, training accuracy is 0.531250, time 222.778 samples/sec
+epoch 5, total_step 79500, total loss is 16.94 , inference loss is 8.02, weight deacy loss is 8.92, training accuracy is 0.437500, time 13.845 samples/sec
+epoch 5, total_step 79520, total loss is 20.78 , inference loss is 11.86, weight deacy loss is 8.92, training accuracy is 0.375000, time 204.853 samples/sec
+epoch 5, total_step 79540, total loss is 14.61 , inference loss is 5.69, weight deacy loss is 8.92, training accuracy is 0.406250, time 298.808 samples/sec
+epoch 5, total_step 79560, total loss is 16.61 , inference loss is 7.69, weight deacy loss is 8.92, training accuracy is 0.562500, time 220.802 samples/sec
+epoch 5, total_step 79580, total loss is 17.04 , inference loss is 8.12, weight deacy loss is 8.92, training accuracy is 0.531250, time 227.021 samples/sec
+epoch 5, total_step 79600, total loss is 17.26 , inference loss is 8.35, weight deacy loss is 8.92, training accuracy is 0.468750, time 213.419 samples/sec
+epoch 5, total_step 79620, total loss is 18.21 , inference loss is 9.29, weight deacy loss is 8.91, training accuracy is 0.281250, time 213.266 samples/sec
+epoch 5, total_step 79640, total loss is 17.85 , inference loss is 8.94, weight deacy loss is 8.91, training accuracy is 0.281250, time 208.645 samples/sec
+epoch 5, total_step 79660, total loss is 21.14 , inference loss is 12.23, weight deacy loss is 8.91, training accuracy is 0.406250, time 210.582 samples/sec
+epoch 5, total_step 79680, total loss is 14.47 , inference loss is 5.56, weight deacy loss is 8.91, training accuracy is 0.437500, time 207.415 samples/sec
+epoch 5, total_step 79700, total loss is 13.96 , inference loss is 5.05, weight deacy loss is 8.91, training accuracy is 0.468750, time 211.189 samples/sec
+epoch 5, total_step 79720, total loss is 18.51 , inference loss is 9.60, weight deacy loss is 8.91, training accuracy is 0.406250, time 206.477 samples/sec
+epoch 5, total_step 79740, total loss is 17.03 , inference loss is 8.12, weight deacy loss is 8.91, training accuracy is 0.406250, time 212.369 samples/sec
+epoch 5, total_step 79760, total loss is 19.20 , inference loss is 10.30, weight deacy loss is 8.91, training accuracy is 0.406250, time 217.264 samples/sec
+epoch 5, total_step 79780, total loss is 15.99 , inference loss is 7.09, weight deacy loss is 8.91, training accuracy is 0.531250, time 217.510 samples/sec
+epoch 5, total_step 79800, total loss is 18.08 , inference loss is 9.18, weight deacy loss is 8.91, training accuracy is 0.468750, time 13.539 samples/sec
+epoch 5, total_step 79820, total loss is 18.44 , inference loss is 9.54, weight deacy loss is 8.90, training accuracy is 0.437500, time 299.756 samples/sec
+epoch 5, total_step 79840, total loss is 17.15 , inference loss is 8.25, weight deacy loss is 8.90, training accuracy is 0.500000, time 207.454 samples/sec
+epoch 5, total_step 79860, total loss is 20.40 , inference loss is 11.50, weight deacy loss is 8.90, training accuracy is 0.406250, time 211.719 samples/sec
+epoch 5, total_step 79880, total loss is 19.43 , inference loss is 10.53, weight deacy loss is 8.90, training accuracy is 0.343750, time 211.551 samples/sec
+epoch 5, total_step 79900, total loss is 16.55 , inference loss is 7.65, weight deacy loss is 8.90, training accuracy is 0.468750, time 211.317 samples/sec
+epoch 5, total_step 79920, total loss is 15.04 , inference loss is 6.14, weight deacy loss is 8.90, training accuracy is 0.406250, time 210.247 samples/sec
+epoch 5, total_step 79940, total loss is 20.27 , inference loss is 11.37, weight deacy loss is 8.90, training accuracy is 0.375000, time 206.533 samples/sec
+epoch 5, total_step 79960, total loss is 17.27 , inference loss is 8.38, weight deacy loss is 8.90, training accuracy is 0.375000, time 216.667 samples/sec
+epoch 5, total_step 79980, total loss is 17.50 , inference loss is 8.61, weight deacy loss is 8.90, training accuracy is 0.437500, time 213.664 samples/sec
+testing verification..
+(12000, 512)
+infer time 22.947141999999985
+best_threshold_index 115 0.9879629629629629
+best_threshold_index 115 0.9859259259259259
+best_threshold_index 115 0.9875925925925926
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9875925925925926
+best_threshold_index 115 0.9862962962962963
+best_threshold_index 116 0.9872222222222222
+best_threshold_index 115 0.9866666666666667
+best_threshold_index 115 0.9855555555555555
+best_threshold_index 115 0.9864814814814815
+[lfw][80000]XNorm: 21.455762
+[lfw][80000]Accuracy-Flip: 0.98650+-0.00685
+test accuracy is: 0.9865
+epoch 5, total_step 80000, total loss is 22.04 , inference loss is 13.14, weight deacy loss is 8.90, training accuracy is 0.156250, time 18.047 samples/sec
+epoch 5, total_step 80020, total loss is 17.62 , inference loss is 8.72, weight deacy loss is 8.90, training accuracy is 0.312500, time 213.477 samples/sec
+epoch 5, total_step 80040, total loss is 13.88 , inference loss is 4.99, weight deacy loss is 8.89, training accuracy is 0.437500, time 212.209 samples/sec
+epoch 5, total_step 80060, total loss is 20.21 , inference loss is 11.32, weight deacy loss is 8.89, training accuracy is 0.375000, time 219.263 samples/sec
+epoch 5, total_step 80080, total loss is 14.35 , inference loss is 5.45, weight deacy loss is 8.89, training accuracy is 0.375000, time 210.237 samples/sec
+epoch 5, total_step 80100, total loss is 22.86 , inference loss is 13.97, weight deacy loss is 8.89, training accuracy is 0.343750, time 13.364 samples/sec
+epoch 5, total_step 80120, total loss is 19.91 , inference loss is 11.02, weight deacy loss is 8.89, training accuracy is 0.375000, time 217.511 samples/sec
+epoch 5, total_step 80140, total loss is 17.94 , inference loss is 9.05, weight deacy loss is 8.89, training accuracy is 0.375000, time 210.415 samples/sec
+epoch 5, total_step 80160, total loss is 18.77 , inference loss is 9.88, weight deacy loss is 8.89, training accuracy is 0.437500, time 294.382 samples/sec
+epoch 5, total_step 80180, total loss is 19.12 , inference loss is 10.23, weight deacy loss is 8.89, training accuracy is 0.250000, time 220.023 samples/sec
+epoch 5, total_step 80200, total loss is 16.93 , inference loss is 8.05, weight deacy loss is 8.89, training accuracy is 0.437500, time 205.523 samples/sec
+epoch 5, total_step 80220, total loss is 17.58 , inference loss is 8.69, weight deacy loss is 8.89, training accuracy is 0.312500, time 209.554 samples/sec
+epoch 5, total_step 80240, total loss is 20.81 , inference loss is 11.92, weight deacy loss is 8.89, training accuracy is 0.218750, time 225.556 samples/sec
+epoch 5, total_step 80260, total loss is 22.30 , inference loss is 13.41, weight deacy loss is 8.88, training accuracy is 0.281250, time 216.821 samples/sec
+epoch 5, total_step 80280, total loss is 20.94 , inference loss is 12.06, weight deacy loss is 8.88, training accuracy is 0.375000, time 215.695 samples/sec
+epoch 5, total_step 80300, total loss is 21.60 , inference loss is 12.72, weight deacy loss is 8.88, training accuracy is 0.250000, time 220.116 samples/sec
+epoch 5, total_step 80320, total loss is 16.70 , inference loss is 7.82, weight deacy loss is 8.88, training accuracy is 0.343750, time 213.945 samples/sec
+epoch 5, total_step 80340, total loss is 19.45 , inference loss is 10.57, weight deacy loss is 8.88, training accuracy is 0.437500, time 210.553 samples/sec
+epoch 5, total_step 80360, total loss is 17.32 , inference loss is 8.44, weight deacy loss is 8.88, training accuracy is 0.343750, time 207.961 samples/sec
+epoch 5, total_step 80380, total loss is 18.33 , inference loss is 9.45, weight deacy loss is 8.88, training accuracy is 0.437500, time 211.442 samples/sec
+epoch 5, total_step 80400, total loss is 20.72 , inference loss is 11.84, weight deacy loss is 8.88, training accuracy is 0.343750, time 13.810 samples/sec
+epoch 5, total_step 80420, total loss is 23.39 , inference loss is 14.51, weight deacy loss is 8.88, training accuracy is 0.250000, time 299.667 samples/sec
+epoch 5, total_step 80440, total loss is 22.19 , inference loss is 13.31, weight deacy loss is 8.88, training accuracy is 0.250000, time 216.669 samples/sec
+epoch 5, total_step 80460, total loss is 25.15 , inference loss is 16.28, weight deacy loss is 8.88, training accuracy is 0.281250, time 211.775 samples/sec
+epoch 5, total_step 80480, total loss is 21.03 , inference loss is 12.16, weight deacy loss is 8.87, training accuracy is 0.250000, time 213.114 samples/sec
+epoch 5, total_step 80500, total loss is 18.26 , inference loss is 9.39, weight deacy loss is 8.87, training accuracy is 0.187500, time 213.891 samples/sec
+epoch 5, total_step 80520, total loss is 18.78 , inference loss is 9.91, weight deacy loss is 8.87, training accuracy is 0.250000, time 214.062 samples/sec
+epoch 5, total_step 80540, total loss is 17.31 , inference loss is 8.44, weight deacy loss is 8.87, training accuracy is 0.281250, time 211.650 samples/sec
+epoch 5, total_step 80560, total loss is 16.08 , inference loss is 7.21, weight deacy loss is 8.87, training accuracy is 0.281250, time 221.547 samples/sec
+epoch 5, total_step 80580, total loss is 20.21 , inference loss is 11.34, weight deacy loss is 8.87, training accuracy is 0.281250, time 294.950 samples/sec
+epoch 5, total_step 80600, total loss is 27.27 , inference loss is 18.40, weight deacy loss is 8.87, training accuracy is 0.187500, time 299.130 samples/sec
+epoch 5, total_step 80620, total loss is 23.28 , inference loss is 14.41, weight deacy loss is 8.87, training accuracy is 0.218750, time 211.254 samples/sec
+epoch 5, total_step 80640, total loss is 22.40 , inference loss is 13.53, weight deacy loss is 8.87, training accuracy is 0.250000, time 202.042 samples/sec
+epoch 5, total_step 80660, total loss is 24.40 , inference loss is 15.53, weight deacy loss is 8.87, training accuracy is 0.218750, time 215.998 samples/sec
+epoch 5, total_step 80680, total loss is 18.03 , inference loss is 9.17, weight deacy loss is 8.87, training accuracy is 0.218750, time 207.931 samples/sec
+epoch 5, total_step 80700, total loss is 17.97 , inference loss is 9.11, weight deacy loss is 8.87, training accuracy is 0.468750, time 14.030 samples/sec
+epoch 5, total_step 80720, total loss is 23.80 , inference loss is 14.94, weight deacy loss is 8.86, training accuracy is 0.281250, time 216.850 samples/sec
+epoch 5, total_step 80740, total loss is 17.83 , inference loss is 8.97, weight deacy loss is 8.86, training accuracy is 0.375000, time 211.298 samples/sec
+epoch 5, total_step 80760, total loss is 20.42 , inference loss is 11.56, weight deacy loss is 8.86, training accuracy is 0.312500, time 201.049 samples/sec
+epoch 5, total_step 80780, total loss is 25.90 , inference loss is 17.04, weight deacy loss is 8.86, training accuracy is 0.218750, time 215.706 samples/sec
+epoch 5, total_step 80800, total loss is 20.40 , inference loss is 11.54, weight deacy loss is 8.86, training accuracy is 0.250000, time 217.935 samples/sec
+epoch 5, total_step 80820, total loss is 21.39 , inference loss is 12.53, weight deacy loss is 8.86, training accuracy is 0.312500, time 222.565 samples/sec
+epoch 5, total_step 80840, total loss is 20.75 , inference loss is 11.89, weight deacy loss is 8.86, training accuracy is 0.281250, time 213.915 samples/sec
+epoch 5, total_step 80860, total loss is 19.35 , inference loss is 10.50, weight deacy loss is 8.86, training accuracy is 0.312500, time 209.363 samples/sec
+epoch 5, total_step 80880, total loss is 20.48 , inference loss is 11.62, weight deacy loss is 8.86, training accuracy is 0.343750, time 176.519 samples/sec
+epoch 5, total_step 80900, total loss is 18.95 , inference loss is 10.09, weight deacy loss is 8.86, training accuracy is 0.375000, time 207.119 samples/sec
+epoch 5, total_step 80920, total loss is 23.57 , inference loss is 14.72, weight deacy loss is 8.86, training accuracy is 0.218750, time 224.312 samples/sec
+epoch 5, total_step 80940, total loss is 18.67 , inference loss is 9.82, weight deacy loss is 8.85, training accuracy is 0.312500, time 210.436 samples/sec
+epoch 5, total_step 80960, total loss is 23.19 , inference loss is 14.34, weight deacy loss is 8.85, training accuracy is 0.312500, time 207.740 samples/sec
+epoch 5, total_step 80980, total loss is 19.36 , inference loss is 10.51, weight deacy loss is 8.85, training accuracy is 0.250000, time 210.738 samples/sec
+epoch 5, total_step 81000, total loss is 18.07 , inference loss is 9.22, weight deacy loss is 8.85, training accuracy is 0.281250, time 13.484 samples/sec
+epoch 5, total_step 81020, total loss is 22.76 , inference loss is 13.91, weight deacy loss is 8.85, training accuracy is 0.281250, time 201.951 samples/sec
+epoch 5, total_step 81040, total loss is 20.32 , inference loss is 11.47, weight deacy loss is 8.85, training accuracy is 0.312500, time 210.216 samples/sec
+epoch 5, total_step 81060, total loss is 20.61 , inference loss is 11.76, weight deacy loss is 8.85, training accuracy is 0.250000, time 208.149 samples/sec
+epoch 5, total_step 81080, total loss is 19.44 , inference loss is 10.60, weight deacy loss is 8.85, training accuracy is 0.250000, time 207.750 samples/sec
+epoch 5, total_step 81100, total loss is 18.54 , inference loss is 9.69, weight deacy loss is 8.85, training accuracy is 0.250000, time 212.590 samples/sec
+epoch 5, total_step 81120, total loss is 15.98 , inference loss is 7.13, weight deacy loss is 8.85, training accuracy is 0.343750, time 181.751 samples/sec
+epoch 5, total_step 81140, total loss is 18.11 , inference loss is 9.26, weight deacy loss is 8.85, training accuracy is 0.312500, time 224.637 samples/sec
+epoch 5, total_step 81160, total loss is 19.94 , inference loss is 11.09, weight deacy loss is 8.84, training accuracy is 0.343750, time 218.403 samples/sec
+epoch 5, total_step 81180, total loss is 18.82 , inference loss is 9.98, weight deacy loss is 8.84, training accuracy is 0.250000, time 215.040 samples/sec
+epoch 5, total_step 81200, total loss is 19.05 , inference loss is 10.21, weight deacy loss is 8.84, training accuracy is 0.312500, time 202.296 samples/sec
+epoch 5, total_step 81220, total loss is 18.89 , inference loss is 10.05, weight deacy loss is 8.84, training accuracy is 0.312500, time 203.842 samples/sec
+epoch 5, total_step 81240, total loss is 24.11 , inference loss is 15.27, weight deacy loss is 8.84, training accuracy is 0.312500, time 211.675 samples/sec
+epoch 5, total_step 81260, total loss is 23.99 , inference loss is 15.15, weight deacy loss is 8.84, training accuracy is 0.312500, time 208.177 samples/sec
+epoch 5, total_step 81280, total loss is 20.48 , inference loss is 11.64, weight deacy loss is 8.84, training accuracy is 0.312500, time 216.807 samples/sec
+epoch 5, total_step 81300, total loss is 22.05 , inference loss is 13.21, weight deacy loss is 8.84, training accuracy is 0.156250, time 13.759 samples/sec
+epoch 5, total_step 81320, total loss is 18.77 , inference loss is 9.94, weight deacy loss is 8.84, training accuracy is 0.312500, time 192.606 samples/sec
+epoch 5, total_step 81340, total loss is 22.72 , inference loss is 13.89, weight deacy loss is 8.84, training accuracy is 0.156250, time 208.557 samples/sec
+epoch 5, total_step 81360, total loss is 21.54 , inference loss is 12.70, weight deacy loss is 8.84, training accuracy is 0.312500, time 218.014 samples/sec
+epoch 5, total_step 81380, total loss is 21.24 , inference loss is 12.41, weight deacy loss is 8.83, training accuracy is 0.250000, time 209.867 samples/sec
+epoch 5, total_step 81400, total loss is 19.76 , inference loss is 10.93, weight deacy loss is 8.83, training accuracy is 0.406250, time 300.873 samples/sec
+epoch 5, total_step 81420, total loss is 20.38 , inference loss is 11.55, weight deacy loss is 8.83, training accuracy is 0.218750, time 204.901 samples/sec
+epoch 5, total_step 81440, total loss is 18.23 , inference loss is 9.39, weight deacy loss is 8.83, training accuracy is 0.281250, time 204.861 samples/sec
+epoch 5, total_step 81460, total loss is 19.20 , inference loss is 10.37, weight deacy loss is 8.83, training accuracy is 0.250000, time 202.560 samples/sec
+epoch 5, total_step 81480, total loss is 19.03 , inference loss is 10.20, weight deacy loss is 8.83, training accuracy is 0.218750, time 207.918 samples/sec
+epoch 5, total_step 81500, total loss is 18.52 , inference loss is 9.69, weight deacy loss is 8.83, training accuracy is 0.406250, time 214.329 samples/sec
+epoch 5, total_step 81520, total loss is 18.85 , inference loss is 10.03, weight deacy loss is 8.83, training accuracy is 0.343750, time 217.137 samples/sec
+epoch 5, total_step 81540, total loss is 18.82 , inference loss is 9.99, weight deacy loss is 8.83, training accuracy is 0.187500, time 200.228 samples/sec
+epoch 5, total_step 81560, total loss is 18.87 , inference loss is 10.04, weight deacy loss is 8.83, training accuracy is 0.312500, time 210.122 samples/sec
+epoch 5, total_step 81580, total loss is 16.62 , inference loss is 7.79, weight deacy loss is 8.82, training accuracy is 0.312500, time 208.448 samples/sec
+epoch 5, total_step 81600, total loss is 21.17 , inference loss is 12.34, weight deacy loss is 8.82, training accuracy is 0.312500, time 13.910 samples/sec
+epoch 5, total_step 81620, total loss is 16.49 , inference loss is 7.67, weight deacy loss is 8.82, training accuracy is 0.343750, time 214.124 samples/sec
+epoch 5, total_step 81640, total loss is 15.91 , inference loss is 7.09, weight deacy loss is 8.82, training accuracy is 0.312500, time 224.057 samples/sec
+epoch 5, total_step 81660, total loss is 19.22 , inference loss is 10.40, weight deacy loss is 8.82, training accuracy is 0.218750, time 295.532 samples/sec
+epoch 5, total_step 81680, total loss is 21.13 , inference loss is 12.31, weight deacy loss is 8.82, training accuracy is 0.312500, time 205.795 samples/sec
+epoch 5, total_step 81700, total loss is 18.80 , inference loss is 9.98, weight deacy loss is 8.82, training accuracy is 0.375000, time 215.518 samples/sec
+epoch 5, total_step 81720, total loss is 16.97 , inference loss is 8.16, weight deacy loss is 8.82, training accuracy is 0.312500, time 206.977 samples/sec
+epoch 5, total_step 81740, total loss is 20.39 , inference loss is 11.57, weight deacy loss is 8.82, training accuracy is 0.406250, time 215.876 samples/sec
+epoch 5, total_step 81760, total loss is 19.87 , inference loss is 11.06, weight deacy loss is 8.82, training accuracy is 0.218750, time 210.767 samples/sec
+epoch 5, total_step 81780, total loss is 19.46 , inference loss is 10.64, weight deacy loss is 8.82, training accuracy is 0.406250, time 208.910 samples/sec
+epoch 5, total_step 81800, total loss is 19.24 , inference loss is 10.43, weight deacy loss is 8.81, training accuracy is 0.312500, time 209.853 samples/sec
+epoch 5, total_step 81820, total loss is 18.94 , inference loss is 10.13, weight deacy loss is 8.81, training accuracy is 0.312500, time 214.488 samples/sec
+epoch 5, total_step 81840, total loss is 18.98 , inference loss is 10.17, weight deacy loss is 8.81, training accuracy is 0.312500, time 213.210 samples/sec
+epoch 5, total_step 81860, total loss is 17.28 , inference loss is 8.47, weight deacy loss is 8.81, training accuracy is 0.500000, time 216.295 samples/sec
+epoch 5, total_step 81880, total loss is 16.43 , inference loss is 7.62, weight deacy loss is 8.81, training accuracy is 0.281250, time 212.300 samples/sec
+epoch 5, total_step 81900, total loss is 22.75 , inference loss is 13.94, weight deacy loss is 8.81, training accuracy is 0.125000, time 13.991 samples/sec
+epoch 5, total_step 81920, total loss is 15.64 , inference loss is 6.83, weight deacy loss is 8.81, training accuracy is 0.625000, time 211.787 samples/sec
+epoch 5, total_step 81940, total loss is 20.81 , inference loss is 12.00, weight deacy loss is 8.81, training accuracy is 0.187500, time 298.805 samples/sec
+epoch 5, total_step 81960, total loss is 20.89 , inference loss is 12.08, weight deacy loss is 8.81, training accuracy is 0.218750, time 207.562 samples/sec
+epoch 5, total_step 81980, total loss is 21.41 , inference loss is 12.60, weight deacy loss is 8.81, training accuracy is 0.375000, time 219.159 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.742426000000005
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 116 0.9842592592592593
+best_threshold_index 116 0.9855555555555555
+best_threshold_index 116 0.9861111111111112
+best_threshold_index 116 0.9851851851851852
+best_threshold_index 116 0.9838888888888889
+best_threshold_index 116 0.9842592592592593
+best_threshold_index 116 0.985
+best_threshold_index 116 0.9838888888888889
+best_threshold_index 116 0.9842592592592593
+[lfw][82000]XNorm: 21.771844
+[lfw][82000]Accuracy-Flip: 0.98483+-0.00713
+test accuracy is: 0.9848333333333334
+epoch 5, total_step 82000, total loss is 21.84 , inference loss is 13.03, weight deacy loss is 8.81, training accuracy is 0.187500, time 19.270 samples/sec
+epoch 5, total_step 82020, total loss is 17.59 , inference loss is 8.79, weight deacy loss is 8.81, training accuracy is 0.312500, time 215.240 samples/sec
+epoch 5, total_step 82040, total loss is 22.18 , inference loss is 13.37, weight deacy loss is 8.80, training accuracy is 0.281250, time 295.642 samples/sec
+epoch 5, total_step 82060, total loss is 23.56 , inference loss is 14.75, weight deacy loss is 8.80, training accuracy is 0.281250, time 208.128 samples/sec
+epoch 5, total_step 82080, total loss is 16.33 , inference loss is 7.52, weight deacy loss is 8.80, training accuracy is 0.437500, time 215.772 samples/sec
+epoch 5, total_step 82100, total loss is 18.04 , inference loss is 9.24, weight deacy loss is 8.80, training accuracy is 0.468750, time 201.341 samples/sec
+epoch 5, total_step 82120, total loss is 16.75 , inference loss is 7.95, weight deacy loss is 8.80, training accuracy is 0.343750, time 212.477 samples/sec
+epoch 5, total_step 82140, total loss is 21.10 , inference loss is 12.30, weight deacy loss is 8.80, training accuracy is 0.312500, time 209.292 samples/sec
+epoch 5, total_step 82160, total loss is 16.62 , inference loss is 7.82, weight deacy loss is 8.80, training accuracy is 0.375000, time 217.320 samples/sec
+epoch 5, total_step 82180, total loss is 20.08 , inference loss is 11.28, weight deacy loss is 8.80, training accuracy is 0.218750, time 218.485 samples/sec
+epoch 5, total_step 82200, total loss is 20.66 , inference loss is 11.86, weight deacy loss is 8.80, training accuracy is 0.343750, time 13.832 samples/sec
+epoch 5, total_step 82220, total loss is 21.22 , inference loss is 12.42, weight deacy loss is 8.80, training accuracy is 0.343750, time 216.678 samples/sec
+epoch 5, total_step 82240, total loss is 17.51 , inference loss is 8.71, weight deacy loss is 8.80, training accuracy is 0.375000, time 211.198 samples/sec
+epoch 5, total_step 82260, total loss is 20.78 , inference loss is 11.99, weight deacy loss is 8.79, training accuracy is 0.187500, time 213.023 samples/sec
+epoch 5, total_step 82280, total loss is 16.42 , inference loss is 7.62, weight deacy loss is 8.79, training accuracy is 0.437500, time 213.894 samples/sec
+epoch 5, total_step 82300, total loss is 16.49 , inference loss is 7.70, weight deacy loss is 8.79, training accuracy is 0.437500, time 215.974 samples/sec
+epoch 5, total_step 82320, total loss is 20.25 , inference loss is 11.46, weight deacy loss is 8.79, training accuracy is 0.250000, time 222.738 samples/sec
+epoch 5, total_step 82340, total loss is 19.07 , inference loss is 10.28, weight deacy loss is 8.79, training accuracy is 0.343750, time 215.844 samples/sec
+epoch 5, total_step 82360, total loss is 17.55 , inference loss is 8.76, weight deacy loss is 8.79, training accuracy is 0.312500, time 299.463 samples/sec
+epoch 5, total_step 82380, total loss is 16.70 , inference loss is 7.91, weight deacy loss is 8.79, training accuracy is 0.406250, time 297.281 samples/sec
+epoch 5, total_step 82400, total loss is 19.28 , inference loss is 10.50, weight deacy loss is 8.79, training accuracy is 0.343750, time 208.298 samples/sec
+epoch 5, total_step 82420, total loss is 18.98 , inference loss is 10.19, weight deacy loss is 8.79, training accuracy is 0.406250, time 229.765 samples/sec
+epoch 5, total_step 82440, total loss is 16.32 , inference loss is 7.54, weight deacy loss is 8.79, training accuracy is 0.281250, time 212.120 samples/sec
+epoch 5, total_step 82460, total loss is 16.50 , inference loss is 7.72, weight deacy loss is 8.79, training accuracy is 0.343750, time 208.025 samples/sec
+epoch 5, total_step 82480, total loss is 19.57 , inference loss is 10.78, weight deacy loss is 8.78, training accuracy is 0.312500, time 217.557 samples/sec
+epoch 5, total_step 82500, total loss is 19.07 , inference loss is 10.28, weight deacy loss is 8.78, training accuracy is 0.218750, time 13.854 samples/sec
+epoch 5, total_step 82520, total loss is 16.94 , inference loss is 8.16, weight deacy loss is 8.78, training accuracy is 0.312500, time 211.032 samples/sec
+epoch 5, total_step 82540, total loss is 25.92 , inference loss is 17.14, weight deacy loss is 8.78, training accuracy is 0.250000, time 212.961 samples/sec
+epoch 5, total_step 82560, total loss is 20.11 , inference loss is 11.33, weight deacy loss is 8.78, training accuracy is 0.281250, time 248.598 samples/sec
+epoch 5, total_step 82580, total loss is 20.63 , inference loss is 11.85, weight deacy loss is 8.78, training accuracy is 0.250000, time 211.777 samples/sec
+epoch 5, total_step 82600, total loss is 20.29 , inference loss is 11.51, weight deacy loss is 8.78, training accuracy is 0.281250, time 206.702 samples/sec
+epoch 5, total_step 82620, total loss is 17.73 , inference loss is 8.95, weight deacy loss is 8.78, training accuracy is 0.437500, time 220.812 samples/sec
+epoch 5, total_step 82640, total loss is 21.49 , inference loss is 12.71, weight deacy loss is 8.78, training accuracy is 0.281250, time 210.324 samples/sec
+epoch 5, total_step 82660, total loss is 23.62 , inference loss is 14.84, weight deacy loss is 8.78, training accuracy is 0.281250, time 216.253 samples/sec
+epoch 5, total_step 82680, total loss is 22.83 , inference loss is 14.05, weight deacy loss is 8.78, training accuracy is 0.281250, time 216.351 samples/sec
+epoch 5, total_step 82700, total loss is 22.29 , inference loss is 13.52, weight deacy loss is 8.77, training accuracy is 0.250000, time 213.399 samples/sec
+epoch 5, total_step 82720, total loss is 17.10 , inference loss is 8.33, weight deacy loss is 8.77, training accuracy is 0.500000, time 219.110 samples/sec
+epoch 5, total_step 82740, total loss is 21.44 , inference loss is 12.66, weight deacy loss is 8.77, training accuracy is 0.187500, time 214.176 samples/sec
+epoch 5, total_step 82760, total loss is 21.32 , inference loss is 12.55, weight deacy loss is 8.77, training accuracy is 0.218750, time 296.582 samples/sec
+epoch 5, total_step 82780, total loss is 21.39 , inference loss is 12.62, weight deacy loss is 8.77, training accuracy is 0.187500, time 213.561 samples/sec
+epoch 5, total_step 82800, total loss is 17.65 , inference loss is 8.88, weight deacy loss is 8.77, training accuracy is 0.343750, time 14.223 samples/sec
+epoch 5, total_step 82820, total loss is 18.07 , inference loss is 9.30, weight deacy loss is 8.77, training accuracy is 0.343750, time 210.607 samples/sec
+epoch 5, total_step 82840, total loss is 19.84 , inference loss is 11.08, weight deacy loss is 8.77, training accuracy is 0.250000, time 212.710 samples/sec
+epoch 5, total_step 82860, total loss is 19.00 , inference loss is 10.23, weight deacy loss is 8.77, training accuracy is 0.375000, time 212.366 samples/sec
+epoch 5, total_step 82880, total loss is 20.77 , inference loss is 12.00, weight deacy loss is 8.77, training accuracy is 0.187500, time 217.094 samples/sec
+epoch 5, total_step 82900, total loss is 15.15 , inference loss is 6.38, weight deacy loss is 8.77, training accuracy is 0.468750, time 218.731 samples/sec
+epoch 5, total_step 82920, total loss is 18.93 , inference loss is 10.16, weight deacy loss is 8.76, training accuracy is 0.343750, time 209.961 samples/sec
+epoch 5, total_step 82940, total loss is 18.09 , inference loss is 9.33, weight deacy loss is 8.76, training accuracy is 0.406250, time 214.898 samples/sec
+epoch 5, total_step 82960, total loss is 21.18 , inference loss is 12.41, weight deacy loss is 8.76, training accuracy is 0.312500, time 210.875 samples/sec
+epoch 5, total_step 82980, total loss is 16.87 , inference loss is 8.11, weight deacy loss is 8.76, training accuracy is 0.187500, time 239.334 samples/sec
+epoch 5, total_step 83000, total loss is 21.62 , inference loss is 12.86, weight deacy loss is 8.76, training accuracy is 0.343750, time 210.395 samples/sec
+epoch 5, total_step 83020, total loss is 18.35 , inference loss is 9.59, weight deacy loss is 8.76, training accuracy is 0.375000, time 277.698 samples/sec
+epoch 5, total_step 83040, total loss is 18.03 , inference loss is 9.27, weight deacy loss is 8.76, training accuracy is 0.281250, time 216.844 samples/sec
+epoch 5, total_step 83060, total loss is 17.93 , inference loss is 9.17, weight deacy loss is 8.76, training accuracy is 0.281250, time 209.725 samples/sec
+epoch 5, total_step 83080, total loss is 18.25 , inference loss is 9.50, weight deacy loss is 8.76, training accuracy is 0.218750, time 211.634 samples/sec
+epoch 5, total_step 83100, total loss is 15.61 , inference loss is 6.85, weight deacy loss is 8.76, training accuracy is 0.375000, time 11.551 samples/sec
+epoch 5, total_step 83120, total loss is 20.29 , inference loss is 11.54, weight deacy loss is 8.76, training accuracy is 0.218750, time 208.676 samples/sec
+epoch 5, total_step 83140, total loss is 18.44 , inference loss is 9.69, weight deacy loss is 8.75, training accuracy is 0.250000, time 211.125 samples/sec
+epoch 5, total_step 83160, total loss is 16.28 , inference loss is 7.53, weight deacy loss is 8.75, training accuracy is 0.437500, time 198.725 samples/sec
+epoch 5, total_step 83180, total loss is 20.47 , inference loss is 11.71, weight deacy loss is 8.75, training accuracy is 0.343750, time 224.325 samples/sec
+epoch 5, total_step 83200, total loss is 18.91 , inference loss is 10.16, weight deacy loss is 8.75, training accuracy is 0.187500, time 214.469 samples/sec
+epoch 5, total_step 83220, total loss is 19.40 , inference loss is 10.65, weight deacy loss is 8.75, training accuracy is 0.437500, time 195.944 samples/sec
+epoch 5, total_step 83240, total loss is 20.41 , inference loss is 11.66, weight deacy loss is 8.75, training accuracy is 0.218750, time 210.994 samples/sec
+epoch 5, total_step 83260, total loss is 22.80 , inference loss is 14.05, weight deacy loss is 8.75, training accuracy is 0.156250, time 203.989 samples/sec
+epoch 5, total_step 83280, total loss is 20.42 , inference loss is 11.67, weight deacy loss is 8.75, training accuracy is 0.250000, time 296.088 samples/sec
+epoch 5, total_step 83300, total loss is 19.25 , inference loss is 10.50, weight deacy loss is 8.75, training accuracy is 0.281250, time 212.781 samples/sec
+epoch 5, total_step 83320, total loss is 16.08 , inference loss is 7.33, weight deacy loss is 8.75, training accuracy is 0.375000, time 209.226 samples/sec
+epoch 5, total_step 83340, total loss is 17.99 , inference loss is 9.24, weight deacy loss is 8.75, training accuracy is 0.312500, time 217.996 samples/sec
+epoch 5, total_step 83360, total loss is 21.86 , inference loss is 13.11, weight deacy loss is 8.74, training accuracy is 0.312500, time 211.956 samples/sec
+epoch 5, total_step 83380, total loss is 24.90 , inference loss is 16.15, weight deacy loss is 8.74, training accuracy is 0.375000, time 207.782 samples/sec
+epoch 5, total_step 83400, total loss is 18.45 , inference loss is 9.70, weight deacy loss is 8.74, training accuracy is 0.218750, time 13.993 samples/sec
+epoch 5, total_step 83420, total loss is 17.60 , inference loss is 8.86, weight deacy loss is 8.74, training accuracy is 0.250000, time 218.538 samples/sec
+epoch 5, total_step 83440, total loss is 17.24 , inference loss is 8.50, weight deacy loss is 8.74, training accuracy is 0.375000, time 225.762 samples/sec
+epoch 5, total_step 83460, total loss is 21.37 , inference loss is 12.64, weight deacy loss is 8.74, training accuracy is 0.218750, time 208.249 samples/sec
+epoch 5, total_step 83480, total loss is 17.04 , inference loss is 8.30, weight deacy loss is 8.74, training accuracy is 0.312500, time 211.708 samples/sec
+epoch 5, total_step 83500, total loss is 18.46 , inference loss is 9.72, weight deacy loss is 8.74, training accuracy is 0.250000, time 210.852 samples/sec
+epoch 5, total_step 83520, total loss is 21.19 , inference loss is 12.45, weight deacy loss is 8.74, training accuracy is 0.250000, time 203.215 samples/sec
+epoch 5, total_step 83540, total loss is 24.36 , inference loss is 15.62, weight deacy loss is 8.74, training accuracy is 0.218750, time 210.818 samples/sec
+epoch 5, total_step 83560, total loss is 17.12 , inference loss is 8.38, weight deacy loss is 8.74, training accuracy is 0.437500, time 296.763 samples/sec
+epoch 5, total_step 83580, total loss is 22.01 , inference loss is 13.28, weight deacy loss is 8.73, training accuracy is 0.125000, time 213.729 samples/sec
+epoch 5, total_step 83600, total loss is 25.42 , inference loss is 16.68, weight deacy loss is 8.73, training accuracy is 0.250000, time 218.863 samples/sec
+epoch 5, total_step 83620, total loss is 18.72 , inference loss is 9.99, weight deacy loss is 8.73, training accuracy is 0.312500, time 214.393 samples/sec
+epoch 5, total_step 83640, total loss is 17.66 , inference loss is 8.92, weight deacy loss is 8.73, training accuracy is 0.218750, time 215.456 samples/sec
+epoch 5, total_step 83660, total loss is 15.87 , inference loss is 7.14, weight deacy loss is 8.73, training accuracy is 0.375000, time 230.801 samples/sec
+epoch 5, total_step 83680, total loss is 17.96 , inference loss is 9.23, weight deacy loss is 8.73, training accuracy is 0.375000, time 218.678 samples/sec
+epoch 5, total_step 83700, total loss is 18.39 , inference loss is 9.66, weight deacy loss is 8.73, training accuracy is 0.406250, time 13.980 samples/sec
+epoch 5, total_step 83720, total loss is 21.90 , inference loss is 13.17, weight deacy loss is 8.73, training accuracy is 0.375000, time 219.955 samples/sec
+epoch 5, total_step 83740, total loss is 18.24 , inference loss is 9.51, weight deacy loss is 8.73, training accuracy is 0.250000, time 212.758 samples/sec
+epoch 5, total_step 83760, total loss is 22.22 , inference loss is 13.49, weight deacy loss is 8.73, training accuracy is 0.187500, time 206.834 samples/sec
+epoch 5, total_step 83780, total loss is 21.07 , inference loss is 12.35, weight deacy loss is 8.73, training accuracy is 0.250000, time 204.059 samples/sec
+epoch 5, total_step 83800, total loss is 19.55 , inference loss is 10.83, weight deacy loss is 8.72, training accuracy is 0.343750, time 227.862 samples/sec
+epoch 5, total_step 83820, total loss is 19.16 , inference loss is 10.43, weight deacy loss is 8.72, training accuracy is 0.343750, time 299.728 samples/sec
+epoch 5, total_step 83840, total loss is 20.04 , inference loss is 11.32, weight deacy loss is 8.72, training accuracy is 0.281250, time 212.397 samples/sec
+epoch 5, total_step 83860, total loss is 17.15 , inference loss is 8.43, weight deacy loss is 8.72, training accuracy is 0.218750, time 206.464 samples/sec
+epoch 5, total_step 83880, total loss is 22.16 , inference loss is 13.44, weight deacy loss is 8.72, training accuracy is 0.250000, time 212.974 samples/sec
+epoch 5, total_step 83900, total loss is 16.29 , inference loss is 7.57, weight deacy loss is 8.72, training accuracy is 0.343750, time 209.790 samples/sec
+epoch 5, total_step 83920, total loss is 19.14 , inference loss is 10.42, weight deacy loss is 8.72, training accuracy is 0.218750, time 206.789 samples/sec
+epoch 5, total_step 83940, total loss is 20.42 , inference loss is 11.70, weight deacy loss is 8.72, training accuracy is 0.312500, time 211.612 samples/sec
+epoch 5, total_step 83960, total loss is 20.14 , inference loss is 11.42, weight deacy loss is 8.72, training accuracy is 0.312500, time 214.043 samples/sec
+epoch 5, total_step 83980, total loss is 23.93 , inference loss is 15.21, weight deacy loss is 8.72, training accuracy is 0.343750, time 213.772 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.650451999999994
+best_threshold_index 113 0.9866666666666667
+best_threshold_index 119 0.9861111111111112
+best_threshold_index 112 0.9857407407407407
+best_threshold_index 114 0.9866666666666667
+best_threshold_index 113 0.9864814814814815
+best_threshold_index 119 0.9853703703703703
+best_threshold_index 113 0.9859259259259259
+best_threshold_index 119 0.9857407407407407
+best_threshold_index 119 0.9848148148148148
+best_threshold_index 119 0.9855555555555555
+[lfw][84000]XNorm: 21.618022
+[lfw][84000]Accuracy-Flip: 0.98417+-0.00569
+test accuracy is: 0.9841666666666666
+epoch 5, total_step 84000, total loss is 19.01 , inference loss is 10.29, weight deacy loss is 8.72, training accuracy is 0.343750, time 14.983 samples/sec
+epoch 5, total_step 84020, total loss is 18.04 , inference loss is 9.33, weight deacy loss is 8.71, training accuracy is 0.281250, time 209.515 samples/sec
+epoch 5, total_step 84040, total loss is 18.22 , inference loss is 9.50, weight deacy loss is 8.71, training accuracy is 0.343750, time 295.868 samples/sec
+epoch 5, total_step 84060, total loss is 19.14 , inference loss is 10.43, weight deacy loss is 8.71, training accuracy is 0.343750, time 220.752 samples/sec
+epoch 5, total_step 84080, total loss is 20.84 , inference loss is 12.13, weight deacy loss is 8.71, training accuracy is 0.343750, time 212.823 samples/sec
+epoch 5, total_step 84100, total loss is 18.72 , inference loss is 10.01, weight deacy loss is 8.71, training accuracy is 0.468750, time 215.131 samples/sec
+epoch 5, total_step 84120, total loss is 20.55 , inference loss is 11.84, weight deacy loss is 8.71, training accuracy is 0.312500, time 218.392 samples/sec
+epoch 5, total_step 84140, total loss is 20.74 , inference loss is 12.03, weight deacy loss is 8.71, training accuracy is 0.281250, time 215.447 samples/sec
+epoch 5, total_step 84160, total loss is 21.72 , inference loss is 13.02, weight deacy loss is 8.71, training accuracy is 0.250000, time 212.984 samples/sec
+epoch 5, total_step 84180, total loss is 20.18 , inference loss is 11.48, weight deacy loss is 8.71, training accuracy is 0.312500, time 227.765 samples/sec
+epoch 5, total_step 84200, total loss is 19.30 , inference loss is 10.59, weight deacy loss is 8.71, training accuracy is 0.218750, time 295.763 samples/sec
+epoch 5, total_step 84220, total loss is 18.34 , inference loss is 9.64, weight deacy loss is 8.71, training accuracy is 0.281250, time 299.951 samples/sec
+epoch 5, total_step 84240, total loss is 22.31 , inference loss is 13.61, weight deacy loss is 8.70, training accuracy is 0.437500, time 220.394 samples/sec
+epoch 5, total_step 84260, total loss is 23.19 , inference loss is 14.49, weight deacy loss is 8.70, training accuracy is 0.187500, time 207.829 samples/sec
+epoch 5, total_step 84280, total loss is 18.22 , inference loss is 9.52, weight deacy loss is 8.70, training accuracy is 0.281250, time 206.687 samples/sec
+epoch 5, total_step 84300, total loss is 20.17 , inference loss is 11.46, weight deacy loss is 8.70, training accuracy is 0.250000, time 12.895 samples/sec
+epoch 5, total_step 84320, total loss is 18.36 , inference loss is 9.66, weight deacy loss is 8.70, training accuracy is 0.343750, time 219.307 samples/sec
+epoch 5, total_step 84340, total loss is 19.60 , inference loss is 10.90, weight deacy loss is 8.70, training accuracy is 0.468750, time 209.577 samples/sec
+epoch 5, total_step 84360, total loss is 20.25 , inference loss is 11.56, weight deacy loss is 8.70, training accuracy is 0.281250, time 213.394 samples/sec
+epoch 5, total_step 84380, total loss is 15.45 , inference loss is 6.75, weight deacy loss is 8.70, training accuracy is 0.468750, time 171.245 samples/sec
+epoch 5, total_step 84400, total loss is 17.05 , inference loss is 8.35, weight deacy loss is 8.70, training accuracy is 0.343750, time 220.039 samples/sec
+epoch 5, total_step 84420, total loss is 18.78 , inference loss is 10.08, weight deacy loss is 8.70, training accuracy is 0.218750, time 248.584 samples/sec
+epoch 5, total_step 84440, total loss is 17.04 , inference loss is 8.34, weight deacy loss is 8.70, training accuracy is 0.531250, time 209.398 samples/sec
+epoch 5, total_step 84460, total loss is 19.10 , inference loss is 10.41, weight deacy loss is 8.69, training accuracy is 0.250000, time 214.297 samples/sec
+epoch 5, total_step 84480, total loss is 18.04 , inference loss is 9.35, weight deacy loss is 8.69, training accuracy is 0.312500, time 198.238 samples/sec
+epoch 5, total_step 84500, total loss is 21.30 , inference loss is 12.61, weight deacy loss is 8.69, training accuracy is 0.218750, time 207.626 samples/sec
+epoch 5, total_step 84520, total loss is 20.90 , inference loss is 12.21, weight deacy loss is 8.69, training accuracy is 0.343750, time 209.993 samples/sec
+epoch 5, total_step 84540, total loss is 17.86 , inference loss is 9.17, weight deacy loss is 8.69, training accuracy is 0.406250, time 202.316 samples/sec
+epoch 5, total_step 84560, total loss is 19.63 , inference loss is 10.94, weight deacy loss is 8.69, training accuracy is 0.187500, time 214.424 samples/sec
+epoch 5, total_step 84580, total loss is 16.10 , inference loss is 7.41, weight deacy loss is 8.69, training accuracy is 0.343750, time 217.568 samples/sec
+epoch 5, total_step 84600, total loss is 19.91 , inference loss is 11.22, weight deacy loss is 8.69, training accuracy is 0.312500, time 13.214 samples/sec
+epoch 5, total_step 84620, total loss is 22.31 , inference loss is 13.62, weight deacy loss is 8.69, training accuracy is 0.343750, time 205.696 samples/sec
+epoch 5, total_step 84640, total loss is 18.43 , inference loss is 9.74, weight deacy loss is 8.69, training accuracy is 0.250000, time 210.370 samples/sec
+epoch 5, total_step 84660, total loss is 22.57 , inference loss is 13.88, weight deacy loss is 8.68, training accuracy is 0.281250, time 297.560 samples/sec
+epoch 5, total_step 84680, total loss is 14.56 , inference loss is 5.88, weight deacy loss is 8.68, training accuracy is 0.375000, time 211.552 samples/sec
+epoch 5, total_step 84700, total loss is 21.15 , inference loss is 12.46, weight deacy loss is 8.68, training accuracy is 0.312500, time 202.684 samples/sec
+epoch 5, total_step 84720, total loss is 19.92 , inference loss is 11.23, weight deacy loss is 8.68, training accuracy is 0.187500, time 206.999 samples/sec
+epoch 5, total_step 84740, total loss is 17.28 , inference loss is 8.59, weight deacy loss is 8.68, training accuracy is 0.437500, time 204.503 samples/sec
+epoch 5, total_step 84760, total loss is 18.88 , inference loss is 10.20, weight deacy loss is 8.68, training accuracy is 0.406250, time 202.733 samples/sec
+epoch 5, total_step 84780, total loss is 16.09 , inference loss is 7.41, weight deacy loss is 8.68, training accuracy is 0.406250, time 213.026 samples/sec
+epoch 5, total_step 84800, total loss is 17.01 , inference loss is 8.34, weight deacy loss is 8.68, training accuracy is 0.468750, time 212.118 samples/sec
+epoch 5, total_step 84820, total loss is 20.00 , inference loss is 11.33, weight deacy loss is 8.68, training accuracy is 0.187500, time 215.849 samples/sec
+epoch 5, total_step 84840, total loss is 17.68 , inference loss is 9.00, weight deacy loss is 8.68, training accuracy is 0.312500, time 209.637 samples/sec
+epoch 5, total_step 84860, total loss is 15.69 , inference loss is 7.02, weight deacy loss is 8.68, training accuracy is 0.437500, time 239.919 samples/sec
+epoch 5, total_step 84880, total loss is 19.16 , inference loss is 10.48, weight deacy loss is 8.67, training accuracy is 0.312500, time 213.917 samples/sec
+epoch 5, total_step 84900, total loss is 15.62 , inference loss is 6.95, weight deacy loss is 8.67, training accuracy is 0.218750, time 11.952 samples/sec
+epoch 5, total_step 84920, total loss is 20.72 , inference loss is 12.05, weight deacy loss is 8.67, training accuracy is 0.281250, time 300.066 samples/sec
+epoch 5, total_step 84940, total loss is 21.15 , inference loss is 12.48, weight deacy loss is 8.67, training accuracy is 0.406250, time 224.018 samples/sec
+epoch 5, total_step 84960, total loss is 20.47 , inference loss is 11.80, weight deacy loss is 8.67, training accuracy is 0.281250, time 212.818 samples/sec
+epoch 5, total_step 84980, total loss is 22.77 , inference loss is 14.10, weight deacy loss is 8.67, training accuracy is 0.156250, time 216.127 samples/sec
+epoch 5, total_step 85000, total loss is 14.91 , inference loss is 6.24, weight deacy loss is 8.67, training accuracy is 0.375000, time 201.049 samples/sec
+epoch 5, total_step 85020, total loss is 25.19 , inference loss is 16.52, weight deacy loss is 8.67, training accuracy is 0.187500, time 214.482 samples/sec
+epoch 5, total_step 85040, total loss is 18.59 , inference loss is 9.92, weight deacy loss is 8.67, training accuracy is 0.156250, time 207.098 samples/sec
+epoch 5, total_step 85060, total loss is 23.28 , inference loss is 14.62, weight deacy loss is 8.67, training accuracy is 0.312500, time 224.665 samples/sec
+epoch 5, total_step 85080, total loss is 20.61 , inference loss is 11.94, weight deacy loss is 8.67, training accuracy is 0.406250, time 207.415 samples/sec
+epoch 5, total_step 85100, total loss is 23.97 , inference loss is 15.31, weight deacy loss is 8.67, training accuracy is 0.218750, time 207.145 samples/sec
+epoch 5, total_step 85120, total loss is 18.63 , inference loss is 9.97, weight deacy loss is 8.66, training accuracy is 0.343750, time 205.882 samples/sec
+epoch 5, total_step 85140, total loss is 20.59 , inference loss is 11.93, weight deacy loss is 8.66, training accuracy is 0.281250, time 208.870 samples/sec
+epoch 5, total_step 85160, total loss is 20.70 , inference loss is 12.04, weight deacy loss is 8.66, training accuracy is 0.312500, time 212.626 samples/sec
+epoch 5, total_step 85180, total loss is 19.59 , inference loss is 10.93, weight deacy loss is 8.66, training accuracy is 0.250000, time 210.369 samples/sec
+epoch 5, total_step 85200, total loss is 24.15 , inference loss is 15.49, weight deacy loss is 8.66, training accuracy is 0.125000, time 15.207 samples/sec
+epoch 5, total_step 85220, total loss is 21.15 , inference loss is 12.49, weight deacy loss is 8.66, training accuracy is 0.250000, time 210.454 samples/sec
+epoch 5, total_step 85240, total loss is 22.42 , inference loss is 13.76, weight deacy loss is 8.66, training accuracy is 0.187500, time 214.850 samples/sec
+epoch 5, total_step 85260, total loss is 19.93 , inference loss is 11.27, weight deacy loss is 8.66, training accuracy is 0.250000, time 210.252 samples/sec
+epoch 5, total_step 85280, total loss is 22.06 , inference loss is 13.40, weight deacy loss is 8.66, training accuracy is 0.218750, time 206.139 samples/sec
+epoch 5, total_step 85300, total loss is 18.70 , inference loss is 10.04, weight deacy loss is 8.66, training accuracy is 0.281250, time 216.260 samples/sec
+epoch 5, total_step 85320, total loss is 19.22 , inference loss is 10.57, weight deacy loss is 8.66, training accuracy is 0.343750, time 210.879 samples/sec
+epoch 5, total_step 85340, total loss is 18.12 , inference loss is 9.47, weight deacy loss is 8.65, training accuracy is 0.375000, time 209.126 samples/sec
+epoch 5, total_step 85360, total loss is 15.85 , inference loss is 7.20, weight deacy loss is 8.65, training accuracy is 0.500000, time 213.352 samples/sec
+epoch 5, total_step 85380, total loss is 17.84 , inference loss is 9.18, weight deacy loss is 8.65, training accuracy is 0.281250, time 208.926 samples/sec
+epoch 5, total_step 85400, total loss is 20.35 , inference loss is 11.70, weight deacy loss is 8.65, training accuracy is 0.281250, time 206.810 samples/sec
+epoch 5, total_step 85420, total loss is 16.42 , inference loss is 7.76, weight deacy loss is 8.65, training accuracy is 0.406250, time 211.753 samples/sec
+epoch 5, total_step 85440, total loss is 19.42 , inference loss is 10.77, weight deacy loss is 8.65, training accuracy is 0.343750, time 204.890 samples/sec
+epoch 5, total_step 85460, total loss is 18.24 , inference loss is 9.59, weight deacy loss is 8.65, training accuracy is 0.343750, time 207.232 samples/sec
+epoch 5, total_step 85480, total loss is 19.12 , inference loss is 10.47, weight deacy loss is 8.65, training accuracy is 0.250000, time 204.107 samples/sec
+epoch 5, total_step 85500, total loss is 20.57 , inference loss is 11.92, weight deacy loss is 8.65, training accuracy is 0.187500, time 13.210 samples/sec
+epoch 5, total_step 85520, total loss is 20.51 , inference loss is 11.86, weight deacy loss is 8.65, training accuracy is 0.218750, time 203.191 samples/sec
+epoch 5, total_step 85540, total loss is 22.20 , inference loss is 13.55, weight deacy loss is 8.65, training accuracy is 0.218750, time 208.419 samples/sec
+epoch 5, total_step 85560, total loss is 19.00 , inference loss is 10.35, weight deacy loss is 8.65, training accuracy is 0.250000, time 209.722 samples/sec
+epoch 5, total_step 85580, total loss is 23.12 , inference loss is 14.47, weight deacy loss is 8.64, training accuracy is 0.312500, time 220.241 samples/sec
+epoch 5, total_step 85600, total loss is 16.90 , inference loss is 8.26, weight deacy loss is 8.64, training accuracy is 0.375000, time 206.930 samples/sec
+epoch 5, total_step 85620, total loss is 22.03 , inference loss is 13.39, weight deacy loss is 8.64, training accuracy is 0.187500, time 208.680 samples/sec
+epoch 5, total_step 85640, total loss is 15.63 , inference loss is 6.99, weight deacy loss is 8.64, training accuracy is 0.375000, time 207.801 samples/sec
+epoch 5, total_step 85660, total loss is 16.51 , inference loss is 7.87, weight deacy loss is 8.64, training accuracy is 0.312500, time 218.416 samples/sec
+epoch 5, total_step 85680, total loss is 19.54 , inference loss is 10.90, weight deacy loss is 8.64, training accuracy is 0.187500, time 213.450 samples/sec
+epoch 5, total_step 85700, total loss is 20.01 , inference loss is 11.37, weight deacy loss is 8.64, training accuracy is 0.250000, time 211.713 samples/sec
+epoch 5, total_step 85720, total loss is 18.28 , inference loss is 9.65, weight deacy loss is 8.64, training accuracy is 0.343750, time 204.273 samples/sec
+epoch 5, total_step 85740, total loss is 21.56 , inference loss is 12.93, weight deacy loss is 8.64, training accuracy is 0.218750, time 206.315 samples/sec
+epoch 5, total_step 85760, total loss is 18.24 , inference loss is 9.60, weight deacy loss is 8.64, training accuracy is 0.343750, time 206.500 samples/sec
+epoch 5, total_step 85780, total loss is 16.69 , inference loss is 8.06, weight deacy loss is 8.64, training accuracy is 0.375000, time 294.513 samples/sec
+epoch 5, total_step 85800, total loss is 22.75 , inference loss is 14.12, weight deacy loss is 8.63, training accuracy is 0.187500, time 13.765 samples/sec
+epoch 5, total_step 85820, total loss is 21.23 , inference loss is 12.60, weight deacy loss is 8.63, training accuracy is 0.218750, time 244.619 samples/sec
+epoch 5, total_step 85840, total loss is 20.58 , inference loss is 11.94, weight deacy loss is 8.63, training accuracy is 0.250000, time 227.788 samples/sec
+epoch 5, total_step 85860, total loss is 17.39 , inference loss is 8.76, weight deacy loss is 8.63, training accuracy is 0.281250, time 212.820 samples/sec
+epoch 5, total_step 85880, total loss is 20.35 , inference loss is 11.72, weight deacy loss is 8.63, training accuracy is 0.281250, time 211.720 samples/sec
+epoch 5, total_step 85900, total loss is 20.77 , inference loss is 12.14, weight deacy loss is 8.63, training accuracy is 0.250000, time 212.896 samples/sec
+epoch 5, total_step 85920, total loss is 19.92 , inference loss is 11.29, weight deacy loss is 8.63, training accuracy is 0.312500, time 205.422 samples/sec
+epoch 5, total_step 85940, total loss is 18.73 , inference loss is 10.10, weight deacy loss is 8.63, training accuracy is 0.187500, time 201.250 samples/sec
+epoch 5, total_step 85960, total loss is 19.57 , inference loss is 10.94, weight deacy loss is 8.63, training accuracy is 0.312500, time 212.819 samples/sec
+epoch 5, total_step 85980, total loss is 19.28 , inference loss is 10.66, weight deacy loss is 8.63, training accuracy is 0.250000, time 214.890 samples/sec
+testing verification..
+(12000, 512)
+infer time 22.956328000000006
+best_threshold_index 112 0.9864814814814815
+best_threshold_index 112 0.9853703703703703
+best_threshold_index 112 0.9861111111111112
+best_threshold_index 112 0.9862962962962963
+best_threshold_index 112 0.9859259259259259
+best_threshold_index 112 0.9851851851851852
+best_threshold_index 112 0.9857407407407407
+best_threshold_index 112 0.9853703703703703
+best_threshold_index 112 0.9851851851851852
+best_threshold_index 112 0.985
+[lfw][86000]XNorm: 21.944362
+[lfw][86000]Accuracy-Flip: 0.98567+-0.00442
+test accuracy is: 0.9856666666666667
+epoch 5, total_step 86000, total loss is 20.02 , inference loss is 11.39, weight deacy loss is 8.63, training accuracy is 0.218750, time 19.141 samples/sec
+epoch 5, total_step 86020, total loss is 16.49 , inference loss is 7.86, weight deacy loss is 8.63, training accuracy is 0.281250, time 275.826 samples/sec
+epoch 5, total_step 86040, total loss is 20.36 , inference loss is 11.73, weight deacy loss is 8.62, training accuracy is 0.187500, time 297.631 samples/sec
+epoch 5, total_step 86060, total loss is 17.88 , inference loss is 9.25, weight deacy loss is 8.62, training accuracy is 0.375000, time 208.781 samples/sec
+epoch 5, total_step 86080, total loss is 22.56 , inference loss is 13.93, weight deacy loss is 8.62, training accuracy is 0.156250, time 208.929 samples/sec
+epoch 5, total_step 86100, total loss is 22.44 , inference loss is 13.82, weight deacy loss is 8.62, training accuracy is 0.218750, time 14.289 samples/sec
+epoch 5, total_step 86120, total loss is 26.08 , inference loss is 17.46, weight deacy loss is 8.62, training accuracy is 0.093750, time 192.182 samples/sec
+epoch 5, total_step 86140, total loss is 19.11 , inference loss is 10.49, weight deacy loss is 8.62, training accuracy is 0.281250, time 207.500 samples/sec
+epoch 5, total_step 86160, total loss is 21.03 , inference loss is 12.41, weight deacy loss is 8.62, training accuracy is 0.281250, time 215.247 samples/sec
+epoch 5, total_step 86180, total loss is 23.32 , inference loss is 14.70, weight deacy loss is 8.62, training accuracy is 0.187500, time 214.230 samples/sec
+epoch 5, total_step 86200, total loss is 19.59 , inference loss is 10.97, weight deacy loss is 8.62, training accuracy is 0.375000, time 213.428 samples/sec
+epoch 5, total_step 86220, total loss is 20.65 , inference loss is 12.03, weight deacy loss is 8.62, training accuracy is 0.312500, time 210.461 samples/sec
+epoch 5, total_step 86240, total loss is 20.75 , inference loss is 12.13, weight deacy loss is 8.62, training accuracy is 0.343750, time 206.942 samples/sec
+epoch 5, total_step 86260, total loss is 18.86 , inference loss is 10.24, weight deacy loss is 8.62, training accuracy is 0.281250, time 213.555 samples/sec
+epoch 5, total_step 86280, total loss is 19.69 , inference loss is 11.08, weight deacy loss is 8.61, training accuracy is 0.218750, time 212.149 samples/sec
+epoch 5, total_step 86300, total loss is 20.29 , inference loss is 11.68, weight deacy loss is 8.61, training accuracy is 0.406250, time 299.449 samples/sec
+epoch 5, total_step 86320, total loss is 19.21 , inference loss is 10.59, weight deacy loss is 8.61, training accuracy is 0.187500, time 208.198 samples/sec
+epoch 5, total_step 86340, total loss is 20.34 , inference loss is 11.73, weight deacy loss is 8.61, training accuracy is 0.343750, time 211.377 samples/sec
+epoch 5, total_step 86360, total loss is 19.79 , inference loss is 11.18, weight deacy loss is 8.61, training accuracy is 0.312500, time 211.074 samples/sec
+epoch 5, total_step 86380, total loss is 21.49 , inference loss is 12.88, weight deacy loss is 8.61, training accuracy is 0.156250, time 209.756 samples/sec
+epoch 5, total_step 86400, total loss is 14.30 , inference loss is 5.70, weight deacy loss is 8.61, training accuracy is 0.343750, time 13.982 samples/sec
+epoch 5, total_step 86420, total loss is 18.15 , inference loss is 9.54, weight deacy loss is 8.61, training accuracy is 0.250000, time 211.785 samples/sec
+epoch 5, total_step 86440, total loss is 18.74 , inference loss is 10.14, weight deacy loss is 8.61, training accuracy is 0.250000, time 221.236 samples/sec
+epoch 5, total_step 86460, total loss is 18.49 , inference loss is 9.88, weight deacy loss is 8.61, training accuracy is 0.375000, time 208.506 samples/sec
+epoch 5, total_step 86480, total loss is 20.63 , inference loss is 12.03, weight deacy loss is 8.61, training accuracy is 0.218750, time 215.422 samples/sec
+epoch 5, total_step 86500, total loss is 18.06 , inference loss is 9.45, weight deacy loss is 8.60, training accuracy is 0.187500, time 214.194 samples/sec
+epoch 5, total_step 86520, total loss is 16.11 , inference loss is 7.50, weight deacy loss is 8.60, training accuracy is 0.406250, time 210.834 samples/sec
+epoch 5, total_step 86540, total loss is 19.45 , inference loss is 10.85, weight deacy loss is 8.60, training accuracy is 0.375000, time 206.987 samples/sec
+epoch 5, total_step 86560, total loss is 24.50 , inference loss is 15.90, weight deacy loss is 8.60, training accuracy is 0.187500, time 208.149 samples/sec
+epoch 5, total_step 86580, total loss is 22.03 , inference loss is 13.43, weight deacy loss is 8.60, training accuracy is 0.312500, time 300.168 samples/sec
+epoch 5, total_step 86600, total loss is 19.34 , inference loss is 10.74, weight deacy loss is 8.60, training accuracy is 0.437500, time 212.345 samples/sec
+epoch 5, total_step 86620, total loss is 19.64 , inference loss is 11.04, weight deacy loss is 8.60, training accuracy is 0.156250, time 207.991 samples/sec
+epoch 5, total_step 86640, total loss is 17.11 , inference loss is 8.51, weight deacy loss is 8.60, training accuracy is 0.187500, time 208.486 samples/sec
+epoch 5, total_step 86660, total loss is 21.40 , inference loss is 12.80, weight deacy loss is 8.60, training accuracy is 0.218750, time 206.445 samples/sec
+epoch 5, total_step 86680, total loss is 15.12 , inference loss is 6.53, weight deacy loss is 8.60, training accuracy is 0.406250, time 207.798 samples/sec
+epoch 5, total_step 86700, total loss is 19.18 , inference loss is 10.58, weight deacy loss is 8.60, training accuracy is 0.312500, time 14.677 samples/sec
+epoch 5, total_step 86720, total loss is 16.05 , inference loss is 7.45, weight deacy loss is 8.59, training accuracy is 0.468750, time 218.543 samples/sec
+epoch 5, total_step 86740, total loss is 17.05 , inference loss is 8.46, weight deacy loss is 8.59, training accuracy is 0.343750, time 206.142 samples/sec
+epoch 5, total_step 86760, total loss is 17.04 , inference loss is 8.45, weight deacy loss is 8.59, training accuracy is 0.312500, time 208.152 samples/sec
+epoch 5, total_step 86780, total loss is 19.42 , inference loss is 10.83, weight deacy loss is 8.59, training accuracy is 0.218750, time 206.162 samples/sec
+epoch 5, total_step 86800, total loss is 15.41 , inference loss is 6.82, weight deacy loss is 8.59, training accuracy is 0.468750, time 238.449 samples/sec
+epoch 5, total_step 86820, total loss is 21.99 , inference loss is 13.39, weight deacy loss is 8.59, training accuracy is 0.187500, time 214.555 samples/sec
+epoch 5, total_step 86840, total loss is 19.25 , inference loss is 10.66, weight deacy loss is 8.59, training accuracy is 0.218750, time 296.927 samples/sec
+epoch 5, total_step 86860, total loss is 13.79 , inference loss is 5.20, weight deacy loss is 8.59, training accuracy is 0.375000, time 224.311 samples/sec
+epoch 5, total_step 86880, total loss is 19.12 , inference loss is 10.53, weight deacy loss is 8.59, training accuracy is 0.218750, time 212.667 samples/sec
+epoch 5, total_step 86900, total loss is 17.44 , inference loss is 8.85, weight deacy loss is 8.59, training accuracy is 0.500000, time 209.018 samples/sec
+epoch 5, total_step 86920, total loss is 19.94 , inference loss is 11.36, weight deacy loss is 8.59, training accuracy is 0.250000, time 216.036 samples/sec
+epoch 5, total_step 86940, total loss is 18.60 , inference loss is 10.01, weight deacy loss is 8.59, training accuracy is 0.343750, time 220.904 samples/sec
+epoch 5, total_step 86960, total loss is 18.77 , inference loss is 10.19, weight deacy loss is 8.58, training accuracy is 0.187500, time 209.489 samples/sec
+epoch 5, total_step 86980, total loss is 16.70 , inference loss is 8.11, weight deacy loss is 8.58, training accuracy is 0.343750, time 208.689 samples/sec
+epoch 5, total_step 87000, total loss is 17.27 , inference loss is 8.69, weight deacy loss is 8.58, training accuracy is 0.250000, time 13.922 samples/sec
+epoch 5, total_step 87020, total loss is 18.25 , inference loss is 9.67, weight deacy loss is 8.58, training accuracy is 0.312500, time 220.508 samples/sec
+epoch 5, total_step 87040, total loss is 21.58 , inference loss is 13.00, weight deacy loss is 8.58, training accuracy is 0.281250, time 207.006 samples/sec
+epoch 5, total_step 87060, total loss is 18.28 , inference loss is 9.70, weight deacy loss is 8.58, training accuracy is 0.250000, time 201.925 samples/sec
+epoch 5, total_step 87080, total loss is 19.55 , inference loss is 10.98, weight deacy loss is 8.58, training accuracy is 0.250000, time 206.812 samples/sec
+epoch 5, total_step 87100, total loss is 17.93 , inference loss is 9.35, weight deacy loss is 8.58, training accuracy is 0.437500, time 212.586 samples/sec
+epoch 5, total_step 87120, total loss is 18.28 , inference loss is 9.70, weight deacy loss is 8.58, training accuracy is 0.343750, time 281.862 samples/sec
+epoch 5, total_step 87140, total loss is 20.93 , inference loss is 12.36, weight deacy loss is 8.58, training accuracy is 0.187500, time 226.150 samples/sec
+epoch 5, total_step 87160, total loss is 15.41 , inference loss is 6.84, weight deacy loss is 8.58, training accuracy is 0.437500, time 222.565 samples/sec
+epoch 5, total_step 87180, total loss is 21.09 , inference loss is 12.52, weight deacy loss is 8.57, training accuracy is 0.343750, time 206.366 samples/sec
+epoch 5, total_step 87200, total loss is 16.19 , inference loss is 7.62, weight deacy loss is 8.57, training accuracy is 0.375000, time 203.941 samples/sec
+epoch 5, total_step 87220, total loss is 20.23 , inference loss is 11.66, weight deacy loss is 8.57, training accuracy is 0.281250, time 208.905 samples/sec
+epoch 5, total_step 87240, total loss is 17.18 , inference loss is 8.61, weight deacy loss is 8.57, training accuracy is 0.250000, time 207.521 samples/sec
+epoch 5, total_step 87260, total loss is 22.46 , inference loss is 13.89, weight deacy loss is 8.57, training accuracy is 0.312500, time 228.613 samples/sec
+epoch 5, total_step 87280, total loss is 18.45 , inference loss is 9.88, weight deacy loss is 8.57, training accuracy is 0.312500, time 204.919 samples/sec
+epoch 5, total_step 87300, total loss is 19.91 , inference loss is 11.34, weight deacy loss is 8.57, training accuracy is 0.218750, time 14.105 samples/sec
+epoch 5, total_step 87320, total loss is 20.48 , inference loss is 11.91, weight deacy loss is 8.57, training accuracy is 0.156250, time 208.549 samples/sec
+epoch 5, total_step 87340, total loss is 17.46 , inference loss is 8.89, weight deacy loss is 8.57, training accuracy is 0.312500, time 199.050 samples/sec
+epoch 5, total_step 87360, total loss is 23.94 , inference loss is 15.38, weight deacy loss is 8.57, training accuracy is 0.125000, time 212.057 samples/sec
+epoch 5, total_step 87380, total loss is 19.92 , inference loss is 11.35, weight deacy loss is 8.57, training accuracy is 0.156250, time 298.423 samples/sec
+epoch 5, total_step 87400, total loss is 20.99 , inference loss is 12.42, weight deacy loss is 8.57, training accuracy is 0.187500, time 208.373 samples/sec
+epoch 5, total_step 87420, total loss is 15.25 , inference loss is 6.69, weight deacy loss is 8.56, training accuracy is 0.281250, time 197.704 samples/sec
+epoch 5, total_step 87440, total loss is 19.16 , inference loss is 10.60, weight deacy loss is 8.56, training accuracy is 0.250000, time 213.558 samples/sec
+epoch 5, total_step 87460, total loss is 20.13 , inference loss is 11.57, weight deacy loss is 8.56, training accuracy is 0.312500, time 208.506 samples/sec
+epoch 5, total_step 87480, total loss is 16.98 , inference loss is 8.42, weight deacy loss is 8.56, training accuracy is 0.375000, time 216.972 samples/sec
+epoch 5, total_step 87500, total loss is 15.48 , inference loss is 6.92, weight deacy loss is 8.56, training accuracy is 0.406250, time 207.137 samples/sec
+epoch 5, total_step 87520, total loss is 20.91 , inference loss is 12.35, weight deacy loss is 8.56, training accuracy is 0.250000, time 215.630 samples/sec
+epoch 5, total_step 87540, total loss is 22.39 , inference loss is 13.83, weight deacy loss is 8.56, training accuracy is 0.187500, time 216.443 samples/sec
+epoch 5, total_step 87560, total loss is 19.40 , inference loss is 10.84, weight deacy loss is 8.56, training accuracy is 0.281250, time 216.031 samples/sec
+epoch 5, total_step 87580, total loss is 17.75 , inference loss is 9.20, weight deacy loss is 8.56, training accuracy is 0.343750, time 207.348 samples/sec
+epoch 5, total_step 87600, total loss is 19.46 , inference loss is 10.91, weight deacy loss is 8.56, training accuracy is 0.187500, time 13.030 samples/sec
+epoch 5, total_step 87620, total loss is 19.18 , inference loss is 10.62, weight deacy loss is 8.56, training accuracy is 0.281250, time 210.188 samples/sec
+epoch 5, total_step 87640, total loss is 21.29 , inference loss is 12.73, weight deacy loss is 8.56, training accuracy is 0.218750, time 297.514 samples/sec
+epoch 5, total_step 87660, total loss is 17.52 , inference loss is 8.97, weight deacy loss is 8.55, training accuracy is 0.375000, time 213.965 samples/sec
+epoch 5, total_step 87680, total loss is 17.42 , inference loss is 8.86, weight deacy loss is 8.55, training accuracy is 0.343750, time 212.156 samples/sec
+epoch 5, total_step 87700, total loss is 15.80 , inference loss is 7.25, weight deacy loss is 8.55, training accuracy is 0.343750, time 197.922 samples/sec
+epoch 5, total_step 87720, total loss is 21.14 , inference loss is 12.59, weight deacy loss is 8.55, training accuracy is 0.218750, time 212.619 samples/sec
+epoch 5, total_step 87740, total loss is 17.82 , inference loss is 9.27, weight deacy loss is 8.55, training accuracy is 0.312500, time 203.741 samples/sec
+epoch 5, total_step 87760, total loss is 19.44 , inference loss is 10.88, weight deacy loss is 8.55, training accuracy is 0.343750, time 211.540 samples/sec
+epoch 5, total_step 87780, total loss is 20.80 , inference loss is 12.25, weight deacy loss is 8.55, training accuracy is 0.250000, time 219.270 samples/sec
+epoch 5, total_step 87800, total loss is 15.61 , inference loss is 7.06, weight deacy loss is 8.55, training accuracy is 0.312500, time 208.875 samples/sec
+epoch 5, total_step 87820, total loss is 24.55 , inference loss is 16.00, weight deacy loss is 8.55, training accuracy is 0.250000, time 214.333 samples/sec
+epoch 5, total_step 87840, total loss is 19.65 , inference loss is 11.10, weight deacy loss is 8.55, training accuracy is 0.218750, time 227.683 samples/sec
+epoch 5, total_step 87860, total loss is 16.84 , inference loss is 8.29, weight deacy loss is 8.55, training accuracy is 0.187500, time 212.813 samples/sec
+epoch 5, total_step 87880, total loss is 18.26 , inference loss is 9.71, weight deacy loss is 8.55, training accuracy is 0.218750, time 215.886 samples/sec
+epoch 5, total_step 87900, total loss is 20.01 , inference loss is 11.46, weight deacy loss is 8.54, training accuracy is 0.437500, time 14.897 samples/sec
+epoch 5, total_step 87920, total loss is 18.18 , inference loss is 9.64, weight deacy loss is 8.54, training accuracy is 0.343750, time 219.229 samples/sec
+epoch 5, total_step 87940, total loss is 19.72 , inference loss is 11.18, weight deacy loss is 8.54, training accuracy is 0.218750, time 218.718 samples/sec
+epoch 5, total_step 87960, total loss is 16.17 , inference loss is 7.63, weight deacy loss is 8.54, training accuracy is 0.343750, time 298.311 samples/sec
+epoch 5, total_step 87980, total loss is 17.32 , inference loss is 8.78, weight deacy loss is 8.54, training accuracy is 0.312500, time 296.736 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.87712899999998
+best_threshold_index 106 0.9838888888888889
+best_threshold_index 106 0.9827777777777778
+best_threshold_index 106 0.9844444444444445
+best_threshold_index 106 0.9853703703703703
+best_threshold_index 106 0.9837037037037037
+best_threshold_index 106 0.9827777777777778
+best_threshold_index 106 0.9835185185185186
+best_threshold_index 106 0.9827777777777778
+best_threshold_index 106 0.9825925925925926
+best_threshold_index 106 0.9831481481481481
+[lfw][88000]XNorm: 21.735142
+[lfw][88000]Accuracy-Flip: 0.98350+-0.00758
+test accuracy is: 0.9834999999999999
+epoch 5, total_step 88000, total loss is 20.30 , inference loss is 11.76, weight deacy loss is 8.54, training accuracy is 0.187500, time 19.302 samples/sec
+epoch 5, total_step 88020, total loss is 21.40 , inference loss is 12.86, weight deacy loss is 8.54, training accuracy is 0.156250, time 202.297 samples/sec
+epoch 5, total_step 88040, total loss is 16.42 , inference loss is 7.89, weight deacy loss is 8.54, training accuracy is 0.343750, time 213.553 samples/sec
+epoch 5, total_step 88060, total loss is 23.04 , inference loss is 14.50, weight deacy loss is 8.54, training accuracy is 0.281250, time 207.314 samples/sec
+epoch 5, total_step 88080, total loss is 20.25 , inference loss is 11.71, weight deacy loss is 8.54, training accuracy is 0.250000, time 216.578 samples/sec
+epoch 5, total_step 88100, total loss is 19.36 , inference loss is 10.82, weight deacy loss is 8.54, training accuracy is 0.375000, time 209.043 samples/sec
+epoch 5, total_step 88120, total loss is 17.77 , inference loss is 9.24, weight deacy loss is 8.54, training accuracy is 0.250000, time 210.881 samples/sec
+epoch 5, total_step 88140, total loss is 20.65 , inference loss is 12.12, weight deacy loss is 8.53, training accuracy is 0.250000, time 207.284 samples/sec
+epoch 5, total_step 88160, total loss is 15.37 , inference loss is 6.84, weight deacy loss is 8.53, training accuracy is 0.281250, time 211.340 samples/sec
+epoch 5, total_step 88180, total loss is 22.12 , inference loss is 13.59, weight deacy loss is 8.53, training accuracy is 0.187500, time 214.273 samples/sec
+epoch 5, total_step 88200, total loss is 21.81 , inference loss is 13.27, weight deacy loss is 8.53, training accuracy is 0.250000, time 14.480 samples/sec
+epoch 5, total_step 88220, total loss is 17.84 , inference loss is 9.31, weight deacy loss is 8.53, training accuracy is 0.281250, time 207.197 samples/sec
+epoch 5, total_step 88240, total loss is 17.26 , inference loss is 8.73, weight deacy loss is 8.53, training accuracy is 0.312500, time 207.423 samples/sec
+epoch 5, total_step 88260, total loss is 23.33 , inference loss is 14.80, weight deacy loss is 8.53, training accuracy is 0.312500, time 210.674 samples/sec
+epoch 5, total_step 88280, total loss is 21.50 , inference loss is 12.97, weight deacy loss is 8.53, training accuracy is 0.312500, time 207.482 samples/sec
+epoch 5, total_step 88300, total loss is 22.55 , inference loss is 14.03, weight deacy loss is 8.53, training accuracy is 0.187500, time 207.633 samples/sec
+epoch 5, total_step 88320, total loss is 23.24 , inference loss is 14.71, weight deacy loss is 8.53, training accuracy is 0.156250, time 215.405 samples/sec
+epoch 5, total_step 88340, total loss is 19.93 , inference loss is 11.40, weight deacy loss is 8.53, training accuracy is 0.281250, time 210.102 samples/sec
+epoch 5, total_step 88360, total loss is 20.04 , inference loss is 11.52, weight deacy loss is 8.52, training accuracy is 0.375000, time 215.522 samples/sec
+epoch 5, total_step 88380, total loss is 21.37 , inference loss is 12.84, weight deacy loss is 8.52, training accuracy is 0.187500, time 217.905 samples/sec
+epoch 5, total_step 88400, total loss is 18.75 , inference loss is 10.23, weight deacy loss is 8.52, training accuracy is 0.281250, time 226.254 samples/sec
+epoch 5, total_step 88420, total loss is 19.05 , inference loss is 10.53, weight deacy loss is 8.52, training accuracy is 0.281250, time 204.361 samples/sec
+epoch 5, total_step 88440, total loss is 20.45 , inference loss is 11.93, weight deacy loss is 8.52, training accuracy is 0.281250, time 212.129 samples/sec
+epoch 5, total_step 88460, total loss is 22.24 , inference loss is 13.72, weight deacy loss is 8.52, training accuracy is 0.156250, time 207.258 samples/sec
+epoch 5, total_step 88480, total loss is 20.46 , inference loss is 11.94, weight deacy loss is 8.52, training accuracy is 0.156250, time 204.905 samples/sec
+epoch 5, total_step 88500, total loss is 17.34 , inference loss is 8.82, weight deacy loss is 8.52, training accuracy is 0.375000, time 12.477 samples/sec
+epoch 5, total_step 88520, total loss is 20.55 , inference loss is 12.03, weight deacy loss is 8.52, training accuracy is 0.218750, time 203.177 samples/sec
+epoch 5, total_step 88540, total loss is 20.18 , inference loss is 11.67, weight deacy loss is 8.52, training accuracy is 0.187500, time 209.514 samples/sec
+epoch 5, total_step 88560, total loss is 20.31 , inference loss is 11.80, weight deacy loss is 8.52, training accuracy is 0.281250, time 212.658 samples/sec
+epoch 5, total_step 88580, total loss is 22.07 , inference loss is 13.55, weight deacy loss is 8.52, training accuracy is 0.281250, time 207.224 samples/sec
+epoch 5, total_step 88600, total loss is 19.47 , inference loss is 10.96, weight deacy loss is 8.51, training accuracy is 0.250000, time 208.480 samples/sec
+epoch 5, total_step 88620, total loss is 20.18 , inference loss is 11.66, weight deacy loss is 8.51, training accuracy is 0.187500, time 231.337 samples/sec
+epoch 5, total_step 88640, total loss is 17.09 , inference loss is 8.58, weight deacy loss is 8.51, training accuracy is 0.312500, time 212.990 samples/sec
+epoch 5, total_step 88660, total loss is 20.33 , inference loss is 11.82, weight deacy loss is 8.51, training accuracy is 0.187500, time 215.007 samples/sec
+epoch 5, total_step 88680, total loss is 20.29 , inference loss is 11.78, weight deacy loss is 8.51, training accuracy is 0.281250, time 213.788 samples/sec
+epoch 5, total_step 88700, total loss is 20.36 , inference loss is 11.85, weight deacy loss is 8.51, training accuracy is 0.093750, time 220.651 samples/sec
+epoch 5, total_step 88720, total loss is 17.93 , inference loss is 9.42, weight deacy loss is 8.51, training accuracy is 0.406250, time 209.115 samples/sec
+epoch 5, total_step 88740, total loss is 20.83 , inference loss is 12.32, weight deacy loss is 8.51, training accuracy is 0.187500, time 205.883 samples/sec
+epoch 5, total_step 88760, total loss is 21.22 , inference loss is 12.71, weight deacy loss is 8.51, training accuracy is 0.312500, time 297.594 samples/sec
+epoch 5, total_step 88780, total loss is 18.57 , inference loss is 10.06, weight deacy loss is 8.51, training accuracy is 0.218750, time 223.450 samples/sec
+epoch 5, total_step 88800, total loss is 15.42 , inference loss is 6.91, weight deacy loss is 8.51, training accuracy is 0.375000, time 14.626 samples/sec
+epoch 5, total_step 88820, total loss is 24.11 , inference loss is 15.61, weight deacy loss is 8.51, training accuracy is 0.250000, time 214.499 samples/sec
+epoch 5, total_step 88840, total loss is 21.05 , inference loss is 12.55, weight deacy loss is 8.50, training accuracy is 0.218750, time 219.106 samples/sec
+epoch 5, total_step 88860, total loss is 19.98 , inference loss is 11.48, weight deacy loss is 8.50, training accuracy is 0.250000, time 227.074 samples/sec
+epoch 5, total_step 88880, total loss is 16.51 , inference loss is 8.01, weight deacy loss is 8.50, training accuracy is 0.406250, time 211.365 samples/sec
+epoch 5, total_step 88900, total loss is 20.72 , inference loss is 12.22, weight deacy loss is 8.50, training accuracy is 0.218750, time 215.977 samples/sec
+epoch 5, total_step 88920, total loss is 21.49 , inference loss is 12.99, weight deacy loss is 8.50, training accuracy is 0.250000, time 212.719 samples/sec
+epoch 5, total_step 88940, total loss is 18.81 , inference loss is 10.31, weight deacy loss is 8.50, training accuracy is 0.250000, time 201.227 samples/sec
+epoch 5, total_step 88960, total loss is 18.36 , inference loss is 9.86, weight deacy loss is 8.50, training accuracy is 0.281250, time 214.036 samples/sec
+epoch 5, total_step 88980, total loss is 21.49 , inference loss is 12.99, weight deacy loss is 8.50, training accuracy is 0.250000, time 211.232 samples/sec
+epoch 5, total_step 89000, total loss is 20.87 , inference loss is 12.37, weight deacy loss is 8.50, training accuracy is 0.218750, time 205.975 samples/sec
+epoch 5, total_step 89020, total loss is 18.48 , inference loss is 9.98, weight deacy loss is 8.50, training accuracy is 0.218750, time 298.491 samples/sec
+epoch 5, total_step 89040, total loss is 19.12 , inference loss is 10.62, weight deacy loss is 8.50, training accuracy is 0.312500, time 210.755 samples/sec
+epoch 5, total_step 89060, total loss is 18.82 , inference loss is 10.32, weight deacy loss is 8.50, training accuracy is 0.156250, time 215.015 samples/sec
+epoch 5, total_step 89080, total loss is 23.32 , inference loss is 14.83, weight deacy loss is 8.49, training accuracy is 0.218750, time 220.518 samples/sec
+epoch 5, total_step 89100, total loss is 19.06 , inference loss is 10.57, weight deacy loss is 8.49, training accuracy is 0.281250, time 13.961 samples/sec
+epoch 5, total_step 89120, total loss is 18.87 , inference loss is 10.37, weight deacy loss is 8.49, training accuracy is 0.125000, time 213.734 samples/sec
+epoch 5, total_step 89140, total loss is 18.75 , inference loss is 10.26, weight deacy loss is 8.49, training accuracy is 0.312500, time 210.233 samples/sec
+epoch 5, total_step 89160, total loss is 21.31 , inference loss is 12.82, weight deacy loss is 8.49, training accuracy is 0.250000, time 207.280 samples/sec
+epoch 5, total_step 89180, total loss is 17.32 , inference loss is 8.83, weight deacy loss is 8.49, training accuracy is 0.312500, time 229.878 samples/sec
+epoch 5, total_step 89200, total loss is 16.33 , inference loss is 7.84, weight deacy loss is 8.49, training accuracy is 0.375000, time 208.794 samples/sec
+epoch 5, total_step 89220, total loss is 19.50 , inference loss is 11.01, weight deacy loss is 8.49, training accuracy is 0.218750, time 211.241 samples/sec
+epoch 5, total_step 89240, total loss is 21.98 , inference loss is 13.49, weight deacy loss is 8.49, training accuracy is 0.250000, time 213.990 samples/sec
+epoch 5, total_step 89260, total loss is 20.95 , inference loss is 12.46, weight deacy loss is 8.49, training accuracy is 0.312500, time 235.855 samples/sec
+epoch 5, total_step 89280, total loss is 21.70 , inference loss is 13.22, weight deacy loss is 8.49, training accuracy is 0.218750, time 298.297 samples/sec
+epoch 5, total_step 89300, total loss is 18.30 , inference loss is 9.81, weight deacy loss is 8.49, training accuracy is 0.406250, time 216.783 samples/sec
+epoch 5, total_step 89320, total loss is 26.00 , inference loss is 17.51, weight deacy loss is 8.48, training accuracy is 0.093750, time 205.867 samples/sec
+epoch 5, total_step 89340, total loss is 16.74 , inference loss is 8.25, weight deacy loss is 8.48, training accuracy is 0.312500, time 211.614 samples/sec
+epoch 5, total_step 89360, total loss is 21.46 , inference loss is 12.98, weight deacy loss is 8.48, training accuracy is 0.281250, time 210.432 samples/sec
+epoch 5, total_step 89380, total loss is 20.50 , inference loss is 12.02, weight deacy loss is 8.48, training accuracy is 0.250000, time 211.761 samples/sec
+epoch 5, total_step 89400, total loss is 19.12 , inference loss is 10.64, weight deacy loss is 8.48, training accuracy is 0.250000, time 12.802 samples/sec
+epoch 5, total_step 89420, total loss is 18.10 , inference loss is 9.62, weight deacy loss is 8.48, training accuracy is 0.281250, time 221.248 samples/sec
+epoch 5, total_step 89440, total loss is 21.58 , inference loss is 13.10, weight deacy loss is 8.48, training accuracy is 0.156250, time 205.539 samples/sec
+epoch 5, total_step 89460, total loss is 20.19 , inference loss is 11.71, weight deacy loss is 8.48, training accuracy is 0.187500, time 210.873 samples/sec
+epoch 5, total_step 89480, total loss is 24.82 , inference loss is 16.34, weight deacy loss is 8.48, training accuracy is 0.187500, time 231.562 samples/sec
+epoch 5, total_step 89500, total loss is 16.64 , inference loss is 8.16, weight deacy loss is 8.48, training accuracy is 0.343750, time 201.779 samples/sec
+epoch 5, total_step 89520, total loss is 21.73 , inference loss is 13.25, weight deacy loss is 8.48, training accuracy is 0.187500, time 211.599 samples/sec
+epoch 5, total_step 89540, total loss is 17.20 , inference loss is 8.73, weight deacy loss is 8.48, training accuracy is 0.343750, time 212.999 samples/sec
+epoch 5, total_step 89560, total loss is 21.77 , inference loss is 13.30, weight deacy loss is 8.47, training accuracy is 0.187500, time 299.581 samples/sec
+epoch 5, total_step 89580, total loss is 27.82 , inference loss is 19.34, weight deacy loss is 8.47, training accuracy is 0.093750, time 206.018 samples/sec
+epoch 5, total_step 89600, total loss is 17.12 , inference loss is 8.65, weight deacy loss is 8.47, training accuracy is 0.250000, time 207.065 samples/sec
+epoch 5, total_step 89620, total loss is 22.22 , inference loss is 13.75, weight deacy loss is 8.47, training accuracy is 0.156250, time 219.818 samples/sec
+epoch 5, total_step 89640, total loss is 17.63 , inference loss is 9.16, weight deacy loss is 8.47, training accuracy is 0.218750, time 214.240 samples/sec
+epoch 5, total_step 89660, total loss is 24.06 , inference loss is 15.59, weight deacy loss is 8.47, training accuracy is 0.093750, time 206.700 samples/sec
+epoch 5, total_step 89680, total loss is 19.34 , inference loss is 10.88, weight deacy loss is 8.47, training accuracy is 0.187500, time 214.656 samples/sec
+epoch 5, total_step 89700, total loss is 18.73 , inference loss is 10.26, weight deacy loss is 8.47, training accuracy is 0.281250, time 11.184 samples/sec
+epoch 5, total_step 89720, total loss is 21.75 , inference loss is 13.28, weight deacy loss is 8.47, training accuracy is 0.187500, time 216.151 samples/sec
+epoch 5, total_step 89740, total loss is 19.83 , inference loss is 11.37, weight deacy loss is 8.47, training accuracy is 0.250000, time 224.769 samples/sec
+epoch 5, total_step 89760, total loss is 21.95 , inference loss is 13.49, weight deacy loss is 8.47, training accuracy is 0.218750, time 298.681 samples/sec
+epoch 5, total_step 89780, total loss is 21.31 , inference loss is 12.85, weight deacy loss is 8.47, training accuracy is 0.187500, time 299.772 samples/sec
+epoch 5, total_step 89800, total loss is 21.09 , inference loss is 12.62, weight deacy loss is 8.46, training accuracy is 0.187500, time 208.983 samples/sec
+epoch 5, total_step 89820, total loss is 25.68 , inference loss is 17.22, weight deacy loss is 8.46, training accuracy is 0.125000, time 203.276 samples/sec
+epoch 5, total_step 89840, total loss is 17.41 , inference loss is 8.95, weight deacy loss is 8.46, training accuracy is 0.281250, time 213.952 samples/sec
+epoch 5, total_step 89860, total loss is 24.31 , inference loss is 15.85, weight deacy loss is 8.46, training accuracy is 0.093750, time 203.697 samples/sec
+epoch 5, total_step 89880, total loss is 17.43 , inference loss is 8.97, weight deacy loss is 8.46, training accuracy is 0.187500, time 208.154 samples/sec
+epoch 5, total_step 89900, total loss is 18.02 , inference loss is 9.56, weight deacy loss is 8.46, training accuracy is 0.312500, time 202.975 samples/sec
+epoch 5, total_step 89920, total loss is 23.54 , inference loss is 15.08, weight deacy loss is 8.46, training accuracy is 0.312500, time 211.501 samples/sec
+epoch 5, total_step 89940, total loss is 19.10 , inference loss is 10.65, weight deacy loss is 8.46, training accuracy is 0.250000, time 210.439 samples/sec
+epoch 5, total_step 89960, total loss is 19.33 , inference loss is 10.87, weight deacy loss is 8.46, training accuracy is 0.250000, time 211.174 samples/sec
+epoch 5, total_step 89980, total loss is 20.29 , inference loss is 11.83, weight deacy loss is 8.46, training accuracy is 0.218750, time 211.766 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.101429000000028
+best_threshold_index 108 0.9840740740740741
+best_threshold_index 108 0.9824074074074074
+best_threshold_index 107 0.9831481481481481
+best_threshold_index 107 0.9851851851851852
+best_threshold_index 104 0.9838888888888889
+best_threshold_index 107 0.9816666666666667
+best_threshold_index 107 0.9829629629629629
+best_threshold_index 107 0.9827777777777778
+best_threshold_index 104 0.9811111111111112
+best_threshold_index 104 0.9814814814814815
+[lfw][90000]XNorm: 22.199885
+[lfw][90000]Accuracy-Flip: 0.98200+-0.01061
+test accuracy is: 0.982
+epoch 5, total_step 90000, total loss is 24.26 , inference loss is 15.81, weight deacy loss is 8.46, training accuracy is 0.187500, time 18.081 samples/sec
+epoch 5, total_step 90020, total loss is 19.30 , inference loss is 10.84, weight deacy loss is 8.46, training accuracy is 0.156250, time 211.136 samples/sec
+epoch 5, total_step 90040, total loss is 14.84 , inference loss is 6.39, weight deacy loss is 8.45, training accuracy is 0.312500, time 297.183 samples/sec
+epoch 5, total_step 90060, total loss is 18.95 , inference loss is 10.50, weight deacy loss is 8.45, training accuracy is 0.218750, time 212.107 samples/sec
+epoch 5, total_step 90080, total loss is 20.35 , inference loss is 11.90, weight deacy loss is 8.45, training accuracy is 0.250000, time 211.343 samples/sec
+epoch 5, total_step 90100, total loss is 20.93 , inference loss is 12.48, weight deacy loss is 8.45, training accuracy is 0.187500, time 213.486 samples/sec
+epoch 5, total_step 90120, total loss is 20.54 , inference loss is 12.09, weight deacy loss is 8.45, training accuracy is 0.343750, time 229.046 samples/sec
+epoch 5, total_step 90140, total loss is 20.61 , inference loss is 12.16, weight deacy loss is 8.45, training accuracy is 0.187500, time 211.178 samples/sec
+epoch 5, total_step 90160, total loss is 20.72 , inference loss is 12.27, weight deacy loss is 8.45, training accuracy is 0.218750, time 204.949 samples/sec
+epoch 5, total_step 90180, total loss is 20.15 , inference loss is 11.71, weight deacy loss is 8.45, training accuracy is 0.218750, time 207.366 samples/sec
+epoch 5, total_step 90200, total loss is 19.21 , inference loss is 10.77, weight deacy loss is 8.45, training accuracy is 0.281250, time 213.021 samples/sec
+epoch 5, total_step 90220, total loss is 16.21 , inference loss is 7.77, weight deacy loss is 8.45, training accuracy is 0.343750, time 215.825 samples/sec
+epoch 5, total_step 90240, total loss is 19.43 , inference loss is 10.98, weight deacy loss is 8.45, training accuracy is 0.312500, time 212.138 samples/sec
+epoch 5, total_step 90260, total loss is 15.28 , inference loss is 6.84, weight deacy loss is 8.45, training accuracy is 0.187500, time 212.237 samples/sec
+epoch 5, total_step 90280, total loss is 22.07 , inference loss is 13.63, weight deacy loss is 8.44, training accuracy is 0.093750, time 213.856 samples/sec
+epoch 5, total_step 90300, total loss is 24.01 , inference loss is 15.56, weight deacy loss is 8.44, training accuracy is 0.156250, time 13.553 samples/sec
+epoch 5, total_step 90320, total loss is 15.24 , inference loss is 6.80, weight deacy loss is 8.44, training accuracy is 0.312500, time 224.454 samples/sec
+epoch 5, total_step 90340, total loss is 19.68 , inference loss is 11.24, weight deacy loss is 8.44, training accuracy is 0.281250, time 206.546 samples/sec
+epoch 5, total_step 90360, total loss is 20.10 , inference loss is 11.66, weight deacy loss is 8.44, training accuracy is 0.187500, time 211.888 samples/sec
+epoch 5, total_step 90380, total loss is 17.22 , inference loss is 8.78, weight deacy loss is 8.44, training accuracy is 0.125000, time 210.854 samples/sec
+epoch 5, total_step 90400, total loss is 21.01 , inference loss is 12.57, weight deacy loss is 8.44, training accuracy is 0.187500, time 202.840 samples/sec
+epoch 5, total_step 90420, total loss is 19.53 , inference loss is 11.09, weight deacy loss is 8.44, training accuracy is 0.187500, time 218.660 samples/sec
+epoch 5, total_step 90440, total loss is 20.88 , inference loss is 12.44, weight deacy loss is 8.44, training accuracy is 0.187500, time 217.088 samples/sec
+epoch 5, total_step 90460, total loss is 18.94 , inference loss is 10.50, weight deacy loss is 8.44, training accuracy is 0.187500, time 210.558 samples/sec
+epoch 5, total_step 90480, total loss is 18.81 , inference loss is 10.37, weight deacy loss is 8.44, training accuracy is 0.281250, time 225.763 samples/sec
+epoch 5, total_step 90500, total loss is 20.01 , inference loss is 11.58, weight deacy loss is 8.43, training accuracy is 0.406250, time 204.767 samples/sec
+epoch 5, total_step 90520, total loss is 18.42 , inference loss is 9.99, weight deacy loss is 8.43, training accuracy is 0.218750, time 208.838 samples/sec
+epoch 5, total_step 90540, total loss is 18.51 , inference loss is 10.08, weight deacy loss is 8.43, training accuracy is 0.250000, time 204.970 samples/sec
+epoch 5, total_step 90560, total loss is 17.10 , inference loss is 8.67, weight deacy loss is 8.43, training accuracy is 0.312500, time 217.639 samples/sec
+epoch 5, total_step 90580, total loss is 17.01 , inference loss is 8.58, weight deacy loss is 8.43, training accuracy is 0.312500, time 207.403 samples/sec
+epoch 5, total_step 90600, total loss is 21.29 , inference loss is 12.85, weight deacy loss is 8.43, training accuracy is 0.218750, time 13.714 samples/sec
+epoch 5, total_step 90620, total loss is 19.00 , inference loss is 10.57, weight deacy loss is 8.43, training accuracy is 0.218750, time 214.796 samples/sec
+epoch 5, total_step 90640, total loss is 23.02 , inference loss is 14.60, weight deacy loss is 8.43, training accuracy is 0.156250, time 211.457 samples/sec
+epoch 5, total_step 90660, total loss is 26.39 , inference loss is 17.96, weight deacy loss is 8.43, training accuracy is 0.125000, time 215.423 samples/sec
+epoch 5, total_step 90680, total loss is 20.58 , inference loss is 12.15, weight deacy loss is 8.43, training accuracy is 0.125000, time 206.627 samples/sec
+epoch 5, total_step 90700, total loss is 23.18 , inference loss is 14.76, weight deacy loss is 8.43, training accuracy is 0.000000, time 198.855 samples/sec
+epoch 5, total_step 90720, total loss is 25.01 , inference loss is 16.58, weight deacy loss is 8.43, training accuracy is 0.187500, time 211.443 samples/sec
+epoch 5, total_step 90740, total loss is 17.54 , inference loss is 9.11, weight deacy loss is 8.43, training accuracy is 0.250000, time 212.912 samples/sec
+epoch 5, total_step 90760, total loss is 18.29 , inference loss is 9.86, weight deacy loss is 8.42, training accuracy is 0.125000, time 210.304 samples/sec
+epoch 5, total_step 90780, total loss is 22.08 , inference loss is 13.65, weight deacy loss is 8.42, training accuracy is 0.156250, time 205.265 samples/sec
+epoch 5, total_step 90800, total loss is 21.95 , inference loss is 13.52, weight deacy loss is 8.42, training accuracy is 0.250000, time 213.820 samples/sec
+epoch 5, total_step 90820, total loss is 19.25 , inference loss is 10.83, weight deacy loss is 8.42, training accuracy is 0.281250, time 213.334 samples/sec
+epoch 5, total_step 90840, total loss is 20.57 , inference loss is 12.15, weight deacy loss is 8.42, training accuracy is 0.156250, time 210.037 samples/sec
+epoch 5, total_step 90860, total loss is 19.01 , inference loss is 10.59, weight deacy loss is 8.42, training accuracy is 0.156250, time 210.108 samples/sec
+epoch 5, total_step 90880, total loss is 22.59 , inference loss is 14.17, weight deacy loss is 8.42, training accuracy is 0.187500, time 227.056 samples/sec
+epoch 5, total_step 90900, total loss is 19.89 , inference loss is 11.48, weight deacy loss is 8.42, training accuracy is 0.125000, time 11.871 samples/sec
+epoch 5, total_step 90920, total loss is 19.74 , inference loss is 11.32, weight deacy loss is 8.42, training accuracy is 0.187500, time 218.489 samples/sec
+epoch 5, total_step 90940, total loss is 20.86 , inference loss is 12.44, weight deacy loss is 8.42, training accuracy is 0.187500, time 210.335 samples/sec
+epoch 5, total_step 90960, total loss is 19.70 , inference loss is 11.28, weight deacy loss is 8.42, training accuracy is 0.218750, time 207.335 samples/sec
+epoch 5, total_step 90980, total loss is 18.18 , inference loss is 9.76, weight deacy loss is 8.42, training accuracy is 0.250000, time 206.987 samples/sec
+epoch 5, total_step 91000, total loss is 21.51 , inference loss is 13.10, weight deacy loss is 8.41, training accuracy is 0.156250, time 202.590 samples/sec2022-12-01 06:16:13.992547: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 8593 of 10000
+2022-12-01 06:16:15.671612: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 5, total_step 91020, total loss is 24.45 , inference loss is 16.03, weight deacy loss is 8.41, training accuracy is 0.125000, time 205.601 samples/sec
+epoch 5, total_step 91040, total loss is 21.23 , inference loss is 12.81, weight deacy loss is 8.41, training accuracy is 0.218750, time 209.338 samples/sec
+epoch 5, total_step 91060, total loss is 20.86 , inference loss is 12.45, weight deacy loss is 8.41, training accuracy is 0.156250, time 205.911 samples/sec
+epoch 5, total_step 91080, total loss is 18.11 , inference loss is 9.70, weight deacy loss is 8.41, training accuracy is 0.281250, time 217.107 samples/sec
+epoch 5, total_step 91100, total loss is 21.63 , inference loss is 13.22, weight deacy loss is 8.41, training accuracy is 0.062500, time 216.131 samples/sec
+epoch 5, total_step 91120, total loss is 20.41 , inference loss is 12.00, weight deacy loss is 8.41, training accuracy is 0.281250, time 219.270 samples/sec
+epoch 5, total_step 91140, total loss is 19.74 , inference loss is 11.33, weight deacy loss is 8.41, training accuracy is 0.156250, time 215.558 samples/sec
+epoch 5, total_step 91160, total loss is 19.45 , inference loss is 11.04, weight deacy loss is 8.41, training accuracy is 0.093750, time 296.386 samples/sec
+epoch 5, total_step 91180, total loss is 19.17 , inference loss is 10.76, weight deacy loss is 8.41, training accuracy is 0.093750, time 218.078 samples/sec
+epoch 5, total_step 91200, total loss is 22.90 , inference loss is 14.49, weight deacy loss is 8.41, training accuracy is 0.281250, time 13.971 samples/sec
+epoch 5, total_step 91220, total loss is 21.53 , inference loss is 13.13, weight deacy loss is 8.41, training accuracy is 0.156250, time 209.485 samples/sec
+epoch 5, total_step 91240, total loss is 16.00 , inference loss is 7.59, weight deacy loss is 8.40, training accuracy is 0.281250, time 212.453 samples/sec
+epoch 5, total_step 91260, total loss is 18.01 , inference loss is 9.60, weight deacy loss is 8.40, training accuracy is 0.250000, time 216.147 samples/sec
+epoch 5, total_step 91280, total loss is 20.70 , inference loss is 12.30, weight deacy loss is 8.40, training accuracy is 0.250000, time 212.845 samples/sec
+epoch 5, total_step 91300, total loss is 21.72 , inference loss is 13.32, weight deacy loss is 8.40, training accuracy is 0.187500, time 210.467 samples/sec
+epoch 5, total_step 91320, total loss is 19.30 , inference loss is 10.90, weight deacy loss is 8.40, training accuracy is 0.187500, time 209.399 samples/sec
+epoch 5, total_step 91340, total loss is 19.32 , inference loss is 10.92, weight deacy loss is 8.40, training accuracy is 0.218750, time 207.347 samples/sec
+epoch 5, total_step 91360, total loss is 20.66 , inference loss is 12.26, weight deacy loss is 8.40, training accuracy is 0.093750, time 219.590 samples/sec
+epoch 5, total_step 91380, total loss is 21.11 , inference loss is 12.71, weight deacy loss is 8.40, training accuracy is 0.250000, time 199.367 samples/sec
+epoch 5, total_step 91400, total loss is 16.57 , inference loss is 8.17, weight deacy loss is 8.40, training accuracy is 0.125000, time 223.811 samples/sec
+epoch 5, total_step 91420, total loss is 18.76 , inference loss is 10.36, weight deacy loss is 8.40, training accuracy is 0.281250, time 299.067 samples/sec
+epoch 5, total_step 91440, total loss is 16.84 , inference loss is 8.44, weight deacy loss is 8.40, training accuracy is 0.343750, time 217.853 samples/sec
+epoch 5, total_step 91460, total loss is 19.22 , inference loss is 10.83, weight deacy loss is 8.40, training accuracy is 0.218750, time 222.325 samples/sec
+epoch 5, total_step 91480, total loss is 20.27 , inference loss is 11.87, weight deacy loss is 8.39, training accuracy is 0.125000, time 225.169 samples/sec
+epoch 5, total_step 91500, total loss is 21.62 , inference loss is 13.23, weight deacy loss is 8.39, training accuracy is 0.093750, time 14.588 samples/sec
+epoch 5, total_step 91520, total loss is 21.76 , inference loss is 13.37, weight deacy loss is 8.39, training accuracy is 0.156250, time 217.659 samples/sec
+epoch 5, total_step 91540, total loss is 21.90 , inference loss is 13.50, weight deacy loss is 8.39, training accuracy is 0.187500, time 301.097 samples/sec
+epoch 5, total_step 91560, total loss is 17.25 , inference loss is 8.86, weight deacy loss is 8.39, training accuracy is 0.312500, time 297.252 samples/sec
+epoch 5, total_step 91580, total loss is 20.60 , inference loss is 12.21, weight deacy loss is 8.39, training accuracy is 0.218750, time 214.444 samples/sec
+epoch 5, total_step 91600, total loss is 21.45 , inference loss is 13.06, weight deacy loss is 8.39, training accuracy is 0.125000, time 206.303 samples/sec
+epoch 5, total_step 91620, total loss is 23.57 , inference loss is 15.18, weight deacy loss is 8.39, training accuracy is 0.187500, time 212.402 samples/sec
+epoch 5, total_step 91640, total loss is 21.09 , inference loss is 12.70, weight deacy loss is 8.39, training accuracy is 0.093750, time 210.333 samples/sec
+epoch 5, total_step 91660, total loss is 22.79 , inference loss is 14.40, weight deacy loss is 8.39, training accuracy is 0.156250, time 214.310 samples/sec
+epoch 5, total_step 91680, total loss is 17.11 , inference loss is 8.73, weight deacy loss is 8.39, training accuracy is 0.343750, time 227.148 samples/sec
+epoch 5, total_step 91700, total loss is 20.67 , inference loss is 12.29, weight deacy loss is 8.39, training accuracy is 0.218750, time 194.629 samples/sec
+epoch 5, total_step 91720, total loss is 18.19 , inference loss is 9.81, weight deacy loss is 8.39, training accuracy is 0.250000, time 191.590 samples/sec
+epoch 5, total_step 91740, total loss is 18.28 , inference loss is 9.90, weight deacy loss is 8.38, training accuracy is 0.250000, time 179.114 samples/sec
+epoch 5, total_step 91760, total loss is 15.78 , inference loss is 7.40, weight deacy loss is 8.38, training accuracy is 0.218750, time 221.799 samples/sec
+epoch 5, total_step 91780, total loss is 19.98 , inference loss is 11.59, weight deacy loss is 8.38, training accuracy is 0.218750, time 182.077 samples/sec
+epoch 5, total_step 91800, total loss is 30.14 , inference loss is 21.76, weight deacy loss is 8.38, training accuracy is 0.062500, time 14.560 samples/sec
+epoch 5, total_step 91820, total loss is 21.75 , inference loss is 13.37, weight deacy loss is 8.38, training accuracy is 0.125000, time 205.571 samples/sec
+epoch 5, total_step 91840, total loss is 22.01 , inference loss is 13.63, weight deacy loss is 8.38, training accuracy is 0.093750, time 175.331 samples/sec
+epoch 5, total_step 91860, total loss is 22.38 , inference loss is 14.00, weight deacy loss is 8.38, training accuracy is 0.218750, time 203.463 samples/sec
+epoch 5, total_step 91880, total loss is 25.17 , inference loss is 16.79, weight deacy loss is 8.38, training accuracy is 0.187500, time 298.985 samples/sec
+epoch 5, total_step 91900, total loss is 24.00 , inference loss is 15.62, weight deacy loss is 8.38, training accuracy is 0.187500, time 204.228 samples/sec
+epoch 5, total_step 91920, total loss is 18.42 , inference loss is 10.05, weight deacy loss is 8.38, training accuracy is 0.343750, time 176.101 samples/sec
+epoch 5, total_step 91940, total loss is 18.89 , inference loss is 10.51, weight deacy loss is 8.38, training accuracy is 0.250000, time 176.260 samples/sec
+epoch 5, total_step 91960, total loss is 16.60 , inference loss is 8.23, weight deacy loss is 8.38, training accuracy is 0.250000, time 175.682 samples/sec
+epoch 5, total_step 91980, total loss is 23.01 , inference loss is 14.64, weight deacy loss is 8.37, training accuracy is 0.281250, time 212.634 samples/sec
+End of epoch 5
+testing verification..
+(12000, 512)
+infer time 23.317968
+best_threshold_index 123 0.9851851851851852
+best_threshold_index 122 0.9840740740740741
+best_threshold_index 122 0.9844444444444445
+best_threshold_index 122 0.985
+best_threshold_index 122 0.9844444444444445
+best_threshold_index 122 0.9831481481481481
+best_threshold_index 123 0.9846296296296296
+best_threshold_index 122 0.9837037037037037
+best_threshold_index 122 0.9829629629629629
+best_threshold_index 122 0.9829629629629629
+[lfw][92000]XNorm: 21.673197
+[lfw][92000]Accuracy-Flip: 0.98350+-0.00705
+test accuracy is: 0.9835
+epoch 6, total_step 92000, total loss is 18.27 , inference loss is 9.90, weight deacy loss is 8.37, training accuracy is 0.218750, time 16.992 samples/sec
+epoch 6, total_step 92020, total loss is 20.89 , inference loss is 12.51, weight deacy loss is 8.37, training accuracy is 0.250000, time 208.507 samples/sec
+epoch 6, total_step 92040, total loss is 15.56 , inference loss is 7.18, weight deacy loss is 8.37, training accuracy is 0.406250, time 214.266 samples/sec
+epoch 6, total_step 92060, total loss is 17.86 , inference loss is 9.49, weight deacy loss is 8.37, training accuracy is 0.343750, time 202.888 samples/sec
+epoch 6, total_step 92080, total loss is 14.38 , inference loss is 6.01, weight deacy loss is 8.37, training accuracy is 0.375000, time 207.076 samples/sec
+epoch 6, total_step 92100, total loss is 10.38 , inference loss is 2.01, weight deacy loss is 8.37, training accuracy is 0.781250, time 12.464 samples/sec
+epoch 6, total_step 92120, total loss is 13.17 , inference loss is 4.80, weight deacy loss is 8.37, training accuracy is 0.593750, time 213.188 samples/sec
+epoch 6, total_step 92140, total loss is 17.37 , inference loss is 9.00, weight deacy loss is 8.37, training accuracy is 0.562500, time 217.252 samples/sec
+epoch 6, total_step 92160, total loss is 12.52 , inference loss is 4.15, weight deacy loss is 8.37, training accuracy is 0.593750, time 210.493 samples/sec
+epoch 6, total_step 92180, total loss is 15.77 , inference loss is 7.40, weight deacy loss is 8.37, training accuracy is 0.500000, time 208.020 samples/sec
+epoch 6, total_step 92200, total loss is 15.56 , inference loss is 7.19, weight deacy loss is 8.37, training accuracy is 0.468750, time 297.960 samples/sec
+epoch 6, total_step 92220, total loss is 19.41 , inference loss is 11.05, weight deacy loss is 8.36, training accuracy is 0.468750, time 219.697 samples/sec
+epoch 6, total_step 92240, total loss is 13.75 , inference loss is 5.39, weight deacy loss is 8.36, training accuracy is 0.500000, time 218.873 samples/sec
+epoch 6, total_step 92260, total loss is 17.22 , inference loss is 8.86, weight deacy loss is 8.36, training accuracy is 0.437500, time 207.699 samples/sec
+epoch 6, total_step 92280, total loss is 14.04 , inference loss is 5.68, weight deacy loss is 8.36, training accuracy is 0.437500, time 194.696 samples/sec
+epoch 6, total_step 92300, total loss is 18.53 , inference loss is 10.17, weight deacy loss is 8.36, training accuracy is 0.437500, time 208.680 samples/sec
+epoch 6, total_step 92320, total loss is 16.39 , inference loss is 8.03, weight deacy loss is 8.36, training accuracy is 0.468750, time 217.637 samples/sec
+epoch 6, total_step 92340, total loss is 15.19 , inference loss is 6.83, weight deacy loss is 8.36, training accuracy is 0.625000, time 234.202 samples/sec
+epoch 6, total_step 92360, total loss is 15.76 , inference loss is 7.40, weight deacy loss is 8.36, training accuracy is 0.375000, time 202.237 samples/sec
+epoch 6, total_step 92380, total loss is 15.79 , inference loss is 7.43, weight deacy loss is 8.36, training accuracy is 0.437500, time 206.831 samples/sec
+epoch 6, total_step 92400, total loss is 15.38 , inference loss is 7.02, weight deacy loss is 8.36, training accuracy is 0.375000, time 13.372 samples/sec
+epoch 6, total_step 92420, total loss is 13.91 , inference loss is 5.55, weight deacy loss is 8.36, training accuracy is 0.562500, time 213.168 samples/sec
+epoch 6, total_step 92440, total loss is 15.50 , inference loss is 7.14, weight deacy loss is 8.36, training accuracy is 0.500000, time 232.380 samples/sec
+epoch 6, total_step 92460, total loss is 18.43 , inference loss is 10.07, weight deacy loss is 8.35, training accuracy is 0.250000, time 213.724 samples/sec
+epoch 6, total_step 92480, total loss is 18.68 , inference loss is 10.32, weight deacy loss is 8.35, training accuracy is 0.437500, time 299.219 samples/sec
+epoch 6, total_step 92500, total loss is 16.99 , inference loss is 8.64, weight deacy loss is 8.35, training accuracy is 0.437500, time 212.672 samples/sec
+epoch 6, total_step 92520, total loss is 16.20 , inference loss is 7.85, weight deacy loss is 8.35, training accuracy is 0.406250, time 221.096 samples/sec
+epoch 6, total_step 92540, total loss is 15.53 , inference loss is 7.18, weight deacy loss is 8.35, training accuracy is 0.406250, time 208.219 samples/sec
+epoch 6, total_step 92560, total loss is 17.19 , inference loss is 8.84, weight deacy loss is 8.35, training accuracy is 0.375000, time 237.731 samples/sec
+epoch 6, total_step 92580, total loss is 22.18 , inference loss is 13.83, weight deacy loss is 8.35, training accuracy is 0.343750, time 213.718 samples/sec
+epoch 6, total_step 92600, total loss is 24.09 , inference loss is 15.74, weight deacy loss is 8.35, training accuracy is 0.281250, time 219.523 samples/sec
+epoch 6, total_step 92620, total loss is 21.08 , inference loss is 12.73, weight deacy loss is 8.35, training accuracy is 0.312500, time 212.804 samples/sec
+epoch 6, total_step 92640, total loss is 19.76 , inference loss is 11.42, weight deacy loss is 8.35, training accuracy is 0.375000, time 215.542 samples/sec
+epoch 6, total_step 92660, total loss is 15.22 , inference loss is 6.88, weight deacy loss is 8.35, training accuracy is 0.468750, time 213.784 samples/sec
+epoch 6, total_step 92680, total loss is 13.52 , inference loss is 5.18, weight deacy loss is 8.35, training accuracy is 0.531250, time 203.105 samples/sec
+epoch 6, total_step 92700, total loss is 18.47 , inference loss is 10.12, weight deacy loss is 8.35, training accuracy is 0.406250, time 13.215 samples/sec
+epoch 6, total_step 92720, total loss is 22.87 , inference loss is 14.53, weight deacy loss is 8.35, training accuracy is 0.312500, time 226.977 samples/sec
+epoch 6, total_step 92740, total loss is 16.47 , inference loss is 8.12, weight deacy loss is 8.34, training accuracy is 0.531250, time 298.019 samples/sec
+epoch 6, total_step 92760, total loss is 17.73 , inference loss is 9.38, weight deacy loss is 8.34, training accuracy is 0.437500, time 212.040 samples/sec
+epoch 6, total_step 92780, total loss is 17.94 , inference loss is 9.60, weight deacy loss is 8.34, training accuracy is 0.437500, time 219.376 samples/sec
+epoch 6, total_step 92800, total loss is 21.34 , inference loss is 12.99, weight deacy loss is 8.34, training accuracy is 0.343750, time 215.338 samples/sec
+epoch 6, total_step 92820, total loss is 22.07 , inference loss is 13.73, weight deacy loss is 8.34, training accuracy is 0.343750, time 217.680 samples/sec
+epoch 6, total_step 92840, total loss is 14.56 , inference loss is 6.22, weight deacy loss is 8.34, training accuracy is 0.562500, time 213.595 samples/sec
+epoch 6, total_step 92860, total loss is 20.51 , inference loss is 12.16, weight deacy loss is 8.34, training accuracy is 0.343750, time 212.299 samples/sec
+epoch 6, total_step 92880, total loss is 17.04 , inference loss is 8.70, weight deacy loss is 8.34, training accuracy is 0.312500, time 218.715 samples/sec
+epoch 6, total_step 92900, total loss is 22.27 , inference loss is 13.93, weight deacy loss is 8.34, training accuracy is 0.312500, time 202.011 samples/sec
+epoch 6, total_step 92920, total loss is 19.68 , inference loss is 11.34, weight deacy loss is 8.34, training accuracy is 0.312500, time 211.540 samples/sec
+epoch 6, total_step 92940, total loss is 16.12 , inference loss is 7.79, weight deacy loss is 8.34, training accuracy is 0.375000, time 212.498 samples/sec
+epoch 6, total_step 92960, total loss is 21.33 , inference loss is 12.99, weight deacy loss is 8.34, training accuracy is 0.312500, time 204.291 samples/sec
+epoch 6, total_step 92980, total loss is 15.93 , inference loss is 7.59, weight deacy loss is 8.34, training accuracy is 0.562500, time 214.647 samples/sec
+epoch 6, total_step 93000, total loss is 19.93 , inference loss is 11.59, weight deacy loss is 8.34, training accuracy is 0.437500, time 13.100 samples/sec
+epoch 6, total_step 93020, total loss is 17.20 , inference loss is 8.86, weight deacy loss is 8.33, training accuracy is 0.343750, time 213.932 samples/sec
+epoch 6, total_step 93040, total loss is 20.65 , inference loss is 12.31, weight deacy loss is 8.33, training accuracy is 0.437500, time 191.943 samples/sec
+epoch 6, total_step 93060, total loss is 15.56 , inference loss is 7.23, weight deacy loss is 8.33, training accuracy is 0.375000, time 208.279 samples/sec
+epoch 6, total_step 93080, total loss is 20.68 , inference loss is 12.35, weight deacy loss is 8.33, training accuracy is 0.437500, time 210.173 samples/sec
+epoch 6, total_step 93100, total loss is 21.78 , inference loss is 13.45, weight deacy loss is 8.33, training accuracy is 0.218750, time 207.901 samples/sec
+epoch 6, total_step 93120, total loss is 16.96 , inference loss is 8.63, weight deacy loss is 8.33, training accuracy is 0.375000, time 207.233 samples/sec
+epoch 6, total_step 93140, total loss is 17.14 , inference loss is 8.81, weight deacy loss is 8.33, training accuracy is 0.468750, time 205.303 samples/sec
+epoch 6, total_step 93160, total loss is 18.39 , inference loss is 10.06, weight deacy loss is 8.33, training accuracy is 0.281250, time 212.331 samples/sec
+epoch 6, total_step 93180, total loss is 13.84 , inference loss is 5.51, weight deacy loss is 8.33, training accuracy is 0.562500, time 208.707 samples/sec
+epoch 6, total_step 93200, total loss is 18.72 , inference loss is 10.39, weight deacy loss is 8.33, training accuracy is 0.343750, time 204.914 samples/sec
+epoch 6, total_step 93220, total loss is 15.26 , inference loss is 6.93, weight deacy loss is 8.33, training accuracy is 0.468750, time 295.840 samples/sec
+epoch 6, total_step 93240, total loss is 17.70 , inference loss is 9.37, weight deacy loss is 8.33, training accuracy is 0.406250, time 296.286 samples/sec
+epoch 6, total_step 93260, total loss is 15.82 , inference loss is 7.50, weight deacy loss is 8.33, training accuracy is 0.343750, time 222.325 samples/sec
+epoch 6, total_step 93280, total loss is 15.19 , inference loss is 6.86, weight deacy loss is 8.33, training accuracy is 0.531250, time 218.900 samples/sec
+epoch 6, total_step 93300, total loss is 16.27 , inference loss is 7.94, weight deacy loss is 8.32, training accuracy is 0.375000, time 14.448 samples/sec
+epoch 6, total_step 93320, total loss is 17.74 , inference loss is 9.42, weight deacy loss is 8.32, training accuracy is 0.406250, time 232.791 samples/sec
+epoch 6, total_step 93340, total loss is 17.35 , inference loss is 9.02, weight deacy loss is 8.32, training accuracy is 0.406250, time 298.311 samples/sec
+epoch 6, total_step 93360, total loss is 26.75 , inference loss is 18.42, weight deacy loss is 8.32, training accuracy is 0.312500, time 299.490 samples/sec
+epoch 6, total_step 93380, total loss is 17.65 , inference loss is 9.33, weight deacy loss is 8.32, training accuracy is 0.312500, time 296.981 samples/sec
+epoch 6, total_step 93400, total loss is 16.57 , inference loss is 8.25, weight deacy loss is 8.32, training accuracy is 0.437500, time 215.451 samples/sec
+epoch 6, total_step 93420, total loss is 18.27 , inference loss is 9.95, weight deacy loss is 8.32, training accuracy is 0.406250, time 203.722 samples/sec
+epoch 6, total_step 93440, total loss is 22.66 , inference loss is 14.34, weight deacy loss is 8.32, training accuracy is 0.218750, time 208.573 samples/sec
+epoch 6, total_step 93460, total loss is 18.48 , inference loss is 10.17, weight deacy loss is 8.32, training accuracy is 0.437500, time 206.274 samples/sec
+epoch 6, total_step 93480, total loss is 14.07 , inference loss is 5.75, weight deacy loss is 8.32, training accuracy is 0.468750, time 204.086 samples/sec
+epoch 6, total_step 93500, total loss is 16.26 , inference loss is 7.94, weight deacy loss is 8.32, training accuracy is 0.375000, time 300.759 samples/sec
+epoch 6, total_step 93520, total loss is 15.90 , inference loss is 7.59, weight deacy loss is 8.32, training accuracy is 0.375000, time 219.025 samples/sec
+epoch 6, total_step 93540, total loss is 20.09 , inference loss is 11.78, weight deacy loss is 8.32, training accuracy is 0.437500, time 215.369 samples/sec
+epoch 6, total_step 93560, total loss is 17.06 , inference loss is 8.75, weight deacy loss is 8.32, training accuracy is 0.312500, time 212.343 samples/sec
+epoch 6, total_step 93580, total loss is 16.74 , inference loss is 8.43, weight deacy loss is 8.32, training accuracy is 0.375000, time 204.864 samples/sec
+epoch 6, total_step 93600, total loss is 18.00 , inference loss is 9.69, weight deacy loss is 8.31, training accuracy is 0.312500, time 13.206 samples/sec
+epoch 6, total_step 93620, total loss is 21.92 , inference loss is 13.60, weight deacy loss is 8.31, training accuracy is 0.250000, time 211.858 samples/sec
+epoch 6, total_step 93640, total loss is 15.93 , inference loss is 7.61, weight deacy loss is 8.31, training accuracy is 0.375000, time 200.714 samples/sec
+epoch 6, total_step 93660, total loss is 18.59 , inference loss is 10.28, weight deacy loss is 8.31, training accuracy is 0.437500, time 208.420 samples/sec
+epoch 6, total_step 93680, total loss is 17.97 , inference loss is 9.66, weight deacy loss is 8.31, training accuracy is 0.375000, time 214.653 samples/sec
+epoch 6, total_step 93700, total loss is 20.12 , inference loss is 11.81, weight deacy loss is 8.31, training accuracy is 0.375000, time 211.503 samples/sec
+epoch 6, total_step 93720, total loss is 16.28 , inference loss is 7.97, weight deacy loss is 8.31, training accuracy is 0.406250, time 212.872 samples/sec
+epoch 6, total_step 93740, total loss is 22.15 , inference loss is 13.84, weight deacy loss is 8.31, training accuracy is 0.281250, time 217.500 samples/sec
+epoch 6, total_step 93760, total loss is 19.79 , inference loss is 11.49, weight deacy loss is 8.31, training accuracy is 0.343750, time 298.590 samples/sec
+epoch 6, total_step 93780, total loss is 17.41 , inference loss is 9.10, weight deacy loss is 8.31, training accuracy is 0.375000, time 204.416 samples/sec
+epoch 6, total_step 93800, total loss is 21.94 , inference loss is 13.63, weight deacy loss is 8.31, training accuracy is 0.218750, time 211.508 samples/sec
+epoch 6, total_step 93820, total loss is 17.47 , inference loss is 9.16, weight deacy loss is 8.31, training accuracy is 0.468750, time 211.489 samples/sec
+epoch 6, total_step 93840, total loss is 19.74 , inference loss is 11.44, weight deacy loss is 8.31, training accuracy is 0.281250, time 206.206 samples/sec
+epoch 6, total_step 93860, total loss is 17.73 , inference loss is 9.42, weight deacy loss is 8.31, training accuracy is 0.500000, time 210.191 samples/sec
+epoch 6, total_step 93880, total loss is 17.48 , inference loss is 9.18, weight deacy loss is 8.30, training accuracy is 0.375000, time 210.268 samples/sec
+epoch 6, total_step 93900, total loss is 19.74 , inference loss is 11.43, weight deacy loss is 8.30, training accuracy is 0.312500, time 13.093 samples/sec
+epoch 6, total_step 93920, total loss is 23.22 , inference loss is 14.92, weight deacy loss is 8.30, training accuracy is 0.218750, time 214.118 samples/sec
+epoch 6, total_step 93940, total loss is 22.24 , inference loss is 13.94, weight deacy loss is 8.30, training accuracy is 0.218750, time 212.386 samples/sec
+epoch 6, total_step 93960, total loss is 14.19 , inference loss is 5.89, weight deacy loss is 8.30, training accuracy is 0.500000, time 207.843 samples/sec
+epoch 6, total_step 93980, total loss is 18.11 , inference loss is 9.81, weight deacy loss is 8.30, training accuracy is 0.437500, time 210.539 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.04129800000001
+best_threshold_index 113 0.9872222222222222
+best_threshold_index 113 0.9857407407407407
+best_threshold_index 113 0.9868518518518519
+best_threshold_index 113 0.987037037037037
+best_threshold_index 113 0.9872222222222222
+best_threshold_index 113 0.9855555555555555
+best_threshold_index 113 0.9866666666666667
+best_threshold_index 113 0.9857407407407407
+best_threshold_index 113 0.9857407407407407
+best_threshold_index 113 0.9855555555555555
+[lfw][94000]XNorm: 23.176093
+[lfw][94000]Accuracy-Flip: 0.98633+-0.00618
+test accuracy is: 0.9863333333333335
+epoch 6, total_step 94000, total loss is 19.87 , inference loss is 11.57, weight deacy loss is 8.30, training accuracy is 0.406250, time 14.693 samples/sec
+epoch 6, total_step 94020, total loss is 25.36 , inference loss is 17.06, weight deacy loss is 8.30, training accuracy is 0.187500, time 211.918 samples/sec
+epoch 6, total_step 94040, total loss is 13.85 , inference loss is 5.55, weight deacy loss is 8.30, training accuracy is 0.437500, time 209.388 samples/sec
+epoch 6, total_step 94060, total loss is 18.81 , inference loss is 10.51, weight deacy loss is 8.30, training accuracy is 0.250000, time 215.815 samples/sec
+epoch 6, total_step 94080, total loss is 13.38 , inference loss is 5.08, weight deacy loss is 8.30, training accuracy is 0.468750, time 210.717 samples/sec
+epoch 6, total_step 94100, total loss is 19.14 , inference loss is 10.84, weight deacy loss is 8.30, training accuracy is 0.312500, time 213.681 samples/sec
+epoch 6, total_step 94120, total loss is 15.76 , inference loss is 7.47, weight deacy loss is 8.29, training accuracy is 0.500000, time 224.964 samples/sec
+epoch 6, total_step 94140, total loss is 20.79 , inference loss is 12.50, weight deacy loss is 8.29, training accuracy is 0.281250, time 224.271 samples/sec
+epoch 6, total_step 94160, total loss is 13.88 , inference loss is 5.59, weight deacy loss is 8.29, training accuracy is 0.468750, time 211.696 samples/sec
+epoch 6, total_step 94180, total loss is 18.62 , inference loss is 10.33, weight deacy loss is 8.29, training accuracy is 0.406250, time 212.817 samples/sec
+epoch 6, total_step 94200, total loss is 15.89 , inference loss is 7.60, weight deacy loss is 8.29, training accuracy is 0.437500, time 13.837 samples/sec
+epoch 6, total_step 94220, total loss is 15.93 , inference loss is 7.64, weight deacy loss is 8.29, training accuracy is 0.562500, time 197.729 samples/sec
+epoch 6, total_step 94240, total loss is 13.91 , inference loss is 5.62, weight deacy loss is 8.29, training accuracy is 0.437500, time 203.699 samples/sec
+epoch 6, total_step 94260, total loss is 11.85 , inference loss is 3.56, weight deacy loss is 8.29, training accuracy is 0.593750, time 221.097 samples/sec
+epoch 6, total_step 94280, total loss is 16.86 , inference loss is 8.57, weight deacy loss is 8.29, training accuracy is 0.593750, time 215.255 samples/sec
+epoch 6, total_step 94300, total loss is 14.96 , inference loss is 6.67, weight deacy loss is 8.29, training accuracy is 0.375000, time 207.339 samples/sec
+epoch 6, total_step 94320, total loss is 17.27 , inference loss is 8.99, weight deacy loss is 8.29, training accuracy is 0.281250, time 215.484 samples/sec
+epoch 6, total_step 94340, total loss is 18.14 , inference loss is 9.85, weight deacy loss is 8.29, training accuracy is 0.406250, time 207.538 samples/sec
+epoch 6, total_step 94360, total loss is 17.21 , inference loss is 8.93, weight deacy loss is 8.28, training accuracy is 0.343750, time 204.193 samples/sec
+epoch 6, total_step 94380, total loss is 16.67 , inference loss is 8.39, weight deacy loss is 8.28, training accuracy is 0.437500, time 296.430 samples/sec
+epoch 6, total_step 94400, total loss is 16.43 , inference loss is 8.14, weight deacy loss is 8.28, training accuracy is 0.437500, time 211.456 samples/sec
+epoch 6, total_step 94420, total loss is 16.52 , inference loss is 8.23, weight deacy loss is 8.28, training accuracy is 0.531250, time 224.399 samples/sec
+epoch 6, total_step 94440, total loss is 12.00 , inference loss is 3.72, weight deacy loss is 8.28, training accuracy is 0.625000, time 205.263 samples/sec
+epoch 6, total_step 94460, total loss is 12.76 , inference loss is 4.48, weight deacy loss is 8.28, training accuracy is 0.625000, time 208.705 samples/sec
+epoch 6, total_step 94480, total loss is 16.75 , inference loss is 8.47, weight deacy loss is 8.28, training accuracy is 0.375000, time 208.963 samples/sec
+epoch 6, total_step 94500, total loss is 18.18 , inference loss is 9.90, weight deacy loss is 8.28, training accuracy is 0.375000, time 12.695 samples/sec
+epoch 6, total_step 94520, total loss is 22.29 , inference loss is 14.01, weight deacy loss is 8.28, training accuracy is 0.312500, time 206.807 samples/sec
+epoch 6, total_step 94540, total loss is 14.18 , inference loss is 5.91, weight deacy loss is 8.28, training accuracy is 0.468750, time 212.681 samples/sec
+epoch 6, total_step 94560, total loss is 17.79 , inference loss is 9.51, weight deacy loss is 8.28, training accuracy is 0.500000, time 188.310 samples/sec
+epoch 6, total_step 94580, total loss is 17.90 , inference loss is 9.62, weight deacy loss is 8.28, training accuracy is 0.187500, time 210.421 samples/sec
+epoch 6, total_step 94600, total loss is 13.36 , inference loss is 5.09, weight deacy loss is 8.27, training accuracy is 0.375000, time 213.273 samples/sec
+epoch 6, total_step 94620, total loss is 14.21 , inference loss is 5.94, weight deacy loss is 8.27, training accuracy is 0.343750, time 209.302 samples/sec
+epoch 6, total_step 94640, total loss is 11.59 , inference loss is 3.31, weight deacy loss is 8.27, training accuracy is 0.593750, time 295.523 samples/sec
+epoch 6, total_step 94660, total loss is 19.06 , inference loss is 10.78, weight deacy loss is 8.27, training accuracy is 0.406250, time 220.823 samples/sec
+epoch 6, total_step 94680, total loss is 18.85 , inference loss is 10.58, weight deacy loss is 8.27, training accuracy is 0.375000, time 212.753 samples/sec
+epoch 6, total_step 94700, total loss is 16.46 , inference loss is 8.19, weight deacy loss is 8.27, training accuracy is 0.531250, time 230.034 samples/sec
+epoch 6, total_step 94720, total loss is 13.76 , inference loss is 5.49, weight deacy loss is 8.27, training accuracy is 0.437500, time 200.661 samples/sec
+epoch 6, total_step 94740, total loss is 15.05 , inference loss is 6.78, weight deacy loss is 8.27, training accuracy is 0.468750, time 213.860 samples/sec
+epoch 6, total_step 94760, total loss is 15.33 , inference loss is 7.07, weight deacy loss is 8.27, training accuracy is 0.562500, time 210.298 samples/sec
+epoch 6, total_step 94780, total loss is 18.80 , inference loss is 10.53, weight deacy loss is 8.27, training accuracy is 0.437500, time 210.835 samples/sec
+epoch 6, total_step 94800, total loss is 15.31 , inference loss is 7.04, weight deacy loss is 8.27, training accuracy is 0.500000, time 11.214 samples/sec
+epoch 6, total_step 94820, total loss is 17.63 , inference loss is 9.37, weight deacy loss is 8.27, training accuracy is 0.468750, time 210.725 samples/sec
+epoch 6, total_step 94840, total loss is 14.38 , inference loss is 6.12, weight deacy loss is 8.27, training accuracy is 0.562500, time 216.959 samples/sec
+epoch 6, total_step 94860, total loss is 16.89 , inference loss is 8.62, weight deacy loss is 8.26, training accuracy is 0.468750, time 207.806 samples/sec
+epoch 6, total_step 94880, total loss is 12.99 , inference loss is 4.73, weight deacy loss is 8.26, training accuracy is 0.406250, time 210.075 samples/sec
+epoch 6, total_step 94900, total loss is 16.66 , inference loss is 8.40, weight deacy loss is 8.26, training accuracy is 0.437500, time 298.072 samples/sec
+epoch 6, total_step 94920, total loss is 16.97 , inference loss is 8.70, weight deacy loss is 8.26, training accuracy is 0.437500, time 217.589 samples/sec
+epoch 6, total_step 94940, total loss is 13.82 , inference loss is 5.56, weight deacy loss is 8.26, training accuracy is 0.593750, time 209.758 samples/sec
+epoch 6, total_step 94960, total loss is 15.84 , inference loss is 7.58, weight deacy loss is 8.26, training accuracy is 0.375000, time 210.088 samples/sec
+epoch 6, total_step 94980, total loss is 17.70 , inference loss is 9.44, weight deacy loss is 8.26, training accuracy is 0.343750, time 216.920 samples/sec
+epoch 6, total_step 95000, total loss is 16.66 , inference loss is 8.40, weight deacy loss is 8.26, training accuracy is 0.531250, time 214.864 samples/sec
+epoch 6, total_step 95020, total loss is 19.32 , inference loss is 11.06, weight deacy loss is 8.26, training accuracy is 0.343750, time 213.644 samples/sec
+epoch 6, total_step 95040, total loss is 14.23 , inference loss is 5.98, weight deacy loss is 8.26, training accuracy is 0.437500, time 217.393 samples/sec
+epoch 6, total_step 95060, total loss is 21.00 , inference loss is 12.74, weight deacy loss is 8.26, training accuracy is 0.375000, time 219.460 samples/sec
+epoch 6, total_step 95080, total loss is 17.82 , inference loss is 9.56, weight deacy loss is 8.26, training accuracy is 0.406250, time 212.886 samples/sec
+epoch 6, total_step 95100, total loss is 11.37 , inference loss is 3.12, weight deacy loss is 8.26, training accuracy is 0.562500, time 14.461 samples/sec
+epoch 6, total_step 95120, total loss is 19.35 , inference loss is 11.09, weight deacy loss is 8.25, training accuracy is 0.281250, time 295.067 samples/sec
+epoch 6, total_step 95140, total loss is 20.00 , inference loss is 11.75, weight deacy loss is 8.25, training accuracy is 0.312500, time 217.292 samples/sec
+epoch 6, total_step 95160, total loss is 16.57 , inference loss is 8.32, weight deacy loss is 8.25, training accuracy is 0.312500, time 204.534 samples/sec
+epoch 6, total_step 95180, total loss is 18.28 , inference loss is 10.02, weight deacy loss is 8.25, training accuracy is 0.437500, time 206.943 samples/sec
+epoch 6, total_step 95200, total loss is 12.65 , inference loss is 4.40, weight deacy loss is 8.25, training accuracy is 0.468750, time 212.861 samples/sec
+epoch 6, total_step 95220, total loss is 19.06 , inference loss is 10.81, weight deacy loss is 8.25, training accuracy is 0.375000, time 217.345 samples/sec
+epoch 6, total_step 95240, total loss is 15.82 , inference loss is 7.57, weight deacy loss is 8.25, training accuracy is 0.437500, time 214.912 samples/sec
+epoch 6, total_step 95260, total loss is 19.32 , inference loss is 11.07, weight deacy loss is 8.25, training accuracy is 0.375000, time 208.607 samples/sec
+epoch 6, total_step 95280, total loss is 18.64 , inference loss is 10.40, weight deacy loss is 8.25, training accuracy is 0.406250, time 214.722 samples/sec
+epoch 6, total_step 95300, total loss is 20.15 , inference loss is 11.90, weight deacy loss is 8.25, training accuracy is 0.375000, time 208.835 samples/sec
+epoch 6, total_step 95320, total loss is 19.18 , inference loss is 10.94, weight deacy loss is 8.25, training accuracy is 0.343750, time 219.697 samples/sec
+epoch 6, total_step 95340, total loss is 14.61 , inference loss is 6.37, weight deacy loss is 8.25, training accuracy is 0.281250, time 297.484 samples/sec
+epoch 6, total_step 95360, total loss is 14.57 , inference loss is 6.32, weight deacy loss is 8.25, training accuracy is 0.312500, time 207.392 samples/sec
+epoch 6, total_step 95380, total loss is 16.24 , inference loss is 8.00, weight deacy loss is 8.24, training accuracy is 0.375000, time 210.298 samples/sec
+epoch 6, total_step 95400, total loss is 19.67 , inference loss is 11.43, weight deacy loss is 8.24, training accuracy is 0.281250, time 13.939 samples/sec
+epoch 6, total_step 95420, total loss is 19.16 , inference loss is 10.92, weight deacy loss is 8.24, training accuracy is 0.312500, time 211.238 samples/sec
+epoch 6, total_step 95440, total loss is 17.17 , inference loss is 8.93, weight deacy loss is 8.24, training accuracy is 0.343750, time 211.353 samples/sec
+epoch 6, total_step 95460, total loss is 15.54 , inference loss is 7.29, weight deacy loss is 8.24, training accuracy is 0.375000, time 210.758 samples/sec
+epoch 6, total_step 95480, total loss is 14.42 , inference loss is 6.18, weight deacy loss is 8.24, training accuracy is 0.500000, time 212.536 samples/sec
+epoch 6, total_step 95500, total loss is 17.55 , inference loss is 9.31, weight deacy loss is 8.24, training accuracy is 0.312500, time 205.894 samples/sec
+epoch 6, total_step 95520, total loss is 17.69 , inference loss is 9.45, weight deacy loss is 8.24, training accuracy is 0.343750, time 212.559 samples/sec
+epoch 6, total_step 95540, total loss is 19.86 , inference loss is 11.62, weight deacy loss is 8.24, training accuracy is 0.312500, time 210.454 samples/sec
+epoch 6, total_step 95560, total loss is 18.40 , inference loss is 10.16, weight deacy loss is 8.24, training accuracy is 0.218750, time 210.594 samples/sec
+epoch 6, total_step 95580, total loss is 16.59 , inference loss is 8.35, weight deacy loss is 8.24, training accuracy is 0.375000, time 209.541 samples/sec
+epoch 6, total_step 95600, total loss is 17.90 , inference loss is 9.67, weight deacy loss is 8.24, training accuracy is 0.406250, time 296.195 samples/sec
+epoch 6, total_step 95620, total loss is 17.17 , inference loss is 8.93, weight deacy loss is 8.24, training accuracy is 0.468750, time 209.926 samples/sec
+epoch 6, total_step 95640, total loss is 17.25 , inference loss is 9.01, weight deacy loss is 8.24, training accuracy is 0.375000, time 225.849 samples/sec
+epoch 6, total_step 95660, total loss is 18.02 , inference loss is 9.78, weight deacy loss is 8.23, training accuracy is 0.343750, time 213.323 samples/sec
+epoch 6, total_step 95680, total loss is 16.39 , inference loss is 8.16, weight deacy loss is 8.23, training accuracy is 0.406250, time 216.869 samples/sec
+epoch 6, total_step 95700, total loss is 20.34 , inference loss is 12.10, weight deacy loss is 8.23, training accuracy is 0.281250, time 13.485 samples/sec
+epoch 6, total_step 95720, total loss is 16.48 , inference loss is 8.24, weight deacy loss is 8.23, training accuracy is 0.375000, time 231.289 samples/sec
+epoch 6, total_step 95740, total loss is 20.63 , inference loss is 12.40, weight deacy loss is 8.23, training accuracy is 0.312500, time 215.969 samples/sec
+epoch 6, total_step 95760, total loss is 16.94 , inference loss is 8.71, weight deacy loss is 8.23, training accuracy is 0.375000, time 212.887 samples/sec
+epoch 6, total_step 95780, total loss is 16.05 , inference loss is 7.82, weight deacy loss is 8.23, training accuracy is 0.281250, time 212.343 samples/sec
+epoch 6, total_step 95800, total loss is 19.12 , inference loss is 10.89, weight deacy loss is 8.23, training accuracy is 0.343750, time 216.803 samples/sec
+epoch 6, total_step 95820, total loss is 16.84 , inference loss is 8.61, weight deacy loss is 8.23, training accuracy is 0.281250, time 212.217 samples/sec
+epoch 6, total_step 95840, total loss is 23.05 , inference loss is 14.83, weight deacy loss is 8.23, training accuracy is 0.218750, time 221.947 samples/sec
+epoch 6, total_step 95860, total loss is 19.38 , inference loss is 11.15, weight deacy loss is 8.23, training accuracy is 0.343750, time 296.642 samples/sec
+epoch 6, total_step 95880, total loss is 23.01 , inference loss is 14.78, weight deacy loss is 8.23, training accuracy is 0.250000, time 207.991 samples/sec
+epoch 6, total_step 95900, total loss is 17.12 , inference loss is 8.89, weight deacy loss is 8.23, training accuracy is 0.468750, time 205.818 samples/sec
+epoch 6, total_step 95920, total loss is 20.08 , inference loss is 11.85, weight deacy loss is 8.23, training accuracy is 0.343750, time 202.340 samples/sec
+epoch 6, total_step 95940, total loss is 20.95 , inference loss is 12.73, weight deacy loss is 8.22, training accuracy is 0.312500, time 206.112 samples/sec
+epoch 6, total_step 95960, total loss is 19.45 , inference loss is 11.23, weight deacy loss is 8.22, training accuracy is 0.437500, time 201.494 samples/sec
+epoch 6, total_step 95980, total loss is 21.45 , inference loss is 13.22, weight deacy loss is 8.22, training accuracy is 0.343750, time 211.970 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.31812000000001
+best_threshold_index 119 0.9874074074074074
+best_threshold_index 119 0.9862962962962963
+best_threshold_index 119 0.9872222222222222
+best_threshold_index 119 0.987037037037037
+best_threshold_index 119 0.9875925925925926
+best_threshold_index 121 0.9857407407407407
+best_threshold_index 120 0.9866666666666667
+best_threshold_index 119 0.9861111111111112
+best_threshold_index 119 0.9853703703703703
+best_threshold_index 120 0.9861111111111112
+[lfw][96000]XNorm: 21.803750
+[lfw][96000]Accuracy-Flip: 0.98600+-0.00638
+test accuracy is: 0.9860000000000001
+epoch 6, total_step 96000, total loss is 19.38 , inference loss is 11.16, weight deacy loss is 8.22, training accuracy is 0.312500, time 15.945 samples/sec
+epoch 6, total_step 96020, total loss is 21.28 , inference loss is 13.05, weight deacy loss is 8.22, training accuracy is 0.281250, time 220.825 samples/sec
+epoch 6, total_step 96040, total loss is 15.78 , inference loss is 7.56, weight deacy loss is 8.22, training accuracy is 0.312500, time 214.241 samples/sec
+epoch 6, total_step 96060, total loss is 17.27 , inference loss is 9.05, weight deacy loss is 8.22, training accuracy is 0.312500, time 209.427 samples/sec
+epoch 6, total_step 96080, total loss is 21.01 , inference loss is 12.79, weight deacy loss is 8.22, training accuracy is 0.343750, time 206.624 samples/sec
+epoch 6, total_step 96100, total loss is 23.14 , inference loss is 14.92, weight deacy loss is 8.22, training accuracy is 0.250000, time 223.619 samples/sec
+epoch 6, total_step 96120, total loss is 22.74 , inference loss is 14.52, weight deacy loss is 8.22, training accuracy is 0.375000, time 211.557 samples/sec
+epoch 6, total_step 96140, total loss is 18.53 , inference loss is 10.31, weight deacy loss is 8.22, training accuracy is 0.343750, time 215.372 samples/sec
+epoch 6, total_step 96160, total loss is 15.89 , inference loss is 7.67, weight deacy loss is 8.22, training accuracy is 0.406250, time 219.626 samples/sec
+epoch 6, total_step 96180, total loss is 22.50 , inference loss is 14.28, weight deacy loss is 8.22, training accuracy is 0.343750, time 209.791 samples/sec
+epoch 6, total_step 96200, total loss is 17.12 , inference loss is 8.90, weight deacy loss is 8.22, training accuracy is 0.343750, time 212.448 samples/sec
+epoch 6, total_step 96220, total loss is 16.32 , inference loss is 8.11, weight deacy loss is 8.21, training accuracy is 0.500000, time 207.114 samples/sec
+epoch 6, total_step 96240, total loss is 13.03 , inference loss is 4.82, weight deacy loss is 8.21, training accuracy is 0.562500, time 204.738 samples/sec
+epoch 6, total_step 96260, total loss is 23.02 , inference loss is 14.81, weight deacy loss is 8.21, training accuracy is 0.125000, time 213.680 samples/sec
+epoch 6, total_step 96280, total loss is 21.71 , inference loss is 13.50, weight deacy loss is 8.21, training accuracy is 0.218750, time 260.995 samples/sec
+epoch 6, total_step 96300, total loss is 14.84 , inference loss is 6.63, weight deacy loss is 8.21, training accuracy is 0.406250, time 13.625 samples/sec
+epoch 6, total_step 96320, total loss is 16.29 , inference loss is 8.08, weight deacy loss is 8.21, training accuracy is 0.375000, time 218.690 samples/sec
+epoch 6, total_step 96340, total loss is 15.37 , inference loss is 7.16, weight deacy loss is 8.21, training accuracy is 0.406250, time 212.250 samples/sec
+epoch 6, total_step 96360, total loss is 17.97 , inference loss is 9.76, weight deacy loss is 8.21, training accuracy is 0.218750, time 217.356 samples/sec
+epoch 6, total_step 96380, total loss is 18.45 , inference loss is 10.24, weight deacy loss is 8.21, training accuracy is 0.250000, time 211.117 samples/sec
+epoch 6, total_step 96400, total loss is 18.06 , inference loss is 9.86, weight deacy loss is 8.21, training accuracy is 0.375000, time 214.140 samples/sec
+epoch 6, total_step 96420, total loss is 15.21 , inference loss is 7.00, weight deacy loss is 8.21, training accuracy is 0.375000, time 208.236 samples/sec
+epoch 6, total_step 96440, total loss is 15.08 , inference loss is 6.87, weight deacy loss is 8.21, training accuracy is 0.437500, time 214.041 samples/sec
+epoch 6, total_step 96460, total loss is 15.80 , inference loss is 7.60, weight deacy loss is 8.21, training accuracy is 0.406250, time 212.031 samples/sec
+epoch 6, total_step 96480, total loss is 18.78 , inference loss is 10.57, weight deacy loss is 8.21, training accuracy is 0.250000, time 214.734 samples/sec
+epoch 6, total_step 96500, total loss is 16.86 , inference loss is 8.65, weight deacy loss is 8.20, training accuracy is 0.343750, time 227.440 samples/sec
+epoch 6, total_step 96520, total loss is 15.42 , inference loss is 7.21, weight deacy loss is 8.20, training accuracy is 0.562500, time 210.922 samples/sec
+epoch 6, total_step 96540, total loss is 21.82 , inference loss is 13.62, weight deacy loss is 8.20, training accuracy is 0.187500, time 255.943 samples/sec
+epoch 6, total_step 96560, total loss is 15.80 , inference loss is 7.60, weight deacy loss is 8.20, training accuracy is 0.437500, time 204.950 samples/sec
+epoch 6, total_step 96580, total loss is 19.01 , inference loss is 10.81, weight deacy loss is 8.20, training accuracy is 0.250000, time 211.056 samples/sec
+epoch 6, total_step 96600, total loss is 18.92 , inference loss is 10.72, weight deacy loss is 8.20, training accuracy is 0.343750, time 13.208 samples/sec
+epoch 6, total_step 96620, total loss is 20.63 , inference loss is 12.43, weight deacy loss is 8.20, training accuracy is 0.375000, time 220.637 samples/sec
+epoch 6, total_step 96640, total loss is 19.63 , inference loss is 11.43, weight deacy loss is 8.20, training accuracy is 0.250000, time 209.695 samples/sec
+epoch 6, total_step 96660, total loss is 19.49 , inference loss is 11.29, weight deacy loss is 8.20, training accuracy is 0.343750, time 217.488 samples/sec
+epoch 6, total_step 96680, total loss is 21.19 , inference loss is 12.99, weight deacy loss is 8.20, training accuracy is 0.187500, time 210.519 samples/sec
+epoch 6, total_step 96700, total loss is 17.88 , inference loss is 9.69, weight deacy loss is 8.20, training accuracy is 0.343750, time 204.575 samples/sec
+epoch 6, total_step 96720, total loss is 20.96 , inference loss is 12.77, weight deacy loss is 8.20, training accuracy is 0.218750, time 203.743 samples/sec
+epoch 6, total_step 96740, total loss is 20.28 , inference loss is 12.08, weight deacy loss is 8.20, training accuracy is 0.312500, time 211.330 samples/sec
+epoch 6, total_step 96760, total loss is 21.12 , inference loss is 12.93, weight deacy loss is 8.19, training accuracy is 0.312500, time 214.439 samples/sec
+epoch 6, total_step 96780, total loss is 18.19 , inference loss is 9.99, weight deacy loss is 8.19, training accuracy is 0.343750, time 211.301 samples/sec
+epoch 6, total_step 96800, total loss is 19.88 , inference loss is 11.69, weight deacy loss is 8.19, training accuracy is 0.218750, time 299.024 samples/sec
+epoch 6, total_step 96820, total loss is 18.23 , inference loss is 10.04, weight deacy loss is 8.19, training accuracy is 0.343750, time 213.143 samples/sec
+epoch 6, total_step 96840, total loss is 23.18 , inference loss is 14.99, weight deacy loss is 8.19, training accuracy is 0.312500, time 209.326 samples/sec
+epoch 6, total_step 96860, total loss is 20.50 , inference loss is 12.30, weight deacy loss is 8.19, training accuracy is 0.218750, time 215.203 samples/sec
+epoch 6, total_step 96880, total loss is 17.03 , inference loss is 8.84, weight deacy loss is 8.19, training accuracy is 0.406250, time 212.046 samples/sec
+epoch 6, total_step 96900, total loss is 18.78 , inference loss is 10.59, weight deacy loss is 8.19, training accuracy is 0.375000, time 13.734 samples/sec
+epoch 6, total_step 96920, total loss is 16.19 , inference loss is 8.00, weight deacy loss is 8.19, training accuracy is 0.375000, time 300.194 samples/sec
+epoch 6, total_step 96940, total loss is 20.59 , inference loss is 12.40, weight deacy loss is 8.19, training accuracy is 0.250000, time 201.497 samples/sec
+epoch 6, total_step 96960, total loss is 20.59 , inference loss is 12.40, weight deacy loss is 8.19, training accuracy is 0.281250, time 214.912 samples/sec
+epoch 6, total_step 96980, total loss is 18.68 , inference loss is 10.50, weight deacy loss is 8.19, training accuracy is 0.343750, time 205.366 samples/sec
+epoch 6, total_step 97000, total loss is 21.53 , inference loss is 13.34, weight deacy loss is 8.19, training accuracy is 0.250000, time 209.446 samples/sec
+epoch 6, total_step 97020, total loss is 18.88 , inference loss is 10.69, weight deacy loss is 8.19, training accuracy is 0.218750, time 202.133 samples/sec
+epoch 6, total_step 97040, total loss is 18.34 , inference loss is 10.16, weight deacy loss is 8.18, training accuracy is 0.250000, time 206.228 samples/sec
+epoch 6, total_step 97060, total loss is 17.13 , inference loss is 8.95, weight deacy loss is 8.18, training accuracy is 0.437500, time 211.048 samples/sec
+epoch 6, total_step 97080, total loss is 14.22 , inference loss is 6.04, weight deacy loss is 8.18, training accuracy is 0.531250, time 211.841 samples/sec
+epoch 6, total_step 97100, total loss is 17.15 , inference loss is 8.97, weight deacy loss is 8.18, training accuracy is 0.250000, time 206.529 samples/sec
+epoch 6, total_step 97120, total loss is 17.59 , inference loss is 9.41, weight deacy loss is 8.18, training accuracy is 0.343750, time 200.508 samples/sec
+epoch 6, total_step 97140, total loss is 18.52 , inference loss is 10.34, weight deacy loss is 8.18, training accuracy is 0.375000, time 203.353 samples/sec
+epoch 6, total_step 97160, total loss is 19.71 , inference loss is 11.53, weight deacy loss is 8.18, training accuracy is 0.281250, time 204.979 samples/sec
+epoch 6, total_step 97180, total loss is 18.14 , inference loss is 9.96, weight deacy loss is 8.18, training accuracy is 0.375000, time 212.324 samples/sec
+epoch 6, total_step 97200, total loss is 19.31 , inference loss is 11.13, weight deacy loss is 8.18, training accuracy is 0.250000, time 12.951 samples/sec
+epoch 6, total_step 97220, total loss is 14.69 , inference loss is 6.51, weight deacy loss is 8.18, training accuracy is 0.375000, time 210.619 samples/sec
+epoch 6, total_step 97240, total loss is 16.14 , inference loss is 7.96, weight deacy loss is 8.18, training accuracy is 0.375000, time 205.762 samples/sec
+epoch 6, total_step 97260, total loss is 19.58 , inference loss is 11.40, weight deacy loss is 8.18, training accuracy is 0.406250, time 204.129 samples/sec
+epoch 6, total_step 97280, total loss is 19.71 , inference loss is 11.54, weight deacy loss is 8.18, training accuracy is 0.312500, time 209.618 samples/sec
+epoch 6, total_step 97300, total loss is 17.23 , inference loss is 9.05, weight deacy loss is 8.17, training accuracy is 0.312500, time 206.529 samples/sec
+epoch 6, total_step 97320, total loss is 13.01 , inference loss is 4.83, weight deacy loss is 8.17, training accuracy is 0.593750, time 211.690 samples/sec
+epoch 6, total_step 97340, total loss is 14.01 , inference loss is 5.84, weight deacy loss is 8.17, training accuracy is 0.531250, time 206.181 samples/sec
+epoch 6, total_step 97360, total loss is 20.90 , inference loss is 12.73, weight deacy loss is 8.17, training accuracy is 0.375000, time 215.096 samples/sec
+epoch 6, total_step 97380, total loss is 16.97 , inference loss is 8.80, weight deacy loss is 8.17, training accuracy is 0.343750, time 208.331 samples/sec
+epoch 6, total_step 97400, total loss is 21.18 , inference loss is 13.01, weight deacy loss is 8.17, training accuracy is 0.343750, time 209.232 samples/sec
+epoch 6, total_step 97420, total loss is 19.93 , inference loss is 11.76, weight deacy loss is 8.17, training accuracy is 0.187500, time 214.435 samples/sec
+epoch 6, total_step 97440, total loss is 16.46 , inference loss is 8.29, weight deacy loss is 8.17, training accuracy is 0.437500, time 212.250 samples/sec
+epoch 6, total_step 97460, total loss is 15.39 , inference loss is 7.22, weight deacy loss is 8.17, training accuracy is 0.406250, time 214.680 samples/sec
+epoch 6, total_step 97480, total loss is 22.06 , inference loss is 13.89, weight deacy loss is 8.17, training accuracy is 0.250000, time 211.812 samples/sec
+epoch 6, total_step 97500, total loss is 18.61 , inference loss is 10.44, weight deacy loss is 8.17, training accuracy is 0.312500, time 13.688 samples/sec
+epoch 6, total_step 97520, total loss is 19.81 , inference loss is 11.64, weight deacy loss is 8.17, training accuracy is 0.250000, time 218.923 samples/sec
+epoch 6, total_step 97540, total loss is 13.89 , inference loss is 5.72, weight deacy loss is 8.17, training accuracy is 0.406250, time 208.217 samples/sec
+epoch 6, total_step 97560, total loss is 17.29 , inference loss is 9.12, weight deacy loss is 8.17, training accuracy is 0.437500, time 217.732 samples/sec
+epoch 6, total_step 97580, total loss is 19.44 , inference loss is 11.28, weight deacy loss is 8.16, training accuracy is 0.281250, time 206.821 samples/sec
+epoch 6, total_step 97600, total loss is 16.79 , inference loss is 8.63, weight deacy loss is 8.16, training accuracy is 0.281250, time 205.689 samples/sec
+epoch 6, total_step 97620, total loss is 20.87 , inference loss is 12.71, weight deacy loss is 8.16, training accuracy is 0.312500, time 210.316 samples/sec
+epoch 6, total_step 97640, total loss is 17.75 , inference loss is 9.59, weight deacy loss is 8.16, training accuracy is 0.406250, time 220.849 samples/sec
+epoch 6, total_step 97660, total loss is 16.28 , inference loss is 8.12, weight deacy loss is 8.16, training accuracy is 0.468750, time 221.315 samples/sec
+epoch 6, total_step 97680, total loss is 20.84 , inference loss is 12.68, weight deacy loss is 8.16, training accuracy is 0.375000, time 204.705 samples/sec
+epoch 6, total_step 97700, total loss is 17.52 , inference loss is 9.36, weight deacy loss is 8.16, training accuracy is 0.312500, time 211.855 samples/sec
+epoch 6, total_step 97720, total loss is 17.19 , inference loss is 9.03, weight deacy loss is 8.16, training accuracy is 0.437500, time 217.585 samples/sec
+epoch 6, total_step 97740, total loss is 18.49 , inference loss is 10.34, weight deacy loss is 8.16, training accuracy is 0.343750, time 215.635 samples/sec
+epoch 6, total_step 97760, total loss is 18.75 , inference loss is 10.59, weight deacy loss is 8.16, training accuracy is 0.281250, time 214.860 samples/sec
+epoch 6, total_step 97780, total loss is 20.51 , inference loss is 12.36, weight deacy loss is 8.16, training accuracy is 0.156250, time 298.205 samples/sec
+epoch 6, total_step 97800, total loss is 15.96 , inference loss is 7.80, weight deacy loss is 8.16, training accuracy is 0.468750, time 13.882 samples/sec
+epoch 6, total_step 97820, total loss is 13.42 , inference loss is 5.26, weight deacy loss is 8.16, training accuracy is 0.468750, time 209.335 samples/sec
+epoch 6, total_step 97840, total loss is 17.17 , inference loss is 9.02, weight deacy loss is 8.16, training accuracy is 0.375000, time 218.224 samples/sec
+epoch 6, total_step 97860, total loss is 19.23 , inference loss is 11.07, weight deacy loss is 8.15, training accuracy is 0.281250, time 214.887 samples/sec
+epoch 6, total_step 97880, total loss is 17.05 , inference loss is 8.89, weight deacy loss is 8.15, training accuracy is 0.312500, time 208.794 samples/sec
+epoch 6, total_step 97900, total loss is 21.14 , inference loss is 12.99, weight deacy loss is 8.15, training accuracy is 0.250000, time 217.658 samples/sec
+epoch 6, total_step 97920, total loss is 18.19 , inference loss is 10.03, weight deacy loss is 8.15, training accuracy is 0.375000, time 210.077 samples/sec
+epoch 6, total_step 97940, total loss is 17.39 , inference loss is 9.24, weight deacy loss is 8.15, training accuracy is 0.250000, time 205.998 samples/sec
+epoch 6, total_step 97960, total loss is 22.41 , inference loss is 14.26, weight deacy loss is 8.15, training accuracy is 0.156250, time 213.587 samples/sec
+epoch 6, total_step 97980, total loss is 17.36 , inference loss is 9.21, weight deacy loss is 8.15, training accuracy is 0.375000, time 211.485 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.03744600000001
+best_threshold_index 111 0.9877777777777778
+best_threshold_index 111 0.9864814814814815
+best_threshold_index 109 0.9874074074074074
+best_threshold_index 111 0.9875925925925926
+best_threshold_index 111 0.9877777777777778
+best_threshold_index 111 0.9861111111111112
+best_threshold_index 111 0.9868518518518519
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 111 0.9859259259259259
+best_threshold_index 111 0.9859259259259259
+[lfw][98000]XNorm: 21.514770
+[lfw][98000]Accuracy-Flip: 0.98650+-0.00652
+test accuracy is: 0.9864999999999998
+epoch 6, total_step 98000, total loss is 18.32 , inference loss is 10.17, weight deacy loss is 8.15, training accuracy is 0.312500, time 17.942 samples/sec
+epoch 6, total_step 98020, total loss is 19.31 , inference loss is 11.16, weight deacy loss is 8.15, training accuracy is 0.218750, time 214.131 samples/sec
+epoch 6, total_step 98040, total loss is 18.61 , inference loss is 10.46, weight deacy loss is 8.15, training accuracy is 0.312500, time 209.991 samples/sec
+epoch 6, total_step 98060, total loss is 17.64 , inference loss is 9.49, weight deacy loss is 8.15, training accuracy is 0.406250, time 224.198 samples/sec
+epoch 6, total_step 98080, total loss is 20.90 , inference loss is 12.76, weight deacy loss is 8.15, training accuracy is 0.312500, time 215.276 samples/sec
+epoch 6, total_step 98100, total loss is 18.87 , inference loss is 10.72, weight deacy loss is 8.15, training accuracy is 0.312500, time 13.552 samples/sec
+epoch 6, total_step 98120, total loss is 15.38 , inference loss is 7.24, weight deacy loss is 8.14, training accuracy is 0.562500, time 190.165 samples/sec
+epoch 6, total_step 98140, total loss is 17.62 , inference loss is 9.48, weight deacy loss is 8.14, training accuracy is 0.343750, time 212.134 samples/sec
+epoch 6, total_step 98160, total loss is 19.29 , inference loss is 11.15, weight deacy loss is 8.14, training accuracy is 0.250000, time 206.994 samples/sec
+epoch 6, total_step 98180, total loss is 19.85 , inference loss is 11.71, weight deacy loss is 8.14, training accuracy is 0.281250, time 206.203 samples/sec
+epoch 6, total_step 98200, total loss is 15.21 , inference loss is 7.07, weight deacy loss is 8.14, training accuracy is 0.531250, time 224.035 samples/sec
+epoch 6, total_step 98220, total loss is 20.14 , inference loss is 12.00, weight deacy loss is 8.14, training accuracy is 0.281250, time 212.875 samples/sec
+epoch 6, total_step 98240, total loss is 19.55 , inference loss is 11.41, weight deacy loss is 8.14, training accuracy is 0.375000, time 214.516 samples/sec
+epoch 6, total_step 98260, total loss is 19.56 , inference loss is 11.42, weight deacy loss is 8.14, training accuracy is 0.187500, time 214.253 samples/sec
+epoch 6, total_step 98280, total loss is 16.51 , inference loss is 8.37, weight deacy loss is 8.14, training accuracy is 0.375000, time 214.002 samples/sec
+epoch 6, total_step 98300, total loss is 19.49 , inference loss is 11.35, weight deacy loss is 8.14, training accuracy is 0.250000, time 207.169 samples/sec
+epoch 6, total_step 98320, total loss is 18.99 , inference loss is 10.86, weight deacy loss is 8.14, training accuracy is 0.375000, time 218.630 samples/sec
+epoch 6, total_step 98340, total loss is 20.62 , inference loss is 12.48, weight deacy loss is 8.14, training accuracy is 0.312500, time 199.690 samples/sec
+epoch 6, total_step 98360, total loss is 17.01 , inference loss is 8.87, weight deacy loss is 8.14, training accuracy is 0.406250, time 204.008 samples/sec
+epoch 6, total_step 98380, total loss is 19.84 , inference loss is 11.70, weight deacy loss is 8.13, training accuracy is 0.187500, time 213.310 samples/sec
+epoch 6, total_step 98400, total loss is 22.61 , inference loss is 14.48, weight deacy loss is 8.13, training accuracy is 0.343750, time 13.991 samples/sec
+epoch 6, total_step 98420, total loss is 22.03 , inference loss is 13.90, weight deacy loss is 8.13, training accuracy is 0.250000, time 218.726 samples/sec
+epoch 6, total_step 98440, total loss is 17.25 , inference loss is 9.12, weight deacy loss is 8.13, training accuracy is 0.312500, time 207.092 samples/sec
+epoch 6, total_step 98460, total loss is 18.20 , inference loss is 10.07, weight deacy loss is 8.13, training accuracy is 0.281250, time 207.148 samples/sec
+epoch 6, total_step 98480, total loss is 18.83 , inference loss is 10.70, weight deacy loss is 8.13, training accuracy is 0.250000, time 210.607 samples/sec
+epoch 6, total_step 98500, total loss is 17.20 , inference loss is 9.07, weight deacy loss is 8.13, training accuracy is 0.218750, time 213.495 samples/sec
+epoch 6, total_step 98520, total loss is 17.18 , inference loss is 9.05, weight deacy loss is 8.13, training accuracy is 0.218750, time 208.302 samples/sec
+epoch 6, total_step 98540, total loss is 17.07 , inference loss is 8.94, weight deacy loss is 8.13, training accuracy is 0.281250, time 210.872 samples/sec
+epoch 6, total_step 98560, total loss is 21.28 , inference loss is 13.15, weight deacy loss is 8.13, training accuracy is 0.312500, time 221.566 samples/sec
+epoch 6, total_step 98580, total loss is 14.61 , inference loss is 6.48, weight deacy loss is 8.13, training accuracy is 0.375000, time 207.886 samples/sec
+epoch 6, total_step 98600, total loss is 15.84 , inference loss is 7.72, weight deacy loss is 8.13, training accuracy is 0.312500, time 212.875 samples/sec
+epoch 6, total_step 98620, total loss is 17.45 , inference loss is 9.33, weight deacy loss is 8.13, training accuracy is 0.375000, time 212.060 samples/sec
+epoch 6, total_step 98640, total loss is 17.83 , inference loss is 9.71, weight deacy loss is 8.13, training accuracy is 0.406250, time 212.803 samples/sec
+epoch 6, total_step 98660, total loss is 20.41 , inference loss is 12.29, weight deacy loss is 8.12, training accuracy is 0.218750, time 213.339 samples/sec
+epoch 6, total_step 98680, total loss is 16.61 , inference loss is 8.49, weight deacy loss is 8.12, training accuracy is 0.343750, time 213.551 samples/sec
+epoch 6, total_step 98700, total loss is 22.13 , inference loss is 14.01, weight deacy loss is 8.12, training accuracy is 0.343750, time 15.066 samples/sec
+epoch 6, total_step 98720, total loss is 20.59 , inference loss is 12.47, weight deacy loss is 8.12, training accuracy is 0.281250, time 209.130 samples/sec
+epoch 6, total_step 98740, total loss is 17.73 , inference loss is 9.61, weight deacy loss is 8.12, training accuracy is 0.312500, time 210.281 samples/sec
+epoch 6, total_step 98760, total loss is 16.60 , inference loss is 8.48, weight deacy loss is 8.12, training accuracy is 0.250000, time 211.201 samples/sec
+epoch 6, total_step 98780, total loss is 18.41 , inference loss is 10.29, weight deacy loss is 8.12, training accuracy is 0.312500, time 203.316 samples/sec
+epoch 6, total_step 98800, total loss is 19.72 , inference loss is 11.60, weight deacy loss is 8.12, training accuracy is 0.187500, time 206.111 samples/sec
+epoch 6, total_step 98820, total loss is 22.12 , inference loss is 14.00, weight deacy loss is 8.12, training accuracy is 0.281250, time 295.711 samples/sec
+epoch 6, total_step 98840, total loss is 20.52 , inference loss is 12.40, weight deacy loss is 8.12, training accuracy is 0.250000, time 220.761 samples/sec
+epoch 6, total_step 98860, total loss is 19.26 , inference loss is 11.14, weight deacy loss is 8.12, training accuracy is 0.187500, time 211.535 samples/sec
+epoch 6, total_step 98880, total loss is 18.85 , inference loss is 10.73, weight deacy loss is 8.12, training accuracy is 0.218750, time 213.422 samples/sec
+epoch 6, total_step 98900, total loss is 18.72 , inference loss is 10.61, weight deacy loss is 8.12, training accuracy is 0.187500, time 213.966 samples/sec
+epoch 6, total_step 98920, total loss is 15.76 , inference loss is 7.65, weight deacy loss is 8.11, training accuracy is 0.406250, time 209.050 samples/sec
+epoch 6, total_step 98940, total loss is 15.94 , inference loss is 7.82, weight deacy loss is 8.11, training accuracy is 0.343750, time 211.085 samples/sec
+epoch 6, total_step 98960, total loss is 17.89 , inference loss is 9.78, weight deacy loss is 8.11, training accuracy is 0.187500, time 222.921 samples/sec
+epoch 6, total_step 98980, total loss is 17.21 , inference loss is 9.10, weight deacy loss is 8.11, training accuracy is 0.343750, time 206.922 samples/sec
+epoch 6, total_step 99000, total loss is 17.95 , inference loss is 9.84, weight deacy loss is 8.11, training accuracy is 0.468750, time 13.614 samples/sec
+epoch 6, total_step 99020, total loss is 20.75 , inference loss is 12.63, weight deacy loss is 8.11, training accuracy is 0.218750, time 207.336 samples/sec
+epoch 6, total_step 99040, total loss is 18.29 , inference loss is 10.18, weight deacy loss is 8.11, training accuracy is 0.218750, time 204.243 samples/sec
+epoch 6, total_step 99060, total loss is 18.67 , inference loss is 10.56, weight deacy loss is 8.11, training accuracy is 0.406250, time 205.728 samples/sec
+epoch 6, total_step 99080, total loss is 19.87 , inference loss is 11.76, weight deacy loss is 8.11, training accuracy is 0.312500, time 204.206 samples/sec
+epoch 6, total_step 99100, total loss is 18.92 , inference loss is 10.81, weight deacy loss is 8.11, training accuracy is 0.250000, time 207.984 samples/sec
+epoch 6, total_step 99120, total loss is 18.09 , inference loss is 9.98, weight deacy loss is 8.11, training accuracy is 0.281250, time 199.618 samples/sec
+epoch 6, total_step 99140, total loss is 23.16 , inference loss is 15.05, weight deacy loss is 8.11, training accuracy is 0.281250, time 206.857 samples/sec
+epoch 6, total_step 99160, total loss is 19.41 , inference loss is 11.30, weight deacy loss is 8.11, training accuracy is 0.312500, time 209.285 samples/sec
+epoch 6, total_step 99180, total loss is 22.16 , inference loss is 14.06, weight deacy loss is 8.11, training accuracy is 0.125000, time 218.053 samples/sec
+epoch 6, total_step 99200, total loss is 20.82 , inference loss is 12.72, weight deacy loss is 8.11, training accuracy is 0.281250, time 212.117 samples/sec
+epoch 6, total_step 99220, total loss is 20.70 , inference loss is 12.60, weight deacy loss is 8.10, training accuracy is 0.375000, time 207.041 samples/sec
+epoch 6, total_step 99240, total loss is 18.46 , inference loss is 10.35, weight deacy loss is 8.10, training accuracy is 0.218750, time 209.989 samples/sec
+epoch 6, total_step 99260, total loss is 12.79 , inference loss is 4.69, weight deacy loss is 8.10, training accuracy is 0.468750, time 222.848 samples/sec
+epoch 6, total_step 99280, total loss is 19.29 , inference loss is 11.18, weight deacy loss is 8.10, training accuracy is 0.281250, time 198.538 samples/sec
+epoch 6, total_step 99300, total loss is 18.77 , inference loss is 10.67, weight deacy loss is 8.10, training accuracy is 0.343750, time 14.002 samples/sec
+epoch 6, total_step 99320, total loss is 21.75 , inference loss is 13.65, weight deacy loss is 8.10, training accuracy is 0.187500, time 202.519 samples/sec
+epoch 6, total_step 99340, total loss is 14.74 , inference loss is 6.64, weight deacy loss is 8.10, training accuracy is 0.437500, time 210.932 samples/sec
+epoch 6, total_step 99360, total loss is 17.09 , inference loss is 8.99, weight deacy loss is 8.10, training accuracy is 0.500000, time 300.193 samples/sec
+epoch 6, total_step 99380, total loss is 14.56 , inference loss is 6.46, weight deacy loss is 8.10, training accuracy is 0.375000, time 210.952 samples/sec
+epoch 6, total_step 99400, total loss is 17.56 , inference loss is 9.46, weight deacy loss is 8.10, training accuracy is 0.281250, time 203.515 samples/sec
+epoch 6, total_step 99420, total loss is 18.16 , inference loss is 10.07, weight deacy loss is 8.10, training accuracy is 0.375000, time 201.944 samples/sec
+epoch 6, total_step 99440, total loss is 20.56 , inference loss is 12.46, weight deacy loss is 8.10, training accuracy is 0.250000, time 213.875 samples/sec
+epoch 6, total_step 99460, total loss is 18.33 , inference loss is 10.23, weight deacy loss is 8.10, training accuracy is 0.437500, time 211.490 samples/sec
+epoch 6, total_step 99480, total loss is 18.58 , inference loss is 10.49, weight deacy loss is 8.09, training accuracy is 0.187500, time 206.478 samples/sec
+epoch 6, total_step 99500, total loss is 18.02 , inference loss is 9.93, weight deacy loss is 8.09, training accuracy is 0.187500, time 206.056 samples/sec
+epoch 6, total_step 99520, total loss is 16.87 , inference loss is 8.78, weight deacy loss is 8.09, training accuracy is 0.312500, time 211.690 samples/sec
+epoch 6, total_step 99540, total loss is 14.66 , inference loss is 6.57, weight deacy loss is 8.09, training accuracy is 0.375000, time 209.303 samples/sec
+epoch 6, total_step 99560, total loss is 21.87 , inference loss is 13.78, weight deacy loss is 8.09, training accuracy is 0.343750, time 208.183 samples/sec
+epoch 6, total_step 99580, total loss is 14.48 , inference loss is 6.39, weight deacy loss is 8.09, training accuracy is 0.343750, time 190.699 samples/sec
+epoch 6, total_step 99600, total loss is 18.22 , inference loss is 10.13, weight deacy loss is 8.09, training accuracy is 0.281250, time 13.736 samples/sec
+epoch 6, total_step 99620, total loss is 15.56 , inference loss is 7.47, weight deacy loss is 8.09, training accuracy is 0.312500, time 298.047 samples/sec
+epoch 6, total_step 99640, total loss is 15.56 , inference loss is 7.47, weight deacy loss is 8.09, training accuracy is 0.593750, time 194.035 samples/sec
+epoch 6, total_step 99660, total loss is 15.63 , inference loss is 7.55, weight deacy loss is 8.09, training accuracy is 0.375000, time 211.042 samples/sec
+epoch 6, total_step 99680, total loss is 17.89 , inference loss is 9.80, weight deacy loss is 8.09, training accuracy is 0.375000, time 215.354 samples/sec
+epoch 6, total_step 99700, total loss is 18.58 , inference loss is 10.50, weight deacy loss is 8.09, training accuracy is 0.312500, time 211.193 samples/sec
+epoch 6, total_step 99720, total loss is 18.32 , inference loss is 10.24, weight deacy loss is 8.09, training accuracy is 0.343750, time 209.467 samples/sec
+epoch 6, total_step 99740, total loss is 18.69 , inference loss is 10.61, weight deacy loss is 8.08, training accuracy is 0.343750, time 210.097 samples/sec
+epoch 6, total_step 99760, total loss is 17.44 , inference loss is 9.36, weight deacy loss is 8.08, training accuracy is 0.343750, time 214.077 samples/sec
+epoch 6, total_step 99780, total loss is 15.96 , inference loss is 7.88, weight deacy loss is 8.08, training accuracy is 0.375000, time 208.649 samples/sec
+epoch 6, total_step 99800, total loss is 18.55 , inference loss is 10.47, weight deacy loss is 8.08, training accuracy is 0.312500, time 222.527 samples/sec
+epoch 6, total_step 99820, total loss is 13.82 , inference loss is 5.74, weight deacy loss is 8.08, training accuracy is 0.437500, time 215.971 samples/sec
+epoch 6, total_step 99840, total loss is 18.55 , inference loss is 10.47, weight deacy loss is 8.08, training accuracy is 0.218750, time 207.415 samples/sec
+epoch 6, total_step 99860, total loss is 19.61 , inference loss is 11.53, weight deacy loss is 8.08, training accuracy is 0.281250, time 217.372 samples/sec
+epoch 6, total_step 99880, total loss is 21.19 , inference loss is 13.11, weight deacy loss is 8.08, training accuracy is 0.250000, time 204.443 samples/sec
+epoch 6, total_step 99900, total loss is 17.14 , inference loss is 9.06, weight deacy loss is 8.08, training accuracy is 0.281250, time 14.643 samples/sec
+epoch 6, total_step 99920, total loss is 18.80 , inference loss is 10.72, weight deacy loss is 8.08, training accuracy is 0.281250, time 213.414 samples/sec
+epoch 6, total_step 99940, total loss is 22.32 , inference loss is 14.24, weight deacy loss is 8.08, training accuracy is 0.406250, time 204.845 samples/sec
+epoch 6, total_step 99960, total loss is 14.01 , inference loss is 5.93, weight deacy loss is 8.08, training accuracy is 0.375000, time 210.307 samples/sec
+epoch 6, total_step 99980, total loss is 19.50 , inference loss is 11.43, weight deacy loss is 8.08, training accuracy is 0.312500, time 209.340 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.091820999999996
+best_threshold_index 112 0.9883333333333333
+best_threshold_index 112 0.987037037037037
+best_threshold_index 110 0.9879629629629629
+best_threshold_index 110 0.9890740740740741
+best_threshold_index 112 0.9877777777777778
+best_threshold_index 112 0.9868518518518519
+best_threshold_index 110 0.9881481481481481
+best_threshold_index 112 0.9868518518518519
+best_threshold_index 112 0.9864814814814815
+best_threshold_index 112 0.9868518518518519
+[lfw][100000]XNorm: 22.323275
+[lfw][100000]Accuracy-Flip: 0.98667+-0.00775
+test accuracy is: 0.9866666666666667
+epoch 6, total_step 100000, total loss is 16.97 , inference loss is 8.90, weight deacy loss is 8.07, training accuracy is 0.281250, time 18.650 samples/sec
+epoch 6, total_step 100020, total loss is 13.99 , inference loss is 5.91, weight deacy loss is 8.07, training accuracy is 0.531250, time 191.798 samples/sec
+epoch 6, total_step 100040, total loss is 19.44 , inference loss is 11.37, weight deacy loss is 8.07, training accuracy is 0.312500, time 214.524 samples/sec
+epoch 6, total_step 100060, total loss is 21.14 , inference loss is 13.06, weight deacy loss is 8.07, training accuracy is 0.125000, time 214.984 samples/sec
+epoch 6, total_step 100080, total loss is 19.35 , inference loss is 11.28, weight deacy loss is 8.07, training accuracy is 0.187500, time 211.736 samples/sec
+epoch 6, total_step 100100, total loss is 16.49 , inference loss is 8.42, weight deacy loss is 8.07, training accuracy is 0.375000, time 204.329 samples/sec
+epoch 6, total_step 100120, total loss is 17.99 , inference loss is 9.92, weight deacy loss is 8.07, training accuracy is 0.406250, time 204.482 samples/sec
+epoch 6, total_step 100140, total loss is 21.10 , inference loss is 13.03, weight deacy loss is 8.07, training accuracy is 0.187500, time 206.003 samples/sec
+epoch 6, total_step 100160, total loss is 17.07 , inference loss is 9.00, weight deacy loss is 8.07, training accuracy is 0.312500, time 210.097 samples/sec
+epoch 6, total_step 100180, total loss is 17.96 , inference loss is 9.89, weight deacy loss is 8.07, training accuracy is 0.406250, time 209.629 samples/sec
+epoch 6, total_step 100200, total loss is 19.85 , inference loss is 11.78, weight deacy loss is 8.07, training accuracy is 0.437500, time 14.594 samples/sec
+epoch 6, total_step 100220, total loss is 20.78 , inference loss is 12.71, weight deacy loss is 8.07, training accuracy is 0.343750, time 204.571 samples/sec
+epoch 6, total_step 100240, total loss is 17.08 , inference loss is 9.02, weight deacy loss is 8.07, training accuracy is 0.375000, time 300.183 samples/sec
+epoch 6, total_step 100260, total loss is 19.44 , inference loss is 11.37, weight deacy loss is 8.07, training accuracy is 0.218750, time 215.825 samples/sec
+epoch 6, total_step 100280, total loss is 17.72 , inference loss is 9.66, weight deacy loss is 8.06, training accuracy is 0.343750, time 210.096 samples/sec
+epoch 6, total_step 100300, total loss is 18.48 , inference loss is 10.41, weight deacy loss is 8.06, training accuracy is 0.281250, time 215.752 samples/sec
+epoch 6, total_step 100320, total loss is 19.31 , inference loss is 11.24, weight deacy loss is 8.06, training accuracy is 0.343750, time 208.767 samples/sec
+epoch 6, total_step 100340, total loss is 20.12 , inference loss is 12.06, weight deacy loss is 8.06, training accuracy is 0.218750, time 218.706 samples/sec
+epoch 6, total_step 100360, total loss is 16.87 , inference loss is 8.81, weight deacy loss is 8.06, training accuracy is 0.250000, time 213.157 samples/sec
+epoch 6, total_step 100380, total loss is 18.71 , inference loss is 10.65, weight deacy loss is 8.06, training accuracy is 0.406250, time 214.891 samples/sec
+epoch 6, total_step 100400, total loss is 17.98 , inference loss is 9.92, weight deacy loss is 8.06, training accuracy is 0.343750, time 225.817 samples/sec
+epoch 6, total_step 100420, total loss is 16.18 , inference loss is 8.12, weight deacy loss is 8.06, training accuracy is 0.343750, time 211.377 samples/sec
+epoch 6, total_step 100440, total loss is 19.71 , inference loss is 11.65, weight deacy loss is 8.06, training accuracy is 0.125000, time 214.711 samples/sec
+epoch 6, total_step 100460, total loss is 17.99 , inference loss is 9.93, weight deacy loss is 8.06, training accuracy is 0.375000, time 213.185 samples/sec
+epoch 6, total_step 100480, total loss is 18.73 , inference loss is 10.68, weight deacy loss is 8.06, training accuracy is 0.281250, time 218.182 samples/sec
+epoch 6, total_step 100500, total loss is 17.04 , inference loss is 8.98, weight deacy loss is 8.06, training accuracy is 0.156250, time 10.704 samples/sec
+epoch 6, total_step 100520, total loss is 19.37 , inference loss is 11.32, weight deacy loss is 8.06, training accuracy is 0.343750, time 222.915 samples/sec
+epoch 6, total_step 100540, total loss is 16.53 , inference loss is 8.48, weight deacy loss is 8.05, training accuracy is 0.312500, time 205.820 samples/sec
+epoch 6, total_step 100560, total loss is 22.00 , inference loss is 13.94, weight deacy loss is 8.05, training accuracy is 0.281250, time 210.465 samples/sec
+epoch 6, total_step 100580, total loss is 20.43 , inference loss is 12.38, weight deacy loss is 8.05, training accuracy is 0.343750, time 207.629 samples/sec
+epoch 6, total_step 100600, total loss is 19.08 , inference loss is 11.03, weight deacy loss is 8.05, training accuracy is 0.250000, time 202.972 samples/sec
+epoch 6, total_step 100620, total loss is 20.38 , inference loss is 12.33, weight deacy loss is 8.05, training accuracy is 0.281250, time 209.000 samples/sec
+epoch 6, total_step 100640, total loss is 19.14 , inference loss is 11.09, weight deacy loss is 8.05, training accuracy is 0.312500, time 210.896 samples/sec
+epoch 6, total_step 100660, total loss is 19.78 , inference loss is 11.73, weight deacy loss is 8.05, training accuracy is 0.343750, time 211.232 samples/sec
+epoch 6, total_step 100680, total loss is 20.97 , inference loss is 12.92, weight deacy loss is 8.05, training accuracy is 0.375000, time 299.578 samples/sec
+epoch 6, total_step 100700, total loss is 17.09 , inference loss is 9.04, weight deacy loss is 8.05, training accuracy is 0.187500, time 214.856 samples/sec
+epoch 6, total_step 100720, total loss is 15.55 , inference loss is 7.50, weight deacy loss is 8.05, training accuracy is 0.312500, time 207.835 samples/sec
+epoch 6, total_step 100740, total loss is 16.22 , inference loss is 8.17, weight deacy loss is 8.05, training accuracy is 0.281250, time 224.753 samples/sec
+epoch 6, total_step 100760, total loss is 18.58 , inference loss is 10.53, weight deacy loss is 8.05, training accuracy is 0.187500, time 209.188 samples/sec
+epoch 6, total_step 100780, total loss is 21.61 , inference loss is 13.56, weight deacy loss is 8.05, training accuracy is 0.312500, time 207.702 samples/sec
+epoch 6, total_step 100800, total loss is 20.02 , inference loss is 11.97, weight deacy loss is 8.05, training accuracy is 0.250000, time 13.464 samples/sec
+epoch 6, total_step 100820, total loss is 16.62 , inference loss is 8.58, weight deacy loss is 8.05, training accuracy is 0.437500, time 221.333 samples/sec
+epoch 6, total_step 100840, total loss is 20.60 , inference loss is 12.56, weight deacy loss is 8.04, training accuracy is 0.343750, time 212.246 samples/sec
+epoch 6, total_step 100860, total loss is 20.08 , inference loss is 12.03, weight deacy loss is 8.04, training accuracy is 0.343750, time 211.450 samples/sec
+epoch 6, total_step 100880, total loss is 17.71 , inference loss is 9.67, weight deacy loss is 8.04, training accuracy is 0.156250, time 216.392 samples/sec
+epoch 6, total_step 100900, total loss is 18.26 , inference loss is 10.22, weight deacy loss is 8.04, training accuracy is 0.343750, time 213.413 samples/sec
+epoch 6, total_step 100920, total loss is 20.20 , inference loss is 12.16, weight deacy loss is 8.04, training accuracy is 0.343750, time 215.249 samples/sec
+epoch 6, total_step 100940, total loss is 20.74 , inference loss is 12.70, weight deacy loss is 8.04, training accuracy is 0.343750, time 298.833 samples/sec
+epoch 6, total_step 100960, total loss is 16.09 , inference loss is 8.05, weight deacy loss is 8.04, training accuracy is 0.281250, time 208.883 samples/sec
+epoch 6, total_step 100980, total loss is 18.45 , inference loss is 10.41, weight deacy loss is 8.04, training accuracy is 0.375000, time 211.467 samples/sec
+epoch 6, total_step 101000, total loss is 18.86 , inference loss is 10.82, weight deacy loss is 8.04, training accuracy is 0.343750, time 221.149 samples/sec
+epoch 6, total_step 101020, total loss is 18.48 , inference loss is 10.45, weight deacy loss is 8.04, training accuracy is 0.343750, time 211.582 samples/sec
+epoch 6, total_step 101040, total loss is 17.28 , inference loss is 9.25, weight deacy loss is 8.04, training accuracy is 0.312500, time 214.332 samples/sec
+epoch 6, total_step 101060, total loss is 20.73 , inference loss is 12.69, weight deacy loss is 8.04, training accuracy is 0.281250, time 227.169 samples/sec
+epoch 6, total_step 101080, total loss is 18.68 , inference loss is 10.64, weight deacy loss is 8.04, training accuracy is 0.281250, time 206.100 samples/sec
+epoch 6, total_step 101100, total loss is 17.10 , inference loss is 9.06, weight deacy loss is 8.04, training accuracy is 0.375000, time 12.544 samples/sec
+epoch 6, total_step 101120, total loss is 21.82 , inference loss is 13.78, weight deacy loss is 8.03, training accuracy is 0.250000, time 205.947 samples/sec
+epoch 6, total_step 101140, total loss is 18.76 , inference loss is 10.72, weight deacy loss is 8.03, training accuracy is 0.312500, time 211.877 samples/sec
+epoch 6, total_step 101160, total loss is 19.51 , inference loss is 11.48, weight deacy loss is 8.03, training accuracy is 0.312500, time 207.625 samples/sec
+epoch 6, total_step 101180, total loss is 19.70 , inference loss is 11.67, weight deacy loss is 8.03, training accuracy is 0.250000, time 210.960 samples/sec
+epoch 6, total_step 101200, total loss is 16.66 , inference loss is 8.63, weight deacy loss is 8.03, training accuracy is 0.500000, time 212.705 samples/sec
+epoch 6, total_step 101220, total loss is 21.72 , inference loss is 13.68, weight deacy loss is 8.03, training accuracy is 0.281250, time 298.193 samples/sec
+epoch 6, total_step 101240, total loss is 17.65 , inference loss is 9.62, weight deacy loss is 8.03, training accuracy is 0.343750, time 217.465 samples/sec
+epoch 6, total_step 101260, total loss is 16.98 , inference loss is 8.95, weight deacy loss is 8.03, training accuracy is 0.406250, time 217.027 samples/sec
+epoch 6, total_step 101280, total loss is 16.65 , inference loss is 8.62, weight deacy loss is 8.03, training accuracy is 0.187500, time 208.105 samples/sec
+epoch 6, total_step 101300, total loss is 17.59 , inference loss is 9.56, weight deacy loss is 8.03, training accuracy is 0.406250, time 221.843 samples/sec
+epoch 6, total_step 101320, total loss is 17.43 , inference loss is 9.40, weight deacy loss is 8.03, training accuracy is 0.468750, time 206.873 samples/sec
+epoch 6, total_step 101340, total loss is 15.69 , inference loss is 7.66, weight deacy loss is 8.03, training accuracy is 0.437500, time 209.508 samples/sec
+epoch 6, total_step 101360, total loss is 18.57 , inference loss is 10.54, weight deacy loss is 8.03, training accuracy is 0.343750, time 220.102 samples/sec
+epoch 6, total_step 101380, total loss is 18.98 , inference loss is 10.95, weight deacy loss is 8.03, training accuracy is 0.312500, time 212.640 samples/sec
+epoch 6, total_step 101400, total loss is 17.33 , inference loss is 9.30, weight deacy loss is 8.02, training accuracy is 0.125000, time 13.519 samples/sec
+epoch 6, total_step 101420, total loss is 18.82 , inference loss is 10.79, weight deacy loss is 8.02, training accuracy is 0.312500, time 207.605 samples/sec
+epoch 6, total_step 101440, total loss is 16.86 , inference loss is 8.83, weight deacy loss is 8.02, training accuracy is 0.343750, time 212.594 samples/sec
+epoch 6, total_step 101460, total loss is 18.69 , inference loss is 10.67, weight deacy loss is 8.02, training accuracy is 0.375000, time 207.777 samples/sec
+epoch 6, total_step 101480, total loss is 23.48 , inference loss is 15.46, weight deacy loss is 8.02, training accuracy is 0.250000, time 300.653 samples/sec
+epoch 6, total_step 101500, total loss is 16.87 , inference loss is 8.85, weight deacy loss is 8.02, training accuracy is 0.281250, time 206.181 samples/sec
+epoch 6, total_step 101520, total loss is 22.29 , inference loss is 14.27, weight deacy loss is 8.02, training accuracy is 0.250000, time 216.409 samples/sec
+epoch 6, total_step 101540, total loss is 15.29 , inference loss is 7.27, weight deacy loss is 8.02, training accuracy is 0.406250, time 207.597 samples/sec
+epoch 6, total_step 101560, total loss is 19.50 , inference loss is 11.48, weight deacy loss is 8.02, training accuracy is 0.343750, time 202.451 samples/sec
+epoch 6, total_step 101580, total loss is 14.83 , inference loss is 6.82, weight deacy loss is 8.02, training accuracy is 0.343750, time 210.792 samples/sec
+epoch 6, total_step 101600, total loss is 14.24 , inference loss is 6.22, weight deacy loss is 8.02, training accuracy is 0.312500, time 218.925 samples/sec
+epoch 6, total_step 101620, total loss is 16.13 , inference loss is 8.11, weight deacy loss is 8.02, training accuracy is 0.437500, time 207.824 samples/sec
+epoch 6, total_step 101640, total loss is 17.07 , inference loss is 9.05, weight deacy loss is 8.02, training accuracy is 0.250000, time 211.445 samples/sec
+epoch 6, total_step 101660, total loss is 18.50 , inference loss is 10.48, weight deacy loss is 8.02, training accuracy is 0.281250, time 215.663 samples/sec
+epoch 6, total_step 101680, total loss is 15.32 , inference loss is 7.31, weight deacy loss is 8.01, training accuracy is 0.281250, time 212.195 samples/sec
+epoch 6, total_step 101700, total loss is 21.38 , inference loss is 13.37, weight deacy loss is 8.01, training accuracy is 0.343750, time 12.184 samples/sec
+epoch 6, total_step 101720, total loss is 18.56 , inference loss is 10.55, weight deacy loss is 8.01, training accuracy is 0.281250, time 208.782 samples/sec
+epoch 6, total_step 101740, total loss is 15.82 , inference loss is 7.81, weight deacy loss is 8.01, training accuracy is 0.406250, time 300.628 samples/sec
+epoch 6, total_step 101760, total loss is 17.02 , inference loss is 9.01, weight deacy loss is 8.01, training accuracy is 0.312500, time 213.757 samples/sec
+epoch 6, total_step 101780, total loss is 16.80 , inference loss is 8.79, weight deacy loss is 8.01, training accuracy is 0.343750, time 211.512 samples/sec
+epoch 6, total_step 101800, total loss is 17.89 , inference loss is 9.88, weight deacy loss is 8.01, training accuracy is 0.312500, time 198.650 samples/sec
+epoch 6, total_step 101820, total loss is 18.19 , inference loss is 10.18, weight deacy loss is 8.01, training accuracy is 0.250000, time 203.479 samples/sec
+epoch 6, total_step 101840, total loss is 21.58 , inference loss is 13.57, weight deacy loss is 8.01, training accuracy is 0.406250, time 211.060 samples/sec
+epoch 6, total_step 101860, total loss is 19.55 , inference loss is 11.54, weight deacy loss is 8.01, training accuracy is 0.250000, time 219.347 samples/sec
+epoch 6, total_step 101880, total loss is 15.69 , inference loss is 7.69, weight deacy loss is 8.01, training accuracy is 0.343750, time 216.447 samples/sec
+epoch 6, total_step 101900, total loss is 20.23 , inference loss is 12.23, weight deacy loss is 8.01, training accuracy is 0.312500, time 214.713 samples/sec
+epoch 6, total_step 101920, total loss is 14.58 , inference loss is 6.57, weight deacy loss is 8.01, training accuracy is 0.437500, time 226.378 samples/sec
+epoch 6, total_step 101940, total loss is 17.74 , inference loss is 9.73, weight deacy loss is 8.01, training accuracy is 0.281250, time 209.025 samples/sec
+epoch 6, total_step 101960, total loss is 18.20 , inference loss is 10.20, weight deacy loss is 8.00, training accuracy is 0.375000, time 226.939 samples/sec
+epoch 6, total_step 101980, total loss is 17.78 , inference loss is 9.78, weight deacy loss is 8.00, training accuracy is 0.250000, time 208.673 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.323135999999984
+best_threshold_index 105 0.9877777777777778
+best_threshold_index 105 0.9864814814814815
+best_threshold_index 105 0.9875925925925926
+best_threshold_index 105 0.9887037037037038
+best_threshold_index 105 0.9875925925925926
+best_threshold_index 105 0.9862962962962963
+best_threshold_index 105 0.9874074074074074
+best_threshold_index 105 0.9861111111111112
+best_threshold_index 105 0.9857407407407407
+best_threshold_index 105 0.9862962962962963
+[lfw][102000]XNorm: 23.505937
+[lfw][102000]Accuracy-Flip: 0.98700+-0.00806
+test accuracy is: 0.9870000000000001
+epoch 6, total_step 102000, total loss is 22.97 , inference loss is 14.97, weight deacy loss is 8.00, training accuracy is 0.125000, time 18.612 samples/sec
+epoch 6, total_step 102020, total loss is 16.36 , inference loss is 8.35, weight deacy loss is 8.00, training accuracy is 0.406250, time 203.286 samples/sec
+epoch 6, total_step 102040, total loss is 20.45 , inference loss is 12.45, weight deacy loss is 8.00, training accuracy is 0.312500, time 206.868 samples/sec
+epoch 6, total_step 102060, total loss is 14.73 , inference loss is 6.73, weight deacy loss is 8.00, training accuracy is 0.406250, time 211.942 samples/sec
+epoch 6, total_step 102080, total loss is 15.08 , inference loss is 7.08, weight deacy loss is 8.00, training accuracy is 0.500000, time 216.918 samples/sec
+epoch 6, total_step 102100, total loss is 17.07 , inference loss is 9.07, weight deacy loss is 8.00, training accuracy is 0.281250, time 213.384 samples/sec
+epoch 6, total_step 102120, total loss is 19.85 , inference loss is 11.85, weight deacy loss is 8.00, training accuracy is 0.156250, time 297.221 samples/sec
+epoch 6, total_step 102140, total loss is 18.79 , inference loss is 10.79, weight deacy loss is 8.00, training accuracy is 0.343750, time 298.327 samples/sec
+epoch 6, total_step 102160, total loss is 18.98 , inference loss is 10.98, weight deacy loss is 8.00, training accuracy is 0.281250, time 298.062 samples/sec
+epoch 6, total_step 102180, total loss is 21.61 , inference loss is 13.62, weight deacy loss is 8.00, training accuracy is 0.250000, time 297.874 samples/sec
+epoch 6, total_step 102200, total loss is 21.43 , inference loss is 13.44, weight deacy loss is 8.00, training accuracy is 0.281250, time 299.301 samples/sec
+epoch 6, total_step 102220, total loss is 17.40 , inference loss is 9.40, weight deacy loss is 8.00, training accuracy is 0.187500, time 216.950 samples/sec
+epoch 6, total_step 102240, total loss is 19.65 , inference loss is 11.66, weight deacy loss is 7.99, training accuracy is 0.250000, time 212.758 samples/sec
+epoch 6, total_step 102260, total loss is 24.34 , inference loss is 16.34, weight deacy loss is 7.99, training accuracy is 0.250000, time 218.131 samples/sec
+epoch 6, total_step 102280, total loss is 17.47 , inference loss is 9.48, weight deacy loss is 7.99, training accuracy is 0.156250, time 219.596 samples/sec
+epoch 6, total_step 102300, total loss is 19.52 , inference loss is 11.52, weight deacy loss is 7.99, training accuracy is 0.218750, time 14.530 samples/sec
+epoch 6, total_step 102320, total loss is 21.10 , inference loss is 13.11, weight deacy loss is 7.99, training accuracy is 0.250000, time 298.101 samples/sec
+epoch 6, total_step 102340, total loss is 21.36 , inference loss is 13.37, weight deacy loss is 7.99, training accuracy is 0.218750, time 297.767 samples/sec
+epoch 6, total_step 102360, total loss is 17.01 , inference loss is 9.02, weight deacy loss is 7.99, training accuracy is 0.312500, time 209.742 samples/sec
+epoch 6, total_step 102380, total loss is 18.06 , inference loss is 10.07, weight deacy loss is 7.99, training accuracy is 0.375000, time 213.476 samples/sec
+epoch 6, total_step 102400, total loss is 17.20 , inference loss is 9.21, weight deacy loss is 7.99, training accuracy is 0.312500, time 212.992 samples/sec
+epoch 6, total_step 102420, total loss is 15.47 , inference loss is 7.48, weight deacy loss is 7.99, training accuracy is 0.281250, time 216.414 samples/sec
+epoch 6, total_step 102440, total loss is 19.60 , inference loss is 11.61, weight deacy loss is 7.99, training accuracy is 0.187500, time 216.285 samples/sec
+epoch 6, total_step 102460, total loss is 16.63 , inference loss is 8.64, weight deacy loss is 7.99, training accuracy is 0.343750, time 205.628 samples/sec
+epoch 6, total_step 102480, total loss is 16.84 , inference loss is 8.86, weight deacy loss is 7.99, training accuracy is 0.437500, time 206.613 samples/sec
+epoch 6, total_step 102500, total loss is 19.52 , inference loss is 11.54, weight deacy loss is 7.99, training accuracy is 0.437500, time 213.975 samples/sec
+epoch 6, total_step 102520, total loss is 15.43 , inference loss is 7.44, weight deacy loss is 7.98, training accuracy is 0.437500, time 212.947 samples/sec
+epoch 6, total_step 102540, total loss is 21.02 , inference loss is 13.04, weight deacy loss is 7.98, training accuracy is 0.218750, time 218.000 samples/sec
+epoch 6, total_step 102560, total loss is 20.80 , inference loss is 12.81, weight deacy loss is 7.98, training accuracy is 0.312500, time 202.633 samples/sec
+epoch 6, total_step 102580, total loss is 17.79 , inference loss is 9.81, weight deacy loss is 7.98, training accuracy is 0.343750, time 212.701 samples/sec
+epoch 6, total_step 102600, total loss is 21.06 , inference loss is 13.08, weight deacy loss is 7.98, training accuracy is 0.312500, time 14.773 samples/sec
+epoch 6, total_step 102620, total loss is 17.88 , inference loss is 9.90, weight deacy loss is 7.98, training accuracy is 0.312500, time 206.000 samples/sec
+epoch 6, total_step 102640, total loss is 16.12 , inference loss is 8.14, weight deacy loss is 7.98, training accuracy is 0.468750, time 218.046 samples/sec
+epoch 6, total_step 102660, total loss is 18.38 , inference loss is 10.40, weight deacy loss is 7.98, training accuracy is 0.375000, time 213.989 samples/sec
+epoch 6, total_step 102680, total loss is 16.81 , inference loss is 8.83, weight deacy loss is 7.98, training accuracy is 0.468750, time 210.903 samples/sec
+epoch 6, total_step 102700, total loss is 21.25 , inference loss is 13.27, weight deacy loss is 7.98, training accuracy is 0.312500, time 209.927 samples/sec
+epoch 6, total_step 102720, total loss is 18.54 , inference loss is 10.56, weight deacy loss is 7.98, training accuracy is 0.218750, time 214.777 samples/sec
+epoch 6, total_step 102740, total loss is 20.61 , inference loss is 12.63, weight deacy loss is 7.98, training accuracy is 0.312500, time 211.201 samples/sec
+epoch 6, total_step 102760, total loss is 20.60 , inference loss is 12.63, weight deacy loss is 7.98, training accuracy is 0.187500, time 213.219 samples/sec
+epoch 6, total_step 102780, total loss is 18.44 , inference loss is 10.47, weight deacy loss is 7.98, training accuracy is 0.218750, time 217.835 samples/sec
+epoch 6, total_step 102800, total loss is 18.83 , inference loss is 10.85, weight deacy loss is 7.97, training accuracy is 0.343750, time 210.345 samples/sec
+epoch 6, total_step 102820, total loss is 16.73 , inference loss is 8.75, weight deacy loss is 7.97, training accuracy is 0.375000, time 209.588 samples/sec
+epoch 6, total_step 102840, total loss is 13.35 , inference loss is 5.38, weight deacy loss is 7.97, training accuracy is 0.437500, time 220.985 samples/sec
+epoch 6, total_step 102860, total loss is 15.48 , inference loss is 7.50, weight deacy loss is 7.97, training accuracy is 0.437500, time 207.285 samples/sec
+epoch 6, total_step 102880, total loss is 12.77 , inference loss is 4.80, weight deacy loss is 7.97, training accuracy is 0.468750, time 211.862 samples/sec
+epoch 6, total_step 102900, total loss is 14.04 , inference loss is 6.07, weight deacy loss is 7.97, training accuracy is 0.500000, time 14.247 samples/sec
+epoch 6, total_step 102920, total loss is 18.00 , inference loss is 10.03, weight deacy loss is 7.97, training accuracy is 0.250000, time 217.814 samples/sec
+epoch 6, total_step 102940, total loss is 23.26 , inference loss is 15.29, weight deacy loss is 7.97, training accuracy is 0.343750, time 221.665 samples/sec
+epoch 6, total_step 102960, total loss is 16.78 , inference loss is 8.81, weight deacy loss is 7.97, training accuracy is 0.406250, time 222.504 samples/sec
+epoch 6, total_step 102980, total loss is 18.76 , inference loss is 10.79, weight deacy loss is 7.97, training accuracy is 0.343750, time 206.826 samples/sec
+epoch 6, total_step 103000, total loss is 20.27 , inference loss is 12.30, weight deacy loss is 7.97, training accuracy is 0.281250, time 210.298 samples/sec
+epoch 6, total_step 103020, total loss is 21.59 , inference loss is 13.63, weight deacy loss is 7.97, training accuracy is 0.437500, time 204.941 samples/sec
+epoch 6, total_step 103040, total loss is 16.20 , inference loss is 8.24, weight deacy loss is 7.97, training accuracy is 0.375000, time 216.831 samples/sec
+epoch 6, total_step 103060, total loss is 16.72 , inference loss is 8.76, weight deacy loss is 7.97, training accuracy is 0.343750, time 209.820 samples/sec
+epoch 6, total_step 103080, total loss is 18.74 , inference loss is 10.77, weight deacy loss is 7.97, training accuracy is 0.312500, time 219.623 samples/sec
+epoch 6, total_step 103100, total loss is 16.88 , inference loss is 8.92, weight deacy loss is 7.96, training accuracy is 0.312500, time 209.665 samples/sec
+epoch 6, total_step 103120, total loss is 15.61 , inference loss is 7.65, weight deacy loss is 7.96, training accuracy is 0.281250, time 207.767 samples/sec
+epoch 6, total_step 103140, total loss is 17.16 , inference loss is 9.20, weight deacy loss is 7.96, training accuracy is 0.437500, time 212.165 samples/sec
+epoch 6, total_step 103160, total loss is 17.92 , inference loss is 9.96, weight deacy loss is 7.96, training accuracy is 0.250000, time 213.953 samples/sec
+epoch 6, total_step 103180, total loss is 19.69 , inference loss is 11.73, weight deacy loss is 7.96, training accuracy is 0.281250, time 298.373 samples/sec
+epoch 6, total_step 103200, total loss is 19.28 , inference loss is 11.32, weight deacy loss is 7.96, training accuracy is 0.250000, time 13.441 samples/sec
+epoch 6, total_step 103220, total loss is 16.42 , inference loss is 8.46, weight deacy loss is 7.96, training accuracy is 0.281250, time 212.591 samples/sec
+epoch 6, total_step 103240, total loss is 20.35 , inference loss is 12.39, weight deacy loss is 7.96, training accuracy is 0.187500, time 225.694 samples/sec
+epoch 6, total_step 103260, total loss is 19.39 , inference loss is 11.43, weight deacy loss is 7.96, training accuracy is 0.218750, time 210.207 samples/sec
+epoch 6, total_step 103280, total loss is 15.85 , inference loss is 7.89, weight deacy loss is 7.96, training accuracy is 0.468750, time 197.653 samples/sec
+epoch 6, total_step 103300, total loss is 18.27 , inference loss is 10.31, weight deacy loss is 7.96, training accuracy is 0.375000, time 215.624 samples/sec
+epoch 6, total_step 103320, total loss is 19.58 , inference loss is 11.62, weight deacy loss is 7.96, training accuracy is 0.312500, time 209.676 samples/sec
+epoch 6, total_step 103340, total loss is 22.29 , inference loss is 14.33, weight deacy loss is 7.96, training accuracy is 0.156250, time 211.239 samples/sec
+epoch 6, total_step 103360, total loss is 21.58 , inference loss is 13.63, weight deacy loss is 7.96, training accuracy is 0.156250, time 220.988 samples/sec
+epoch 6, total_step 103380, total loss is 17.38 , inference loss is 9.42, weight deacy loss is 7.95, training accuracy is 0.312500, time 219.471 samples/sec
+epoch 6, total_step 103400, total loss is 15.94 , inference loss is 7.99, weight deacy loss is 7.95, training accuracy is 0.343750, time 205.587 samples/sec
+epoch 6, total_step 103420, total loss is 17.33 , inference loss is 9.37, weight deacy loss is 7.95, training accuracy is 0.250000, time 209.174 samples/sec
+epoch 6, total_step 103440, total loss is 18.22 , inference loss is 10.27, weight deacy loss is 7.95, training accuracy is 0.156250, time 299.998 samples/sec
+epoch 6, total_step 103460, total loss is 24.98 , inference loss is 17.03, weight deacy loss is 7.95, training accuracy is 0.156250, time 212.470 samples/sec
+epoch 6, total_step 103480, total loss is 16.20 , inference loss is 8.25, weight deacy loss is 7.95, training accuracy is 0.312500, time 207.387 samples/sec
+epoch 6, total_step 103500, total loss is 17.60 , inference loss is 9.65, weight deacy loss is 7.95, training accuracy is 0.375000, time 10.954 samples/sec
+epoch 6, total_step 103520, total loss is 19.63 , inference loss is 11.68, weight deacy loss is 7.95, training accuracy is 0.437500, time 202.060 samples/sec
+epoch 6, total_step 103540, total loss is 17.83 , inference loss is 9.88, weight deacy loss is 7.95, training accuracy is 0.187500, time 223.887 samples/sec
+epoch 6, total_step 103560, total loss is 17.39 , inference loss is 9.44, weight deacy loss is 7.95, training accuracy is 0.343750, time 207.905 samples/sec
+epoch 6, total_step 103580, total loss is 16.58 , inference loss is 8.63, weight deacy loss is 7.95, training accuracy is 0.437500, time 216.122 samples/sec
+epoch 6, total_step 103600, total loss is 20.55 , inference loss is 12.60, weight deacy loss is 7.95, training accuracy is 0.281250, time 212.744 samples/sec
+epoch 6, total_step 103620, total loss is 21.19 , inference loss is 13.24, weight deacy loss is 7.95, training accuracy is 0.312500, time 201.029 samples/sec
+epoch 6, total_step 103640, total loss is 18.93 , inference loss is 10.98, weight deacy loss is 7.95, training accuracy is 0.187500, time 217.858 samples/sec
+epoch 6, total_step 103660, total loss is 17.86 , inference loss is 9.91, weight deacy loss is 7.94, training accuracy is 0.250000, time 210.269 samples/sec
+epoch 6, total_step 103680, total loss is 21.25 , inference loss is 13.30, weight deacy loss is 7.94, training accuracy is 0.187500, time 216.704 samples/sec
+epoch 6, total_step 103700, total loss is 16.80 , inference loss is 8.86, weight deacy loss is 7.94, training accuracy is 0.437500, time 183.598 samples/sec
+epoch 6, total_step 103720, total loss is 19.09 , inference loss is 11.15, weight deacy loss is 7.94, training accuracy is 0.187500, time 208.979 samples/sec
+epoch 6, total_step 103740, total loss is 20.82 , inference loss is 12.87, weight deacy loss is 7.94, training accuracy is 0.218750, time 199.784 samples/sec
+epoch 6, total_step 103760, total loss is 16.63 , inference loss is 8.69, weight deacy loss is 7.94, training accuracy is 0.343750, time 208.154 samples/sec
+epoch 6, total_step 103780, total loss is 14.93 , inference loss is 6.98, weight deacy loss is 7.94, training accuracy is 0.406250, time 206.459 samples/sec
+epoch 6, total_step 103800, total loss is 18.60 , inference loss is 10.66, weight deacy loss is 7.94, training accuracy is 0.187500, time 13.666 samples/sec
+epoch 6, total_step 103820, total loss is 20.50 , inference loss is 12.56, weight deacy loss is 7.94, training accuracy is 0.281250, time 204.603 samples/sec
+epoch 6, total_step 103840, total loss is 23.31 , inference loss is 15.37, weight deacy loss is 7.94, training accuracy is 0.156250, time 220.150 samples/sec
+epoch 6, total_step 103860, total loss is 16.45 , inference loss is 8.51, weight deacy loss is 7.94, training accuracy is 0.375000, time 215.274 samples/sec
+epoch 6, total_step 103880, total loss is 22.80 , inference loss is 14.87, weight deacy loss is 7.94, training accuracy is 0.156250, time 205.340 samples/sec
+epoch 6, total_step 103900, total loss is 16.72 , inference loss is 8.78, weight deacy loss is 7.94, training accuracy is 0.468750, time 222.145 samples/sec
+epoch 6, total_step 103920, total loss is 19.88 , inference loss is 11.94, weight deacy loss is 7.94, training accuracy is 0.250000, time 211.349 samples/sec
+epoch 6, total_step 103940, total loss is 18.50 , inference loss is 10.57, weight deacy loss is 7.94, training accuracy is 0.281250, time 212.732 samples/sec
+epoch 6, total_step 103960, total loss is 20.92 , inference loss is 12.99, weight deacy loss is 7.93, training accuracy is 0.375000, time 210.541 samples/sec
+epoch 6, total_step 103980, total loss is 14.95 , inference loss is 7.01, weight deacy loss is 7.93, training accuracy is 0.343750, time 296.243 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.89715400000003
+best_threshold_index 116 0.987037037037037
+best_threshold_index 117 0.9861111111111112
+best_threshold_index 117 0.9866666666666667
+best_threshold_index 108 0.987037037037037
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 117 0.9866666666666667
+best_threshold_index 117 0.9862962962962963
+best_threshold_index 117 0.9866666666666667
+best_threshold_index 117 0.9851851851851852
+best_threshold_index 117 0.9857407407407407
+[lfw][104000]XNorm: 21.685005
+[lfw][104000]Accuracy-Flip: 0.98483+-0.00845
+test accuracy is: 0.9848333333333334
+epoch 6, total_step 104000, total loss is 19.84 , inference loss is 11.91, weight deacy loss is 7.93, training accuracy is 0.250000, time 18.277 samples/sec
+epoch 6, total_step 104020, total loss is 22.33 , inference loss is 14.40, weight deacy loss is 7.93, training accuracy is 0.218750, time 217.363 samples/sec
+epoch 6, total_step 104040, total loss is 21.13 , inference loss is 13.19, weight deacy loss is 7.93, training accuracy is 0.125000, time 212.496 samples/sec
+epoch 6, total_step 104060, total loss is 19.24 , inference loss is 11.30, weight deacy loss is 7.93, training accuracy is 0.437500, time 216.087 samples/sec
+epoch 6, total_step 104080, total loss is 17.40 , inference loss is 9.47, weight deacy loss is 7.93, training accuracy is 0.250000, time 213.453 samples/sec
+epoch 6, total_step 104100, total loss is 19.62 , inference loss is 11.69, weight deacy loss is 7.93, training accuracy is 0.218750, time 14.943 samples/sec
+epoch 6, total_step 104120, total loss is 18.43 , inference loss is 10.50, weight deacy loss is 7.93, training accuracy is 0.250000, time 178.637 samples/sec
+epoch 6, total_step 104140, total loss is 20.04 , inference loss is 12.11, weight deacy loss is 7.93, training accuracy is 0.312500, time 209.520 samples/sec
+epoch 6, total_step 104160, total loss is 15.67 , inference loss is 7.74, weight deacy loss is 7.93, training accuracy is 0.187500, time 213.745 samples/sec
+epoch 6, total_step 104180, total loss is 15.66 , inference loss is 7.74, weight deacy loss is 7.93, training accuracy is 0.281250, time 218.811 samples/sec
+epoch 6, total_step 104200, total loss is 15.95 , inference loss is 8.03, weight deacy loss is 7.93, training accuracy is 0.312500, time 203.613 samples/sec
+epoch 6, total_step 104220, total loss is 19.80 , inference loss is 11.88, weight deacy loss is 7.93, training accuracy is 0.156250, time 298.211 samples/sec
+epoch 6, total_step 104240, total loss is 21.21 , inference loss is 13.28, weight deacy loss is 7.93, training accuracy is 0.281250, time 210.357 samples/sec
+epoch 6, total_step 104260, total loss is 22.31 , inference loss is 14.39, weight deacy loss is 7.92, training accuracy is 0.218750, time 200.723 samples/sec
+epoch 6, total_step 104280, total loss is 18.68 , inference loss is 10.75, weight deacy loss is 7.92, training accuracy is 0.375000, time 207.522 samples/sec
+epoch 6, total_step 104300, total loss is 23.48 , inference loss is 15.56, weight deacy loss is 7.92, training accuracy is 0.125000, time 224.434 samples/sec
+epoch 6, total_step 104320, total loss is 19.36 , inference loss is 11.44, weight deacy loss is 7.92, training accuracy is 0.281250, time 208.629 samples/sec
+epoch 6, total_step 104340, total loss is 18.21 , inference loss is 10.29, weight deacy loss is 7.92, training accuracy is 0.343750, time 205.138 samples/sec
+epoch 6, total_step 104360, total loss is 17.74 , inference loss is 9.82, weight deacy loss is 7.92, training accuracy is 0.250000, time 217.514 samples/sec
+epoch 6, total_step 104380, total loss is 19.44 , inference loss is 11.52, weight deacy loss is 7.92, training accuracy is 0.312500, time 215.797 samples/sec
+epoch 6, total_step 104400, total loss is 20.38 , inference loss is 12.46, weight deacy loss is 7.92, training accuracy is 0.250000, time 13.570 samples/sec
+epoch 6, total_step 104420, total loss is 17.90 , inference loss is 9.98, weight deacy loss is 7.92, training accuracy is 0.250000, time 209.815 samples/sec
+epoch 6, total_step 104440, total loss is 19.70 , inference loss is 11.79, weight deacy loss is 7.92, training accuracy is 0.281250, time 211.279 samples/sec
+epoch 6, total_step 104460, total loss is 18.66 , inference loss is 10.74, weight deacy loss is 7.92, training accuracy is 0.375000, time 209.200 samples/sec
+epoch 6, total_step 104480, total loss is 15.86 , inference loss is 7.94, weight deacy loss is 7.92, training accuracy is 0.437500, time 210.418 samples/sec
+epoch 6, total_step 104500, total loss is 22.55 , inference loss is 14.63, weight deacy loss is 7.92, training accuracy is 0.125000, time 298.377 samples/sec
+epoch 6, total_step 104520, total loss is 18.30 , inference loss is 10.39, weight deacy loss is 7.92, training accuracy is 0.281250, time 205.059 samples/sec
+epoch 6, total_step 104540, total loss is 17.43 , inference loss is 9.52, weight deacy loss is 7.91, training accuracy is 0.250000, time 210.961 samples/sec
+epoch 6, total_step 104560, total loss is 17.25 , inference loss is 9.34, weight deacy loss is 7.91, training accuracy is 0.312500, time 216.061 samples/sec
+epoch 6, total_step 104580, total loss is 17.92 , inference loss is 10.01, weight deacy loss is 7.91, training accuracy is 0.187500, time 204.317 samples/sec
+epoch 6, total_step 104600, total loss is 16.61 , inference loss is 8.70, weight deacy loss is 7.91, training accuracy is 0.375000, time 208.530 samples/sec
+epoch 6, total_step 104620, total loss is 16.17 , inference loss is 8.26, weight deacy loss is 7.91, training accuracy is 0.187500, time 207.786 samples/sec
+epoch 6, total_step 104640, total loss is 15.27 , inference loss is 7.36, weight deacy loss is 7.91, training accuracy is 0.406250, time 229.303 samples/sec
+epoch 6, total_step 104660, total loss is 15.79 , inference loss is 7.88, weight deacy loss is 7.91, training accuracy is 0.250000, time 202.915 samples/sec
+epoch 6, total_step 104680, total loss is 20.36 , inference loss is 12.45, weight deacy loss is 7.91, training accuracy is 0.187500, time 212.405 samples/sec
+epoch 6, total_step 104700, total loss is 22.02 , inference loss is 14.11, weight deacy loss is 7.91, training accuracy is 0.156250, time 13.318 samples/sec
+epoch 6, total_step 104720, total loss is 21.61 , inference loss is 13.70, weight deacy loss is 7.91, training accuracy is 0.281250, time 220.750 samples/sec
+epoch 6, total_step 104740, total loss is 17.16 , inference loss is 9.25, weight deacy loss is 7.91, training accuracy is 0.218750, time 212.425 samples/sec
+epoch 6, total_step 104760, total loss is 18.27 , inference loss is 10.37, weight deacy loss is 7.91, training accuracy is 0.312500, time 298.167 samples/sec
+epoch 6, total_step 104780, total loss is 17.12 , inference loss is 9.21, weight deacy loss is 7.91, training accuracy is 0.281250, time 202.561 samples/sec
+epoch 6, total_step 104800, total loss is 20.25 , inference loss is 12.34, weight deacy loss is 7.91, training accuracy is 0.312500, time 237.419 samples/sec
+epoch 6, total_step 104820, total loss is 18.05 , inference loss is 10.15, weight deacy loss is 7.90, training accuracy is 0.312500, time 226.492 samples/sec
+epoch 6, total_step 104840, total loss is 16.57 , inference loss is 8.66, weight deacy loss is 7.90, training accuracy is 0.312500, time 213.666 samples/sec
+epoch 6, total_step 104860, total loss is 18.41 , inference loss is 10.50, weight deacy loss is 7.90, training accuracy is 0.250000, time 209.423 samples/sec
+epoch 6, total_step 104880, total loss is 18.58 , inference loss is 10.67, weight deacy loss is 7.90, training accuracy is 0.312500, time 212.205 samples/sec
+epoch 6, total_step 104900, total loss is 19.58 , inference loss is 11.68, weight deacy loss is 7.90, training accuracy is 0.218750, time 220.369 samples/sec
+epoch 6, total_step 104920, total loss is 18.20 , inference loss is 10.30, weight deacy loss is 7.90, training accuracy is 0.343750, time 206.075 samples/sec
+epoch 6, total_step 104940, total loss is 15.27 , inference loss is 7.37, weight deacy loss is 7.90, training accuracy is 0.312500, time 226.637 samples/sec
+epoch 6, total_step 104960, total loss is 23.08 , inference loss is 15.18, weight deacy loss is 7.90, training accuracy is 0.156250, time 215.611 samples/sec
+epoch 6, total_step 104980, total loss is 19.36 , inference loss is 11.46, weight deacy loss is 7.90, training accuracy is 0.312500, time 214.670 samples/sec
+epoch 6, total_step 105000, total loss is 21.58 , inference loss is 13.68, weight deacy loss is 7.90, training accuracy is 0.218750, time 11.930 samples/sec
+epoch 6, total_step 105020, total loss is 14.86 , inference loss is 6.97, weight deacy loss is 7.90, training accuracy is 0.343750, time 299.929 samples/sec
+epoch 6, total_step 105040, total loss is 16.60 , inference loss is 8.70, weight deacy loss is 7.90, training accuracy is 0.312500, time 212.991 samples/sec
+epoch 6, total_step 105060, total loss is 20.46 , inference loss is 12.57, weight deacy loss is 7.90, training accuracy is 0.281250, time 204.384 samples/sec
+epoch 6, total_step 105080, total loss is 17.35 , inference loss is 9.46, weight deacy loss is 7.90, training accuracy is 0.187500, time 211.048 samples/sec
+epoch 6, total_step 105100, total loss is 22.33 , inference loss is 14.43, weight deacy loss is 7.90, training accuracy is 0.187500, time 218.642 samples/sec
+epoch 6, total_step 105120, total loss is 19.48 , inference loss is 11.58, weight deacy loss is 7.89, training accuracy is 0.250000, time 209.246 samples/sec
+epoch 6, total_step 105140, total loss is 22.61 , inference loss is 14.72, weight deacy loss is 7.89, training accuracy is 0.156250, time 212.844 samples/sec
+epoch 6, total_step 105160, total loss is 19.52 , inference loss is 11.63, weight deacy loss is 7.89, training accuracy is 0.312500, time 214.457 samples/sec
+epoch 6, total_step 105180, total loss is 19.33 , inference loss is 11.44, weight deacy loss is 7.89, training accuracy is 0.343750, time 209.659 samples/sec
+epoch 6, total_step 105200, total loss is 18.38 , inference loss is 10.49, weight deacy loss is 7.89, training accuracy is 0.312500, time 219.025 samples/sec
+epoch 6, total_step 105220, total loss is 19.93 , inference loss is 12.04, weight deacy loss is 7.89, training accuracy is 0.250000, time 223.726 samples/sec
+epoch 6, total_step 105240, total loss is 18.98 , inference loss is 11.09, weight deacy loss is 7.89, training accuracy is 0.187500, time 214.364 samples/sec
+epoch 6, total_step 105260, total loss is 18.93 , inference loss is 11.04, weight deacy loss is 7.89, training accuracy is 0.218750, time 213.022 samples/sec
+epoch 6, total_step 105280, total loss is 21.76 , inference loss is 13.87, weight deacy loss is 7.89, training accuracy is 0.125000, time 214.014 samples/sec
+epoch 6, total_step 105300, total loss is 17.24 , inference loss is 9.35, weight deacy loss is 7.89, training accuracy is 0.218750, time 15.469 samples/sec
+epoch 6, total_step 105320, total loss is 16.97 , inference loss is 9.08, weight deacy loss is 7.89, training accuracy is 0.125000, time 229.168 samples/sec
+epoch 6, total_step 105340, total loss is 19.21 , inference loss is 11.32, weight deacy loss is 7.89, training accuracy is 0.312500, time 212.378 samples/sec
+epoch 6, total_step 105360, total loss is 23.47 , inference loss is 15.59, weight deacy loss is 7.89, training accuracy is 0.187500, time 208.970 samples/sec
+epoch 6, total_step 105380, total loss is 17.45 , inference loss is 9.57, weight deacy loss is 7.89, training accuracy is 0.312500, time 216.073 samples/sec
+epoch 6, total_step 105400, total loss is 14.36 , inference loss is 6.48, weight deacy loss is 7.88, training accuracy is 0.281250, time 213.959 samples/sec
+epoch 6, total_step 105420, total loss is 18.54 , inference loss is 10.65, weight deacy loss is 7.88, training accuracy is 0.312500, time 209.612 samples/sec
+epoch 6, total_step 105440, total loss is 18.36 , inference loss is 10.48, weight deacy loss is 7.88, training accuracy is 0.343750, time 210.462 samples/sec
+epoch 6, total_step 105460, total loss is 16.98 , inference loss is 9.09, weight deacy loss is 7.88, training accuracy is 0.406250, time 220.757 samples/sec
+epoch 6, total_step 105480, total loss is 17.91 , inference loss is 10.03, weight deacy loss is 7.88, training accuracy is 0.406250, time 217.511 samples/sec
+epoch 6, total_step 105500, total loss is 17.37 , inference loss is 9.49, weight deacy loss is 7.88, training accuracy is 0.281250, time 207.916 samples/sec
+epoch 6, total_step 105520, total loss is 17.69 , inference loss is 9.81, weight deacy loss is 7.88, training accuracy is 0.343750, time 211.031 samples/sec
+epoch 6, total_step 105540, total loss is 20.59 , inference loss is 12.71, weight deacy loss is 7.88, training accuracy is 0.281250, time 203.308 samples/sec
+epoch 6, total_step 105560, total loss is 15.67 , inference loss is 7.79, weight deacy loss is 7.88, training accuracy is 0.437500, time 211.667 samples/sec
+epoch 6, total_step 105580, total loss is 16.37 , inference loss is 8.49, weight deacy loss is 7.88, training accuracy is 0.312500, time 211.866 samples/sec
+epoch 6, total_step 105600, total loss is 19.27 , inference loss is 11.39, weight deacy loss is 7.88, training accuracy is 0.281250, time 14.410 samples/sec
+epoch 6, total_step 105620, total loss is 24.22 , inference loss is 16.34, weight deacy loss is 7.88, training accuracy is 0.281250, time 222.200 samples/sec
+epoch 6, total_step 105640, total loss is 14.88 , inference loss is 7.01, weight deacy loss is 7.88, training accuracy is 0.375000, time 206.831 samples/sec
+epoch 6, total_step 105660, total loss is 16.03 , inference loss is 8.16, weight deacy loss is 7.88, training accuracy is 0.312500, time 215.564 samples/sec
+epoch 6, total_step 105680, total loss is 14.49 , inference loss is 6.61, weight deacy loss is 7.87, training accuracy is 0.343750, time 224.397 samples/sec
+epoch 6, total_step 105700, total loss is 21.67 , inference loss is 13.80, weight deacy loss is 7.87, training accuracy is 0.250000, time 199.334 samples/sec
+epoch 6, total_step 105720, total loss is 18.62 , inference loss is 10.74, weight deacy loss is 7.87, training accuracy is 0.156250, time 213.883 samples/sec
+epoch 6, total_step 105740, total loss is 18.76 , inference loss is 10.89, weight deacy loss is 7.87, training accuracy is 0.187500, time 211.843 samples/sec
+epoch 6, total_step 105760, total loss is 23.73 , inference loss is 15.86, weight deacy loss is 7.87, training accuracy is 0.156250, time 209.858 samples/sec
+epoch 6, total_step 105780, total loss is 17.91 , inference loss is 10.04, weight deacy loss is 7.87, training accuracy is 0.281250, time 242.254 samples/sec
+epoch 6, total_step 105800, total loss is 20.39 , inference loss is 12.52, weight deacy loss is 7.87, training accuracy is 0.093750, time 212.410 samples/sec
+epoch 6, total_step 105820, total loss is 18.21 , inference loss is 10.34, weight deacy loss is 7.87, training accuracy is 0.250000, time 215.459 samples/sec
+epoch 6, total_step 105840, total loss is 21.98 , inference loss is 14.11, weight deacy loss is 7.87, training accuracy is 0.281250, time 211.316 samples/sec
+epoch 6, total_step 105860, total loss is 20.46 , inference loss is 12.59, weight deacy loss is 7.87, training accuracy is 0.218750, time 216.334 samples/sec
+epoch 6, total_step 105880, total loss is 20.40 , inference loss is 12.53, weight deacy loss is 7.87, training accuracy is 0.218750, time 300.456 samples/sec
+epoch 6, total_step 105900, total loss is 19.44 , inference loss is 11.57, weight deacy loss is 7.87, training accuracy is 0.250000, time 13.611 samples/sec
+epoch 6, total_step 105920, total loss is 20.05 , inference loss is 12.18, weight deacy loss is 7.87, training accuracy is 0.156250, time 204.381 samples/sec
+epoch 6, total_step 105940, total loss is 22.54 , inference loss is 14.67, weight deacy loss is 7.87, training accuracy is 0.250000, time 215.129 samples/sec
+epoch 6, total_step 105960, total loss is 17.53 , inference loss is 9.66, weight deacy loss is 7.87, training accuracy is 0.156250, time 214.645 samples/sec
+epoch 6, total_step 105980, total loss is 20.92 , inference loss is 13.05, weight deacy loss is 7.86, training accuracy is 0.218750, time 212.771 samples/sec
+testing verification..
+(12000, 512)
+infer time 20.667990999999997
+best_threshold_index 107 0.9838888888888889
+best_threshold_index 104 0.9824074074074074
+best_threshold_index 107 0.9835185185185186
+best_threshold_index 104 0.9853703703703703
+best_threshold_index 104 0.9837037037037037
+best_threshold_index 105 0.9822222222222222
+best_threshold_index 104 0.9825925925925926
+best_threshold_index 107 0.9822222222222222
+best_threshold_index 107 0.9816666666666667
+best_threshold_index 104 0.9816666666666667
+[lfw][106000]XNorm: 23.818581
+[lfw][106000]Accuracy-Flip: 0.98200+-0.01002
+test accuracy is: 0.982
+epoch 6, total_step 106000, total loss is 16.87 , inference loss is 9.00, weight deacy loss is 7.86, training accuracy is 0.375000, time 16.437 samples/sec
+epoch 6, total_step 106020, total loss is 21.65 , inference loss is 13.79, weight deacy loss is 7.86, training accuracy is 0.281250, time 209.609 samples/sec
+epoch 6, total_step 106040, total loss is 20.58 , inference loss is 12.72, weight deacy loss is 7.86, training accuracy is 0.218750, time 213.702 samples/sec
+epoch 6, total_step 106060, total loss is 18.22 , inference loss is 10.36, weight deacy loss is 7.86, training accuracy is 0.250000, time 217.979 samples/sec
+epoch 6, total_step 106080, total loss is 16.68 , inference loss is 8.82, weight deacy loss is 7.86, training accuracy is 0.281250, time 216.684 samples/sec
+epoch 6, total_step 106100, total loss is 20.19 , inference loss is 12.33, weight deacy loss is 7.86, training accuracy is 0.281250, time 216.532 samples/sec
+epoch 6, total_step 106120, total loss is 17.17 , inference loss is 9.31, weight deacy loss is 7.86, training accuracy is 0.281250, time 215.052 samples/sec
+epoch 6, total_step 106140, total loss is 13.70 , inference loss is 5.84, weight deacy loss is 7.86, training accuracy is 0.406250, time 207.821 samples/sec
+epoch 6, total_step 106160, total loss is 16.42 , inference loss is 8.56, weight deacy loss is 7.86, training accuracy is 0.250000, time 298.369 samples/sec
+epoch 6, total_step 106180, total loss is 19.82 , inference loss is 11.97, weight deacy loss is 7.86, training accuracy is 0.312500, time 206.754 samples/sec
+epoch 6, total_step 106200, total loss is 17.75 , inference loss is 9.89, weight deacy loss is 7.86, training accuracy is 0.250000, time 13.562 samples/sec
+epoch 6, total_step 106220, total loss is 19.58 , inference loss is 11.73, weight deacy loss is 7.86, training accuracy is 0.250000, time 213.033 samples/sec
+epoch 6, total_step 106240, total loss is 19.24 , inference loss is 11.39, weight deacy loss is 7.86, training accuracy is 0.218750, time 209.048 samples/sec
+epoch 6, total_step 106260, total loss is 17.22 , inference loss is 9.36, weight deacy loss is 7.86, training accuracy is 0.406250, time 212.452 samples/sec
+epoch 6, total_step 106280, total loss is 16.42 , inference loss is 8.56, weight deacy loss is 7.85, training accuracy is 0.375000, time 201.826 samples/sec
+epoch 6, total_step 106300, total loss is 23.17 , inference loss is 15.32, weight deacy loss is 7.85, training accuracy is 0.281250, time 212.614 samples/sec
+epoch 6, total_step 106320, total loss is 19.90 , inference loss is 12.05, weight deacy loss is 7.85, training accuracy is 0.031250, time 218.069 samples/sec
+epoch 6, total_step 106340, total loss is 19.79 , inference loss is 11.94, weight deacy loss is 7.85, training accuracy is 0.125000, time 207.632 samples/sec
+epoch 6, total_step 106360, total loss is 15.56 , inference loss is 7.71, weight deacy loss is 7.85, training accuracy is 0.281250, time 212.654 samples/sec
+epoch 6, total_step 106380, total loss is 19.42 , inference loss is 11.56, weight deacy loss is 7.85, training accuracy is 0.218750, time 209.325 samples/sec
+epoch 6, total_step 106400, total loss is 20.02 , inference loss is 12.17, weight deacy loss is 7.85, training accuracy is 0.156250, time 214.341 samples/sec
+epoch 6, total_step 106420, total loss is 23.27 , inference loss is 15.42, weight deacy loss is 7.85, training accuracy is 0.125000, time 297.970 samples/sec
+epoch 6, total_step 106440, total loss is 21.87 , inference loss is 14.02, weight deacy loss is 7.85, training accuracy is 0.218750, time 221.329 samples/sec
+epoch 6, total_step 106460, total loss is 18.92 , inference loss is 11.07, weight deacy loss is 7.85, training accuracy is 0.250000, time 217.815 samples/sec
+epoch 6, total_step 106480, total loss is 18.55 , inference loss is 10.70, weight deacy loss is 7.85, training accuracy is 0.281250, time 214.278 samples/sec
+epoch 6, total_step 106500, total loss is 15.46 , inference loss is 7.61, weight deacy loss is 7.85, training accuracy is 0.250000, time 13.311 samples/sec
+epoch 6, total_step 106520, total loss is 21.61 , inference loss is 13.76, weight deacy loss is 7.85, training accuracy is 0.187500, time 211.924 samples/sec
+epoch 6, total_step 106540, total loss is 19.33 , inference loss is 11.48, weight deacy loss is 7.85, training accuracy is 0.125000, time 209.133 samples/sec
+epoch 6, total_step 106560, total loss is 19.46 , inference loss is 11.62, weight deacy loss is 7.85, training accuracy is 0.250000, time 207.225 samples/sec
+epoch 6, total_step 106580, total loss is 22.76 , inference loss is 14.91, weight deacy loss is 7.84, training accuracy is 0.125000, time 208.577 samples/sec
+epoch 6, total_step 106600, total loss is 22.83 , inference loss is 14.99, weight deacy loss is 7.84, training accuracy is 0.218750, time 207.762 samples/sec
+epoch 6, total_step 106620, total loss is 18.75 , inference loss is 10.90, weight deacy loss is 7.84, training accuracy is 0.250000, time 207.989 samples/sec
+epoch 6, total_step 106640, total loss is 15.56 , inference loss is 7.72, weight deacy loss is 7.84, training accuracy is 0.250000, time 213.801 samples/sec
+epoch 6, total_step 106660, total loss is 19.53 , inference loss is 11.69, weight deacy loss is 7.84, training accuracy is 0.156250, time 217.848 samples/sec
+epoch 6, total_step 106680, total loss is 15.53 , inference loss is 7.69, weight deacy loss is 7.84, training accuracy is 0.218750, time 299.552 samples/sec
+epoch 6, total_step 106700, total loss is 19.15 , inference loss is 11.31, weight deacy loss is 7.84, training accuracy is 0.281250, time 211.491 samples/sec
+epoch 6, total_step 106720, total loss is 19.57 , inference loss is 11.73, weight deacy loss is 7.84, training accuracy is 0.281250, time 202.100 samples/sec
+epoch 6, total_step 106740, total loss is 14.78 , inference loss is 6.94, weight deacy loss is 7.84, training accuracy is 0.156250, time 205.305 samples/sec
+epoch 6, total_step 106760, total loss is 20.91 , inference loss is 13.07, weight deacy loss is 7.84, training accuracy is 0.281250, time 226.054 samples/sec
+epoch 6, total_step 106780, total loss is 19.08 , inference loss is 11.25, weight deacy loss is 7.84, training accuracy is 0.187500, time 207.940 samples/sec
+epoch 6, total_step 106800, total loss is 16.20 , inference loss is 8.36, weight deacy loss is 7.84, training accuracy is 0.187500, time 11.728 samples/sec
+epoch 6, total_step 106820, total loss is 23.31 , inference loss is 15.48, weight deacy loss is 7.84, training accuracy is 0.250000, time 214.398 samples/sec
+epoch 6, total_step 106840, total loss is 18.45 , inference loss is 10.62, weight deacy loss is 7.84, training accuracy is 0.281250, time 217.853 samples/sec
+epoch 6, total_step 106860, total loss is 20.76 , inference loss is 12.92, weight deacy loss is 7.84, training accuracy is 0.281250, time 209.929 samples/sec
+epoch 6, total_step 106880, total loss is 20.03 , inference loss is 12.20, weight deacy loss is 7.83, training accuracy is 0.218750, time 231.916 samples/sec
+epoch 6, total_step 106900, total loss is 17.66 , inference loss is 9.83, weight deacy loss is 7.83, training accuracy is 0.343750, time 210.186 samples/sec
+epoch 6, total_step 106920, total loss is 18.63 , inference loss is 10.80, weight deacy loss is 7.83, training accuracy is 0.156250, time 219.650 samples/sec
+epoch 6, total_step 106940, total loss is 16.93 , inference loss is 9.09, weight deacy loss is 7.83, training accuracy is 0.187500, time 214.162 samples/sec
+epoch 6, total_step 106960, total loss is 21.32 , inference loss is 13.48, weight deacy loss is 7.83, training accuracy is 0.218750, time 296.964 samples/sec
+epoch 6, total_step 106980, total loss is 23.23 , inference loss is 15.40, weight deacy loss is 7.83, training accuracy is 0.125000, time 213.026 samples/sec
+epoch 6, total_step 107000, total loss is 17.08 , inference loss is 9.25, weight deacy loss is 7.83, training accuracy is 0.156250, time 213.434 samples/sec
+epoch 6, total_step 107020, total loss is 23.40 , inference loss is 15.57, weight deacy loss is 7.83, training accuracy is 0.125000, time 195.969 samples/sec
+epoch 6, total_step 107040, total loss is 27.31 , inference loss is 19.48, weight deacy loss is 7.83, training accuracy is 0.125000, time 233.212 samples/sec
+epoch 6, total_step 107060, total loss is 25.38 , inference loss is 17.55, weight deacy loss is 7.83, training accuracy is 0.187500, time 201.413 samples/sec
+epoch 6, total_step 107080, total loss is 18.51 , inference loss is 10.68, weight deacy loss is 7.83, training accuracy is 0.312500, time 175.609 samples/sec
+epoch 6, total_step 107100, total loss is 20.87 , inference loss is 13.04, weight deacy loss is 7.83, training accuracy is 0.156250, time 13.293 samples/sec
+epoch 6, total_step 107120, total loss is 24.66 , inference loss is 16.84, weight deacy loss is 7.83, training accuracy is 0.250000, time 200.424 samples/sec
+epoch 6, total_step 107140, total loss is 20.15 , inference loss is 12.32, weight deacy loss is 7.83, training accuracy is 0.250000, time 226.020 samples/sec
+epoch 6, total_step 107160, total loss is 19.19 , inference loss is 11.37, weight deacy loss is 7.83, training accuracy is 0.187500, time 177.118 samples/sec
+epoch 6, total_step 107180, total loss is 18.67 , inference loss is 10.85, weight deacy loss is 7.82, training accuracy is 0.218750, time 217.004 samples/sec
+epoch 6, total_step 107200, total loss is 23.04 , inference loss is 15.22, weight deacy loss is 7.82, training accuracy is 0.250000, time 241.482 samples/sec
+epoch 6, total_step 107220, total loss is 17.40 , inference loss is 9.58, weight deacy loss is 7.82, training accuracy is 0.281250, time 205.811 samples/sec
+epoch 6, total_step 107240, total loss is 20.67 , inference loss is 12.84, weight deacy loss is 7.82, training accuracy is 0.093750, time 226.347 samples/sec2022-12-01 07:16:54.714470: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 9024 of 10000
+2022-12-01 07:16:55.915096: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 6, total_step 107260, total loss is 19.36 , inference loss is 11.54, weight deacy loss is 7.82, training accuracy is 0.187500, time 194.474 samples/sec
+epoch 6, total_step 107280, total loss is 20.35 , inference loss is 12.53, weight deacy loss is 7.82, training accuracy is 0.250000, time 209.671 samples/sec
+epoch 6, total_step 107300, total loss is 18.02 , inference loss is 10.20, weight deacy loss is 7.82, training accuracy is 0.312500, time 296.484 samples/sec
+End of epoch 6
+epoch 7, total_step 107320, total loss is 30.03 , inference loss is 22.21, weight deacy loss is 7.82, training accuracy is 0.000000, time 219.256 samples/sec
+epoch 7, total_step 107340, total loss is 19.04 , inference loss is 11.22, weight deacy loss is 7.82, training accuracy is 0.375000, time 226.953 samples/sec
+epoch 7, total_step 107360, total loss is 14.33 , inference loss is 6.51, weight deacy loss is 7.82, training accuracy is 0.437500, time 207.672 samples/sec
+epoch 7, total_step 107380, total loss is 10.93 , inference loss is 3.11, weight deacy loss is 7.82, training accuracy is 0.687500, time 216.039 samples/sec
+epoch 7, total_step 107400, total loss is 12.23 , inference loss is 4.41, weight deacy loss is 7.82, training accuracy is 0.656250, time 14.322 samples/sec
+epoch 7, total_step 107420, total loss is 14.30 , inference loss is 6.48, weight deacy loss is 7.82, training accuracy is 0.593750, time 210.541 samples/sec
+epoch 7, total_step 107440, total loss is 11.86 , inference loss is 4.05, weight deacy loss is 7.82, training accuracy is 0.562500, time 207.999 samples/sec
+epoch 7, total_step 107460, total loss is 16.26 , inference loss is 8.45, weight deacy loss is 7.81, training accuracy is 0.406250, time 212.970 samples/sec
+epoch 7, total_step 107480, total loss is 13.69 , inference loss is 5.88, weight deacy loss is 7.81, training accuracy is 0.531250, time 201.623 samples/sec
+epoch 7, total_step 107500, total loss is 12.50 , inference loss is 4.69, weight deacy loss is 7.81, training accuracy is 0.687500, time 299.162 samples/sec
+epoch 7, total_step 107520, total loss is 18.97 , inference loss is 11.16, weight deacy loss is 7.81, training accuracy is 0.468750, time 213.485 samples/sec
+epoch 7, total_step 107540, total loss is 13.80 , inference loss is 5.99, weight deacy loss is 7.81, training accuracy is 0.562500, time 207.819 samples/sec
+epoch 7, total_step 107560, total loss is 14.53 , inference loss is 6.71, weight deacy loss is 7.81, training accuracy is 0.500000, time 210.432 samples/sec
+epoch 7, total_step 107580, total loss is 17.78 , inference loss is 9.97, weight deacy loss is 7.81, training accuracy is 0.468750, time 210.763 samples/sec
+epoch 7, total_step 107600, total loss is 16.16 , inference loss is 8.35, weight deacy loss is 7.81, training accuracy is 0.593750, time 215.907 samples/sec
+epoch 7, total_step 107620, total loss is 14.72 , inference loss is 6.91, weight deacy loss is 7.81, training accuracy is 0.468750, time 205.062 samples/sec
+epoch 7, total_step 107640, total loss is 14.95 , inference loss is 7.14, weight deacy loss is 7.81, training accuracy is 0.593750, time 208.001 samples/sec
+epoch 7, total_step 107660, total loss is 11.80 , inference loss is 3.99, weight deacy loss is 7.81, training accuracy is 0.593750, time 212.164 samples/sec
+epoch 7, total_step 107680, total loss is 16.62 , inference loss is 8.81, weight deacy loss is 7.81, training accuracy is 0.375000, time 211.955 samples/sec
+epoch 7, total_step 107700, total loss is 19.78 , inference loss is 11.98, weight deacy loss is 7.81, training accuracy is 0.343750, time 11.674 samples/sec
+epoch 7, total_step 107720, total loss is 16.49 , inference loss is 8.69, weight deacy loss is 7.81, training accuracy is 0.562500, time 214.126 samples/sec
+epoch 7, total_step 107740, total loss is 14.26 , inference loss is 6.45, weight deacy loss is 7.81, training accuracy is 0.375000, time 206.912 samples/sec
+epoch 7, total_step 107760, total loss is 12.86 , inference loss is 5.06, weight deacy loss is 7.80, training accuracy is 0.500000, time 298.842 samples/sec
+epoch 7, total_step 107780, total loss is 13.99 , inference loss is 6.19, weight deacy loss is 7.80, training accuracy is 0.406250, time 218.526 samples/sec
+epoch 7, total_step 107800, total loss is 15.20 , inference loss is 7.40, weight deacy loss is 7.80, training accuracy is 0.468750, time 214.597 samples/sec
+epoch 7, total_step 107820, total loss is 13.86 , inference loss is 6.06, weight deacy loss is 7.80, training accuracy is 0.687500, time 212.824 samples/sec
+epoch 7, total_step 107840, total loss is 13.21 , inference loss is 5.41, weight deacy loss is 7.80, training accuracy is 0.500000, time 214.249 samples/sec
+epoch 7, total_step 107860, total loss is 15.44 , inference loss is 7.64, weight deacy loss is 7.80, training accuracy is 0.531250, time 217.352 samples/sec
+epoch 7, total_step 107880, total loss is 15.01 , inference loss is 7.21, weight deacy loss is 7.80, training accuracy is 0.656250, time 215.140 samples/sec
+epoch 7, total_step 107900, total loss is 21.37 , inference loss is 13.57, weight deacy loss is 7.80, training accuracy is 0.406250, time 215.807 samples/sec
+epoch 7, total_step 107920, total loss is 22.50 , inference loss is 14.70, weight deacy loss is 7.80, training accuracy is 0.375000, time 298.621 samples/sec
+epoch 7, total_step 107940, total loss is 18.53 , inference loss is 10.73, weight deacy loss is 7.80, training accuracy is 0.375000, time 298.745 samples/sec
+epoch 7, total_step 107960, total loss is 20.21 , inference loss is 12.41, weight deacy loss is 7.80, training accuracy is 0.375000, time 238.847 samples/sec
+epoch 7, total_step 107980, total loss is 16.85 , inference loss is 9.05, weight deacy loss is 7.80, training accuracy is 0.562500, time 207.507 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.57679399999998
+best_threshold_index 113 0.9859259259259259
+best_threshold_index 113 0.9857407407407407
+best_threshold_index 113 0.985
+best_threshold_index 113 0.9866666666666667
+best_threshold_index 113 0.9859259259259259
+best_threshold_index 113 0.9851851851851852
+best_threshold_index 113 0.9851851851851852
+best_threshold_index 113 0.9861111111111112
+best_threshold_index 113 0.9848148148148148
+best_threshold_index 113 0.9844444444444445
+[lfw][108000]XNorm: 20.910872
+[lfw][108000]Accuracy-Flip: 0.98550+-0.00582
+test accuracy is: 0.9854999999999998
+epoch 7, total_step 108000, total loss is 14.22 , inference loss is 6.42, weight deacy loss is 7.80, training accuracy is 0.531250, time 18.350 samples/sec
+epoch 7, total_step 108020, total loss is 13.78 , inference loss is 5.98, weight deacy loss is 7.80, training accuracy is 0.531250, time 207.027 samples/sec
+epoch 7, total_step 108040, total loss is 18.27 , inference loss is 10.47, weight deacy loss is 7.80, training accuracy is 0.437500, time 297.012 samples/sec
+epoch 7, total_step 108060, total loss is 15.21 , inference loss is 7.42, weight deacy loss is 7.80, training accuracy is 0.500000, time 203.156 samples/sec
+epoch 7, total_step 108080, total loss is 17.68 , inference loss is 9.89, weight deacy loss is 7.80, training accuracy is 0.468750, time 209.997 samples/sec
+epoch 7, total_step 108100, total loss is 14.02 , inference loss is 6.22, weight deacy loss is 7.80, training accuracy is 0.593750, time 218.798 samples/sec
+epoch 7, total_step 108120, total loss is 17.05 , inference loss is 9.25, weight deacy loss is 7.80, training accuracy is 0.312500, time 210.956 samples/sec
+epoch 7, total_step 108140, total loss is 12.82 , inference loss is 5.03, weight deacy loss is 7.80, training accuracy is 0.531250, time 211.622 samples/sec
+epoch 7, total_step 108160, total loss is 14.92 , inference loss is 7.13, weight deacy loss is 7.79, training accuracy is 0.593750, time 193.067 samples/sec
+epoch 7, total_step 108180, total loss is 18.49 , inference loss is 10.70, weight deacy loss is 7.79, training accuracy is 0.406250, time 214.190 samples/sec
+epoch 7, total_step 108200, total loss is 16.51 , inference loss is 8.72, weight deacy loss is 7.79, training accuracy is 0.343750, time 218.299 samples/sec
+epoch 7, total_step 108220, total loss is 18.21 , inference loss is 10.42, weight deacy loss is 7.79, training accuracy is 0.375000, time 211.585 samples/sec
+epoch 7, total_step 108240, total loss is 23.51 , inference loss is 15.72, weight deacy loss is 7.79, training accuracy is 0.281250, time 199.998 samples/sec
+epoch 7, total_step 108260, total loss is 18.90 , inference loss is 11.11, weight deacy loss is 7.79, training accuracy is 0.437500, time 212.222 samples/sec
+epoch 7, total_step 108280, total loss is 15.87 , inference loss is 8.08, weight deacy loss is 7.79, training accuracy is 0.375000, time 217.045 samples/sec
+epoch 7, total_step 108300, total loss is 16.26 , inference loss is 8.47, weight deacy loss is 7.79, training accuracy is 0.468750, time 14.596 samples/sec
+epoch 7, total_step 108320, total loss is 16.44 , inference loss is 8.65, weight deacy loss is 7.79, training accuracy is 0.343750, time 211.787 samples/sec
+epoch 7, total_step 108340, total loss is 16.27 , inference loss is 8.48, weight deacy loss is 7.79, training accuracy is 0.500000, time 214.233 samples/sec
+epoch 7, total_step 108360, total loss is 12.79 , inference loss is 5.00, weight deacy loss is 7.79, training accuracy is 0.406250, time 213.267 samples/sec
+epoch 7, total_step 108380, total loss is 16.76 , inference loss is 8.97, weight deacy loss is 7.79, training accuracy is 0.468750, time 206.338 samples/sec
+epoch 7, total_step 108400, total loss is 14.64 , inference loss is 6.85, weight deacy loss is 7.79, training accuracy is 0.312500, time 211.896 samples/sec
+epoch 7, total_step 108420, total loss is 17.79 , inference loss is 10.00, weight deacy loss is 7.79, training accuracy is 0.468750, time 216.288 samples/sec
+epoch 7, total_step 108440, total loss is 16.94 , inference loss is 9.15, weight deacy loss is 7.79, training accuracy is 0.437500, time 211.340 samples/sec
+epoch 7, total_step 108460, total loss is 12.65 , inference loss is 4.86, weight deacy loss is 7.79, training accuracy is 0.531250, time 226.212 samples/sec
+epoch 7, total_step 108480, total loss is 17.04 , inference loss is 9.26, weight deacy loss is 7.79, training accuracy is 0.375000, time 217.466 samples/sec
+epoch 7, total_step 108500, total loss is 22.42 , inference loss is 14.63, weight deacy loss is 7.79, training accuracy is 0.187500, time 219.256 samples/sec
+epoch 7, total_step 108520, total loss is 16.74 , inference loss is 8.96, weight deacy loss is 7.79, training accuracy is 0.406250, time 207.076 samples/sec
+epoch 7, total_step 108540, total loss is 15.58 , inference loss is 7.80, weight deacy loss is 7.79, training accuracy is 0.468750, time 211.617 samples/sec
+epoch 7, total_step 108560, total loss is 17.68 , inference loss is 9.90, weight deacy loss is 7.78, training accuracy is 0.593750, time 212.551 samples/sec
+epoch 7, total_step 108580, total loss is 19.08 , inference loss is 11.29, weight deacy loss is 7.78, training accuracy is 0.250000, time 207.967 samples/sec
+epoch 7, total_step 108600, total loss is 14.66 , inference loss is 6.88, weight deacy loss is 7.78, training accuracy is 0.406250, time 13.492 samples/sec
+epoch 7, total_step 108620, total loss is 14.18 , inference loss is 6.39, weight deacy loss is 7.78, training accuracy is 0.468750, time 210.316 samples/sec
+epoch 7, total_step 108640, total loss is 16.27 , inference loss is 8.49, weight deacy loss is 7.78, training accuracy is 0.375000, time 212.997 samples/sec
+epoch 7, total_step 108660, total loss is 16.19 , inference loss is 8.40, weight deacy loss is 7.78, training accuracy is 0.500000, time 215.761 samples/sec
+epoch 7, total_step 108680, total loss is 16.83 , inference loss is 9.04, weight deacy loss is 7.78, training accuracy is 0.500000, time 218.294 samples/sec
+epoch 7, total_step 108700, total loss is 19.69 , inference loss is 11.91, weight deacy loss is 7.78, training accuracy is 0.406250, time 209.988 samples/sec
+epoch 7, total_step 108720, total loss is 17.35 , inference loss is 9.57, weight deacy loss is 7.78, training accuracy is 0.375000, time 206.297 samples/sec
+epoch 7, total_step 108740, total loss is 19.28 , inference loss is 11.50, weight deacy loss is 7.78, training accuracy is 0.406250, time 212.911 samples/sec
+epoch 7, total_step 108760, total loss is 20.37 , inference loss is 12.59, weight deacy loss is 7.78, training accuracy is 0.375000, time 210.096 samples/sec
+epoch 7, total_step 108780, total loss is 15.74 , inference loss is 7.96, weight deacy loss is 7.78, training accuracy is 0.437500, time 211.417 samples/sec
+epoch 7, total_step 108800, total loss is 14.97 , inference loss is 7.19, weight deacy loss is 7.78, training accuracy is 0.406250, time 211.917 samples/sec
+epoch 7, total_step 108820, total loss is 14.92 , inference loss is 7.14, weight deacy loss is 7.78, training accuracy is 0.468750, time 211.273 samples/sec
+epoch 7, total_step 108840, total loss is 19.17 , inference loss is 11.39, weight deacy loss is 7.78, training accuracy is 0.375000, time 192.123 samples/sec
+epoch 7, total_step 108860, total loss is 20.00 , inference loss is 12.22, weight deacy loss is 7.78, training accuracy is 0.343750, time 214.813 samples/sec
+epoch 7, total_step 108880, total loss is 15.51 , inference loss is 7.73, weight deacy loss is 7.78, training accuracy is 0.468750, time 298.945 samples/sec
+epoch 7, total_step 108900, total loss is 16.53 , inference loss is 8.76, weight deacy loss is 7.78, training accuracy is 0.468750, time 13.694 samples/sec
+epoch 7, total_step 108920, total loss is 14.58 , inference loss is 6.81, weight deacy loss is 7.78, training accuracy is 0.437500, time 213.784 samples/sec
+epoch 7, total_step 108940, total loss is 18.91 , inference loss is 11.13, weight deacy loss is 7.78, training accuracy is 0.375000, time 219.713 samples/sec
+epoch 7, total_step 108960, total loss is 19.75 , inference loss is 11.97, weight deacy loss is 7.77, training accuracy is 0.343750, time 214.505 samples/sec
+epoch 7, total_step 108980, total loss is 20.39 , inference loss is 12.62, weight deacy loss is 7.77, training accuracy is 0.375000, time 215.466 samples/sec
+epoch 7, total_step 109000, total loss is 14.87 , inference loss is 7.10, weight deacy loss is 7.77, training accuracy is 0.500000, time 208.345 samples/sec
+epoch 7, total_step 109020, total loss is 17.34 , inference loss is 9.57, weight deacy loss is 7.77, training accuracy is 0.343750, time 215.934 samples/sec
+epoch 7, total_step 109040, total loss is 15.69 , inference loss is 7.92, weight deacy loss is 7.77, training accuracy is 0.500000, time 202.682 samples/sec
+epoch 7, total_step 109060, total loss is 14.91 , inference loss is 7.14, weight deacy loss is 7.77, training accuracy is 0.562500, time 214.154 samples/sec
+epoch 7, total_step 109080, total loss is 16.84 , inference loss is 9.07, weight deacy loss is 7.77, training accuracy is 0.406250, time 223.467 samples/sec
+epoch 7, total_step 109100, total loss is 13.84 , inference loss is 6.07, weight deacy loss is 7.77, training accuracy is 0.406250, time 212.060 samples/sec
+epoch 7, total_step 109120, total loss is 16.27 , inference loss is 8.50, weight deacy loss is 7.77, training accuracy is 0.531250, time 216.372 samples/sec
+epoch 7, total_step 109140, total loss is 16.81 , inference loss is 9.04, weight deacy loss is 7.77, training accuracy is 0.343750, time 297.623 samples/sec
+epoch 7, total_step 109160, total loss is 16.35 , inference loss is 8.58, weight deacy loss is 7.77, training accuracy is 0.375000, time 195.393 samples/sec
+epoch 7, total_step 109180, total loss is 19.93 , inference loss is 12.16, weight deacy loss is 7.77, training accuracy is 0.406250, time 210.196 samples/sec
+epoch 7, total_step 109200, total loss is 17.14 , inference loss is 9.37, weight deacy loss is 7.77, training accuracy is 0.406250, time 13.483 samples/sec
+epoch 7, total_step 109220, total loss is 15.48 , inference loss is 7.72, weight deacy loss is 7.77, training accuracy is 0.437500, time 210.160 samples/sec
+epoch 7, total_step 109240, total loss is 18.25 , inference loss is 10.49, weight deacy loss is 7.77, training accuracy is 0.312500, time 210.287 samples/sec
+epoch 7, total_step 109260, total loss is 17.65 , inference loss is 9.88, weight deacy loss is 7.77, training accuracy is 0.312500, time 217.315 samples/sec
+epoch 7, total_step 109280, total loss is 12.92 , inference loss is 5.15, weight deacy loss is 7.77, training accuracy is 0.500000, time 219.133 samples/sec
+epoch 7, total_step 109300, total loss is 21.42 , inference loss is 13.66, weight deacy loss is 7.77, training accuracy is 0.406250, time 206.372 samples/sec
+epoch 7, total_step 109320, total loss is 19.26 , inference loss is 11.50, weight deacy loss is 7.76, training accuracy is 0.406250, time 212.120 samples/sec
+epoch 7, total_step 109340, total loss is 19.43 , inference loss is 11.67, weight deacy loss is 7.76, training accuracy is 0.437500, time 210.827 samples/sec
+epoch 7, total_step 109360, total loss is 16.17 , inference loss is 8.41, weight deacy loss is 7.76, training accuracy is 0.406250, time 223.357 samples/sec
+epoch 7, total_step 109380, total loss is 16.26 , inference loss is 8.50, weight deacy loss is 7.76, training accuracy is 0.312500, time 214.524 samples/sec
+epoch 7, total_step 109400, total loss is 18.62 , inference loss is 10.85, weight deacy loss is 7.76, training accuracy is 0.406250, time 295.442 samples/sec
+epoch 7, total_step 109420, total loss is 15.16 , inference loss is 7.40, weight deacy loss is 7.76, training accuracy is 0.406250, time 209.911 samples/sec
+epoch 7, total_step 109440, total loss is 13.06 , inference loss is 5.30, weight deacy loss is 7.76, training accuracy is 0.312500, time 205.490 samples/sec
+epoch 7, total_step 109460, total loss is 15.43 , inference loss is 7.67, weight deacy loss is 7.76, training accuracy is 0.437500, time 215.472 samples/sec
+epoch 7, total_step 109480, total loss is 16.59 , inference loss is 8.83, weight deacy loss is 7.76, training accuracy is 0.468750, time 213.254 samples/sec
+epoch 7, total_step 109500, total loss is 13.84 , inference loss is 6.08, weight deacy loss is 7.76, training accuracy is 0.531250, time 13.789 samples/sec
+epoch 7, total_step 109520, total loss is 15.67 , inference loss is 7.91, weight deacy loss is 7.76, training accuracy is 0.500000, time 205.593 samples/sec
+epoch 7, total_step 109540, total loss is 12.57 , inference loss is 4.81, weight deacy loss is 7.76, training accuracy is 0.593750, time 210.818 samples/sec
+epoch 7, total_step 109560, total loss is 16.46 , inference loss is 8.70, weight deacy loss is 7.76, training accuracy is 0.437500, time 213.497 samples/sec
+epoch 7, total_step 109580, total loss is 14.19 , inference loss is 6.44, weight deacy loss is 7.76, training accuracy is 0.375000, time 214.065 samples/sec
+epoch 7, total_step 109600, total loss is 19.22 , inference loss is 11.47, weight deacy loss is 7.76, training accuracy is 0.437500, time 214.010 samples/sec
+epoch 7, total_step 109620, total loss is 13.68 , inference loss is 5.92, weight deacy loss is 7.75, training accuracy is 0.500000, time 219.117 samples/sec
+epoch 7, total_step 109640, total loss is 14.85 , inference loss is 7.09, weight deacy loss is 7.75, training accuracy is 0.437500, time 215.148 samples/sec
+epoch 7, total_step 109660, total loss is 13.47 , inference loss is 5.72, weight deacy loss is 7.75, training accuracy is 0.531250, time 215.887 samples/sec
+epoch 7, total_step 109680, total loss is 13.11 , inference loss is 5.36, weight deacy loss is 7.75, training accuracy is 0.437500, time 209.907 samples/sec
+epoch 7, total_step 109700, total loss is 13.26 , inference loss is 5.51, weight deacy loss is 7.75, training accuracy is 0.593750, time 217.496 samples/sec
+epoch 7, total_step 109720, total loss is 13.57 , inference loss is 5.82, weight deacy loss is 7.75, training accuracy is 0.562500, time 300.431 samples/sec
+epoch 7, total_step 109740, total loss is 16.34 , inference loss is 8.59, weight deacy loss is 7.75, training accuracy is 0.437500, time 298.054 samples/sec
+epoch 7, total_step 109760, total loss is 19.61 , inference loss is 11.86, weight deacy loss is 7.75, training accuracy is 0.312500, time 212.821 samples/sec
+epoch 7, total_step 109780, total loss is 12.06 , inference loss is 4.31, weight deacy loss is 7.75, training accuracy is 0.625000, time 211.527 samples/sec
+epoch 7, total_step 109800, total loss is 13.71 , inference loss is 5.96, weight deacy loss is 7.75, training accuracy is 0.468750, time 11.180 samples/sec
+epoch 7, total_step 109820, total loss is 12.80 , inference loss is 5.05, weight deacy loss is 7.75, training accuracy is 0.531250, time 213.616 samples/sec
+epoch 7, total_step 109840, total loss is 15.04 , inference loss is 7.30, weight deacy loss is 7.75, training accuracy is 0.500000, time 213.147 samples/sec
+epoch 7, total_step 109860, total loss is 13.49 , inference loss is 5.75, weight deacy loss is 7.75, training accuracy is 0.656250, time 196.883 samples/sec
+epoch 7, total_step 109880, total loss is 11.56 , inference loss is 3.81, weight deacy loss is 7.75, training accuracy is 0.625000, time 211.825 samples/sec
+epoch 7, total_step 109900, total loss is 18.18 , inference loss is 10.43, weight deacy loss is 7.75, training accuracy is 0.281250, time 213.507 samples/sec
+epoch 7, total_step 109920, total loss is 11.95 , inference loss is 4.21, weight deacy loss is 7.74, training accuracy is 0.593750, time 219.191 samples/sec
+epoch 7, total_step 109940, total loss is 12.78 , inference loss is 5.04, weight deacy loss is 7.74, training accuracy is 0.593750, time 217.958 samples/sec
+epoch 7, total_step 109960, total loss is 14.24 , inference loss is 6.49, weight deacy loss is 7.74, training accuracy is 0.468750, time 215.836 samples/sec
+epoch 7, total_step 109980, total loss is 14.83 , inference loss is 7.08, weight deacy loss is 7.74, training accuracy is 0.625000, time 207.450 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.236351999999975
+best_threshold_index 116 0.9868518518518519
+best_threshold_index 115 0.9853703703703703
+best_threshold_index 115 0.9853703703703703
+best_threshold_index 116 0.9864814814814815
+best_threshold_index 118 0.9861111111111112
+best_threshold_index 116 0.9844444444444445
+best_threshold_index 114 0.985
+best_threshold_index 116 0.9848148148148148
+best_threshold_index 116 0.9846296296296296
+best_threshold_index 116 0.9844444444444445
+[lfw][110000]XNorm: 21.664058
+[lfw][110000]Accuracy-Flip: 0.98433+-0.00746
+test accuracy is: 0.9843333333333334
+epoch 7, total_step 110000, total loss is 12.84 , inference loss is 5.10, weight deacy loss is 7.74, training accuracy is 0.593750, time 17.857 samples/sec
+epoch 7, total_step 110020, total loss is 16.42 , inference loss is 8.67, weight deacy loss is 7.74, training accuracy is 0.343750, time 215.185 samples/sec
+epoch 7, total_step 110040, total loss is 13.80 , inference loss is 6.05, weight deacy loss is 7.74, training accuracy is 0.406250, time 210.848 samples/sec
+epoch 7, total_step 110060, total loss is 17.99 , inference loss is 10.25, weight deacy loss is 7.74, training accuracy is 0.375000, time 198.707 samples/sec
+epoch 7, total_step 110080, total loss is 12.92 , inference loss is 5.18, weight deacy loss is 7.74, training accuracy is 0.593750, time 214.631 samples/sec
+epoch 7, total_step 110100, total loss is 14.59 , inference loss is 6.85, weight deacy loss is 7.74, training accuracy is 0.531250, time 13.101 samples/sec
+epoch 7, total_step 110120, total loss is 12.04 , inference loss is 4.30, weight deacy loss is 7.74, training accuracy is 0.406250, time 216.827 samples/sec
+epoch 7, total_step 110140, total loss is 14.23 , inference loss is 6.49, weight deacy loss is 7.74, training accuracy is 0.562500, time 278.106 samples/sec
+epoch 7, total_step 110160, total loss is 15.51 , inference loss is 7.77, weight deacy loss is 7.74, training accuracy is 0.406250, time 203.969 samples/sec
+epoch 7, total_step 110180, total loss is 14.29 , inference loss is 6.55, weight deacy loss is 7.74, training accuracy is 0.593750, time 199.146 samples/sec
+epoch 7, total_step 110200, total loss is 12.38 , inference loss is 4.64, weight deacy loss is 7.74, training accuracy is 0.468750, time 206.974 samples/sec
+epoch 7, total_step 110220, total loss is 17.80 , inference loss is 10.06, weight deacy loss is 7.73, training accuracy is 0.500000, time 206.018 samples/sec
+epoch 7, total_step 110240, total loss is 13.24 , inference loss is 5.51, weight deacy loss is 7.73, training accuracy is 0.593750, time 219.644 samples/sec
+epoch 7, total_step 110260, total loss is 14.70 , inference loss is 6.96, weight deacy loss is 7.73, training accuracy is 0.593750, time 224.662 samples/sec
+epoch 7, total_step 110280, total loss is 13.70 , inference loss is 5.97, weight deacy loss is 7.73, training accuracy is 0.593750, time 223.239 samples/sec
+epoch 7, total_step 110300, total loss is 14.02 , inference loss is 6.28, weight deacy loss is 7.73, training accuracy is 0.437500, time 215.471 samples/sec
+epoch 7, total_step 110320, total loss is 14.34 , inference loss is 6.61, weight deacy loss is 7.73, training accuracy is 0.562500, time 211.017 samples/sec
+epoch 7, total_step 110340, total loss is 18.14 , inference loss is 10.41, weight deacy loss is 7.73, training accuracy is 0.375000, time 208.257 samples/sec
+epoch 7, total_step 110360, total loss is 14.21 , inference loss is 6.48, weight deacy loss is 7.73, training accuracy is 0.468750, time 208.377 samples/sec
+epoch 7, total_step 110380, total loss is 14.69 , inference loss is 6.96, weight deacy loss is 7.73, training accuracy is 0.531250, time 209.780 samples/sec
+epoch 7, total_step 110400, total loss is 13.54 , inference loss is 5.81, weight deacy loss is 7.73, training accuracy is 0.562500, time 13.616 samples/sec
+epoch 7, total_step 110420, total loss is 15.75 , inference loss is 8.02, weight deacy loss is 7.73, training accuracy is 0.437500, time 296.901 samples/sec
+epoch 7, total_step 110440, total loss is 17.28 , inference loss is 9.56, weight deacy loss is 7.73, training accuracy is 0.500000, time 214.669 samples/sec
+epoch 7, total_step 110460, total loss is 22.93 , inference loss is 15.20, weight deacy loss is 7.73, training accuracy is 0.375000, time 213.570 samples/sec
+epoch 7, total_step 110480, total loss is 19.30 , inference loss is 11.57, weight deacy loss is 7.73, training accuracy is 0.343750, time 225.356 samples/sec
+epoch 7, total_step 110500, total loss is 12.91 , inference loss is 5.18, weight deacy loss is 7.73, training accuracy is 0.593750, time 206.348 samples/sec
+epoch 7, total_step 110520, total loss is 18.62 , inference loss is 10.90, weight deacy loss is 7.73, training accuracy is 0.343750, time 220.082 samples/sec
+epoch 7, total_step 110540, total loss is 18.51 , inference loss is 10.78, weight deacy loss is 7.73, training accuracy is 0.343750, time 202.156 samples/sec
+epoch 7, total_step 110560, total loss is 18.68 , inference loss is 10.96, weight deacy loss is 7.72, training accuracy is 0.375000, time 207.329 samples/sec
+epoch 7, total_step 110580, total loss is 20.26 , inference loss is 12.53, weight deacy loss is 7.72, training accuracy is 0.343750, time 209.229 samples/sec
+epoch 7, total_step 110600, total loss is 14.97 , inference loss is 7.25, weight deacy loss is 7.72, training accuracy is 0.468750, time 212.967 samples/sec
+epoch 7, total_step 110620, total loss is 16.21 , inference loss is 8.49, weight deacy loss is 7.72, training accuracy is 0.593750, time 220.907 samples/sec
+epoch 7, total_step 110640, total loss is 16.64 , inference loss is 8.92, weight deacy loss is 7.72, training accuracy is 0.437500, time 221.042 samples/sec
+epoch 7, total_step 110660, total loss is 12.98 , inference loss is 5.26, weight deacy loss is 7.72, training accuracy is 0.531250, time 212.471 samples/sec
+epoch 7, total_step 110680, total loss is 18.46 , inference loss is 10.73, weight deacy loss is 7.72, training accuracy is 0.375000, time 198.982 samples/sec
+epoch 7, total_step 110700, total loss is 18.83 , inference loss is 11.11, weight deacy loss is 7.72, training accuracy is 0.406250, time 14.253 samples/sec
+epoch 7, total_step 110720, total loss is 17.97 , inference loss is 10.25, weight deacy loss is 7.72, training accuracy is 0.500000, time 210.212 samples/sec
+epoch 7, total_step 110740, total loss is 18.49 , inference loss is 10.77, weight deacy loss is 7.72, training accuracy is 0.375000, time 212.559 samples/sec
+epoch 7, total_step 110760, total loss is 18.89 , inference loss is 11.17, weight deacy loss is 7.72, training accuracy is 0.531250, time 195.193 samples/sec
+epoch 7, total_step 110780, total loss is 20.08 , inference loss is 12.36, weight deacy loss is 7.72, training accuracy is 0.406250, time 212.061 samples/sec
+epoch 7, total_step 110800, total loss is 20.46 , inference loss is 12.74, weight deacy loss is 7.72, training accuracy is 0.500000, time 209.129 samples/sec
+epoch 7, total_step 110820, total loss is 14.79 , inference loss is 7.07, weight deacy loss is 7.72, training accuracy is 0.406250, time 210.867 samples/sec
+epoch 7, total_step 110840, total loss is 15.63 , inference loss is 7.91, weight deacy loss is 7.72, training accuracy is 0.500000, time 227.138 samples/sec
+epoch 7, total_step 110860, total loss is 16.71 , inference loss is 9.00, weight deacy loss is 7.72, training accuracy is 0.343750, time 216.204 samples/sec
+epoch 7, total_step 110880, total loss is 15.06 , inference loss is 7.35, weight deacy loss is 7.72, training accuracy is 0.531250, time 222.762 samples/sec
+epoch 7, total_step 110900, total loss is 18.96 , inference loss is 11.24, weight deacy loss is 7.71, training accuracy is 0.531250, time 211.898 samples/sec
+epoch 7, total_step 110920, total loss is 16.14 , inference loss is 8.43, weight deacy loss is 7.71, training accuracy is 0.437500, time 210.513 samples/sec
+epoch 7, total_step 110940, total loss is 15.92 , inference loss is 8.20, weight deacy loss is 7.71, training accuracy is 0.500000, time 218.882 samples/sec
+epoch 7, total_step 110960, total loss is 21.13 , inference loss is 13.41, weight deacy loss is 7.71, training accuracy is 0.281250, time 208.493 samples/sec
+epoch 7, total_step 110980, total loss is 14.79 , inference loss is 7.08, weight deacy loss is 7.71, training accuracy is 0.343750, time 219.542 samples/sec
+epoch 7, total_step 111000, total loss is 17.56 , inference loss is 9.85, weight deacy loss is 7.71, training accuracy is 0.343750, time 13.602 samples/sec
+epoch 7, total_step 111020, total loss is 15.87 , inference loss is 8.16, weight deacy loss is 7.71, training accuracy is 0.375000, time 211.235 samples/sec
+epoch 7, total_step 111040, total loss is 17.41 , inference loss is 9.70, weight deacy loss is 7.71, training accuracy is 0.218750, time 206.904 samples/sec
+epoch 7, total_step 111060, total loss is 17.43 , inference loss is 9.72, weight deacy loss is 7.71, training accuracy is 0.500000, time 206.946 samples/sec
+epoch 7, total_step 111080, total loss is 19.37 , inference loss is 11.66, weight deacy loss is 7.71, training accuracy is 0.468750, time 207.460 samples/sec
+epoch 7, total_step 111100, total loss is 19.62 , inference loss is 11.91, weight deacy loss is 7.71, training accuracy is 0.343750, time 205.395 samples/sec
+epoch 7, total_step 111120, total loss is 15.46 , inference loss is 7.75, weight deacy loss is 7.71, training accuracy is 0.312500, time 204.682 samples/sec
+epoch 7, total_step 111140, total loss is 18.03 , inference loss is 10.32, weight deacy loss is 7.71, training accuracy is 0.437500, time 210.018 samples/sec
+epoch 7, total_step 111160, total loss is 22.14 , inference loss is 14.44, weight deacy loss is 7.71, training accuracy is 0.281250, time 210.190 samples/sec
+epoch 7, total_step 111180, total loss is 18.01 , inference loss is 10.30, weight deacy loss is 7.71, training accuracy is 0.437500, time 215.685 samples/sec
+epoch 7, total_step 111200, total loss is 17.19 , inference loss is 9.49, weight deacy loss is 7.71, training accuracy is 0.375000, time 215.067 samples/sec
+epoch 7, total_step 111220, total loss is 19.26 , inference loss is 11.56, weight deacy loss is 7.71, training accuracy is 0.218750, time 203.780 samples/sec
+epoch 7, total_step 111240, total loss is 16.82 , inference loss is 9.12, weight deacy loss is 7.71, training accuracy is 0.500000, time 212.061 samples/sec
+epoch 7, total_step 111260, total loss is 16.44 , inference loss is 8.73, weight deacy loss is 7.70, training accuracy is 0.375000, time 298.385 samples/sec
+epoch 7, total_step 111280, total loss is 19.12 , inference loss is 11.41, weight deacy loss is 7.70, training accuracy is 0.437500, time 205.436 samples/sec
+epoch 7, total_step 111300, total loss is 13.67 , inference loss is 5.96, weight deacy loss is 7.70, training accuracy is 0.468750, time 13.137 samples/sec
+epoch 7, total_step 111320, total loss is 19.31 , inference loss is 11.60, weight deacy loss is 7.70, training accuracy is 0.343750, time 252.835 samples/sec
+epoch 7, total_step 111340, total loss is 16.66 , inference loss is 8.96, weight deacy loss is 7.70, training accuracy is 0.375000, time 298.479 samples/sec
+epoch 7, total_step 111360, total loss is 19.44 , inference loss is 11.74, weight deacy loss is 7.70, training accuracy is 0.312500, time 219.314 samples/sec
+epoch 7, total_step 111380, total loss is 17.62 , inference loss is 9.92, weight deacy loss is 7.70, training accuracy is 0.281250, time 215.002 samples/sec
+epoch 7, total_step 111400, total loss is 15.30 , inference loss is 7.60, weight deacy loss is 7.70, training accuracy is 0.531250, time 219.918 samples/sec
+epoch 7, total_step 111420, total loss is 13.64 , inference loss is 5.94, weight deacy loss is 7.70, training accuracy is 0.437500, time 216.220 samples/sec
+epoch 7, total_step 111440, total loss is 16.55 , inference loss is 8.85, weight deacy loss is 7.70, training accuracy is 0.343750, time 216.137 samples/sec
+epoch 7, total_step 111460, total loss is 13.12 , inference loss is 5.42, weight deacy loss is 7.70, training accuracy is 0.468750, time 219.423 samples/sec
+epoch 7, total_step 111480, total loss is 19.46 , inference loss is 11.76, weight deacy loss is 7.70, training accuracy is 0.281250, time 215.922 samples/sec
+epoch 7, total_step 111500, total loss is 19.92 , inference loss is 12.23, weight deacy loss is 7.70, training accuracy is 0.250000, time 297.610 samples/sec
+epoch 7, total_step 111520, total loss is 16.73 , inference loss is 9.04, weight deacy loss is 7.70, training accuracy is 0.406250, time 295.647 samples/sec
+epoch 7, total_step 111540, total loss is 19.28 , inference loss is 11.58, weight deacy loss is 7.70, training accuracy is 0.218750, time 204.576 samples/sec
+epoch 7, total_step 111560, total loss is 15.24 , inference loss is 7.55, weight deacy loss is 7.70, training accuracy is 0.312500, time 210.092 samples/sec
+epoch 7, total_step 111580, total loss is 17.14 , inference loss is 9.45, weight deacy loss is 7.70, training accuracy is 0.343750, time 208.707 samples/sec
+epoch 7, total_step 111600, total loss is 14.90 , inference loss is 7.21, weight deacy loss is 7.69, training accuracy is 0.406250, time 11.787 samples/sec
+epoch 7, total_step 111620, total loss is 20.63 , inference loss is 12.94, weight deacy loss is 7.69, training accuracy is 0.218750, time 213.550 samples/sec
+epoch 7, total_step 111640, total loss is 19.57 , inference loss is 11.87, weight deacy loss is 7.69, training accuracy is 0.343750, time 203.953 samples/sec
+epoch 7, total_step 111660, total loss is 15.46 , inference loss is 7.77, weight deacy loss is 7.69, training accuracy is 0.406250, time 211.317 samples/sec
+epoch 7, total_step 111680, total loss is 14.94 , inference loss is 7.25, weight deacy loss is 7.69, training accuracy is 0.468750, time 299.333 samples/sec
+epoch 7, total_step 111700, total loss is 15.09 , inference loss is 7.40, weight deacy loss is 7.69, training accuracy is 0.375000, time 213.172 samples/sec
+epoch 7, total_step 111720, total loss is 13.48 , inference loss is 5.78, weight deacy loss is 7.69, training accuracy is 0.468750, time 220.715 samples/sec
+epoch 7, total_step 111740, total loss is 17.10 , inference loss is 9.41, weight deacy loss is 7.69, training accuracy is 0.437500, time 246.317 samples/sec
+epoch 7, total_step 111760, total loss is 21.80 , inference loss is 14.11, weight deacy loss is 7.69, training accuracy is 0.312500, time 212.562 samples/sec
+epoch 7, total_step 111780, total loss is 16.55 , inference loss is 8.86, weight deacy loss is 7.69, training accuracy is 0.406250, time 223.724 samples/sec
+epoch 7, total_step 111800, total loss is 17.98 , inference loss is 10.29, weight deacy loss is 7.69, training accuracy is 0.281250, time 214.837 samples/sec
+epoch 7, total_step 111820, total loss is 15.81 , inference loss is 8.12, weight deacy loss is 7.69, training accuracy is 0.437500, time 214.590 samples/sec
+epoch 7, total_step 111840, total loss is 18.56 , inference loss is 10.88, weight deacy loss is 7.69, training accuracy is 0.250000, time 233.135 samples/sec
+epoch 7, total_step 111860, total loss is 13.91 , inference loss is 6.23, weight deacy loss is 7.69, training accuracy is 0.468750, time 214.988 samples/sec
+epoch 7, total_step 111880, total loss is 15.81 , inference loss is 8.12, weight deacy loss is 7.69, training accuracy is 0.468750, time 216.954 samples/sec
+epoch 7, total_step 111900, total loss is 21.14 , inference loss is 13.46, weight deacy loss is 7.69, training accuracy is 0.281250, time 13.477 samples/sec
+epoch 7, total_step 111920, total loss is 14.60 , inference loss is 6.92, weight deacy loss is 7.69, training accuracy is 0.375000, time 214.960 samples/sec
+epoch 7, total_step 111940, total loss is 17.81 , inference loss is 10.12, weight deacy loss is 7.69, training accuracy is 0.343750, time 210.420 samples/sec
+epoch 7, total_step 111960, total loss is 17.34 , inference loss is 9.65, weight deacy loss is 7.68, training accuracy is 0.312500, time 297.613 samples/sec
+epoch 7, total_step 111980, total loss is 22.08 , inference loss is 14.39, weight deacy loss is 7.68, training accuracy is 0.343750, time 209.116 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.725526000000013
+best_threshold_index 114 0.9875925925925926
+best_threshold_index 113 0.9866666666666667
+best_threshold_index 113 0.9877777777777778
+best_threshold_index 111 0.9872222222222222
+best_threshold_index 113 0.9874074074074074
+best_threshold_index 120 0.9861111111111112
+best_threshold_index 113 0.987037037037037
+best_threshold_index 113 0.9862962962962963
+best_threshold_index 125 0.9857407407407407
+best_threshold_index 113 0.9861111111111112
+[lfw][112000]XNorm: 20.903893
+[lfw][112000]Accuracy-Flip: 0.98433+-0.00539
+test accuracy is: 0.9843333333333334
+epoch 7, total_step 112000, total loss is 16.30 , inference loss is 8.62, weight deacy loss is 7.68, training accuracy is 0.406250, time 19.422 samples/sec
+epoch 7, total_step 112020, total loss is 16.06 , inference loss is 8.38, weight deacy loss is 7.68, training accuracy is 0.343750, time 210.623 samples/sec
+epoch 7, total_step 112040, total loss is 17.03 , inference loss is 9.34, weight deacy loss is 7.68, training accuracy is 0.343750, time 226.819 samples/sec
+epoch 7, total_step 112060, total loss is 16.00 , inference loss is 8.32, weight deacy loss is 7.68, training accuracy is 0.468750, time 210.541 samples/sec
+epoch 7, total_step 112080, total loss is 19.27 , inference loss is 11.59, weight deacy loss is 7.68, training accuracy is 0.343750, time 297.415 samples/sec
+epoch 7, total_step 112100, total loss is 13.21 , inference loss is 5.53, weight deacy loss is 7.68, training accuracy is 0.718750, time 216.774 samples/sec
+epoch 7, total_step 112120, total loss is 14.41 , inference loss is 6.73, weight deacy loss is 7.68, training accuracy is 0.406250, time 221.622 samples/sec
+epoch 7, total_step 112140, total loss is 23.45 , inference loss is 15.77, weight deacy loss is 7.68, training accuracy is 0.312500, time 211.169 samples/sec
+epoch 7, total_step 112160, total loss is 18.73 , inference loss is 11.05, weight deacy loss is 7.68, training accuracy is 0.312500, time 211.399 samples/sec
+epoch 7, total_step 112180, total loss is 16.56 , inference loss is 8.88, weight deacy loss is 7.68, training accuracy is 0.250000, time 226.847 samples/sec
+epoch 7, total_step 112200, total loss is 18.98 , inference loss is 11.30, weight deacy loss is 7.68, training accuracy is 0.218750, time 14.799 samples/sec
+epoch 7, total_step 112220, total loss is 18.19 , inference loss is 10.51, weight deacy loss is 7.68, training accuracy is 0.343750, time 212.619 samples/sec
+epoch 7, total_step 112240, total loss is 21.57 , inference loss is 13.90, weight deacy loss is 7.68, training accuracy is 0.250000, time 213.782 samples/sec
+epoch 7, total_step 112260, total loss is 17.56 , inference loss is 9.88, weight deacy loss is 7.68, training accuracy is 0.375000, time 224.354 samples/sec
+epoch 7, total_step 112280, total loss is 21.80 , inference loss is 14.12, weight deacy loss is 7.68, training accuracy is 0.250000, time 213.563 samples/sec
+epoch 7, total_step 112300, total loss is 18.87 , inference loss is 11.19, weight deacy loss is 7.67, training accuracy is 0.343750, time 204.559 samples/sec
+epoch 7, total_step 112320, total loss is 16.28 , inference loss is 8.60, weight deacy loss is 7.67, training accuracy is 0.343750, time 209.626 samples/sec
+epoch 7, total_step 112340, total loss is 19.20 , inference loss is 11.53, weight deacy loss is 7.67, training accuracy is 0.250000, time 298.581 samples/sec
+epoch 7, total_step 112360, total loss is 15.78 , inference loss is 8.11, weight deacy loss is 7.67, training accuracy is 0.468750, time 177.748 samples/sec
+epoch 7, total_step 112380, total loss is 16.97 , inference loss is 9.30, weight deacy loss is 7.67, training accuracy is 0.312500, time 213.102 samples/sec
+epoch 7, total_step 112400, total loss is 15.42 , inference loss is 7.74, weight deacy loss is 7.67, training accuracy is 0.468750, time 230.575 samples/sec
+epoch 7, total_step 112420, total loss is 14.17 , inference loss is 6.50, weight deacy loss is 7.67, training accuracy is 0.593750, time 200.621 samples/sec
+epoch 7, total_step 112440, total loss is 19.88 , inference loss is 12.21, weight deacy loss is 7.67, training accuracy is 0.250000, time 215.032 samples/sec
+epoch 7, total_step 112460, total loss is 17.88 , inference loss is 10.21, weight deacy loss is 7.67, training accuracy is 0.406250, time 218.093 samples/sec
+epoch 7, total_step 112480, total loss is 14.81 , inference loss is 7.14, weight deacy loss is 7.67, training accuracy is 0.343750, time 215.981 samples/sec
+epoch 7, total_step 112500, total loss is 14.90 , inference loss is 7.23, weight deacy loss is 7.67, training accuracy is 0.468750, time 12.115 samples/sec
+epoch 7, total_step 112520, total loss is 16.35 , inference loss is 8.68, weight deacy loss is 7.67, training accuracy is 0.343750, time 209.844 samples/sec
+epoch 7, total_step 112540, total loss is 12.74 , inference loss is 5.07, weight deacy loss is 7.67, training accuracy is 0.406250, time 213.120 samples/sec
+epoch 7, total_step 112560, total loss is 16.75 , inference loss is 9.08, weight deacy loss is 7.67, training accuracy is 0.468750, time 204.881 samples/sec
+epoch 7, total_step 112580, total loss is 16.37 , inference loss is 8.70, weight deacy loss is 7.67, training accuracy is 0.500000, time 212.271 samples/sec
+epoch 7, total_step 112600, total loss is 15.07 , inference loss is 7.41, weight deacy loss is 7.67, training accuracy is 0.406250, time 208.568 samples/sec
+epoch 7, total_step 112620, total loss is 15.85 , inference loss is 8.18, weight deacy loss is 7.67, training accuracy is 0.500000, time 298.769 samples/sec
+epoch 7, total_step 112640, total loss is 19.14 , inference loss is 11.48, weight deacy loss is 7.66, training accuracy is 0.218750, time 211.377 samples/sec
+epoch 7, total_step 112660, total loss is 23.07 , inference loss is 15.41, weight deacy loss is 7.66, training accuracy is 0.187500, time 214.309 samples/sec
+epoch 7, total_step 112680, total loss is 14.54 , inference loss is 6.87, weight deacy loss is 7.66, training accuracy is 0.343750, time 212.722 samples/sec
+epoch 7, total_step 112700, total loss is 18.51 , inference loss is 10.84, weight deacy loss is 7.66, training accuracy is 0.250000, time 214.462 samples/sec
+epoch 7, total_step 112720, total loss is 16.51 , inference loss is 8.85, weight deacy loss is 7.66, training accuracy is 0.437500, time 221.273 samples/sec
+epoch 7, total_step 112740, total loss is 20.70 , inference loss is 13.04, weight deacy loss is 7.66, training accuracy is 0.218750, time 199.737 samples/sec
+epoch 7, total_step 112760, total loss is 14.13 , inference loss is 6.47, weight deacy loss is 7.66, training accuracy is 0.468750, time 208.153 samples/sec
+epoch 7, total_step 112780, total loss is 18.79 , inference loss is 11.13, weight deacy loss is 7.66, training accuracy is 0.343750, time 212.692 samples/sec
+epoch 7, total_step 112800, total loss is 17.84 , inference loss is 10.18, weight deacy loss is 7.66, training accuracy is 0.406250, time 13.825 samples/sec
+epoch 7, total_step 112820, total loss is 15.11 , inference loss is 7.45, weight deacy loss is 7.66, training accuracy is 0.343750, time 218.677 samples/sec
+epoch 7, total_step 112840, total loss is 15.57 , inference loss is 7.91, weight deacy loss is 7.66, training accuracy is 0.437500, time 210.513 samples/sec
+epoch 7, total_step 112860, total loss is 15.01 , inference loss is 7.35, weight deacy loss is 7.66, training accuracy is 0.406250, time 211.053 samples/sec
+epoch 7, total_step 112880, total loss is 14.72 , inference loss is 7.06, weight deacy loss is 7.66, training accuracy is 0.468750, time 298.604 samples/sec
+epoch 7, total_step 112900, total loss is 16.32 , inference loss is 8.66, weight deacy loss is 7.66, training accuracy is 0.406250, time 208.961 samples/sec
+epoch 7, total_step 112920, total loss is 18.02 , inference loss is 10.36, weight deacy loss is 7.66, training accuracy is 0.406250, time 221.242 samples/sec
+epoch 7, total_step 112940, total loss is 14.35 , inference loss is 6.70, weight deacy loss is 7.66, training accuracy is 0.468750, time 203.169 samples/sec
+epoch 7, total_step 112960, total loss is 16.78 , inference loss is 9.13, weight deacy loss is 7.66, training accuracy is 0.406250, time 207.915 samples/sec
+epoch 7, total_step 112980, total loss is 16.16 , inference loss is 8.51, weight deacy loss is 7.65, training accuracy is 0.187500, time 210.927 samples/sec
+epoch 7, total_step 113000, total loss is 14.13 , inference loss is 6.47, weight deacy loss is 7.65, training accuracy is 0.468750, time 215.773 samples/sec
+epoch 7, total_step 113020, total loss is 15.46 , inference loss is 7.81, weight deacy loss is 7.65, training accuracy is 0.343750, time 204.109 samples/sec
+epoch 7, total_step 113040, total loss is 16.25 , inference loss is 8.60, weight deacy loss is 7.65, training accuracy is 0.343750, time 215.534 samples/sec
+epoch 7, total_step 113060, total loss is 20.10 , inference loss is 12.45, weight deacy loss is 7.65, training accuracy is 0.312500, time 202.708 samples/sec
+epoch 7, total_step 113080, total loss is 12.69 , inference loss is 5.04, weight deacy loss is 7.65, training accuracy is 0.437500, time 218.658 samples/sec
+epoch 7, total_step 113100, total loss is 19.26 , inference loss is 11.61, weight deacy loss is 7.65, training accuracy is 0.343750, time 13.961 samples/sec
+epoch 7, total_step 113120, total loss is 20.48 , inference loss is 12.83, weight deacy loss is 7.65, training accuracy is 0.437500, time 212.887 samples/sec
+epoch 7, total_step 113140, total loss is 13.09 , inference loss is 5.44, weight deacy loss is 7.65, training accuracy is 0.468750, time 180.090 samples/sec
+epoch 7, total_step 113160, total loss is 15.06 , inference loss is 7.41, weight deacy loss is 7.65, training accuracy is 0.468750, time 213.582 samples/sec
+epoch 7, total_step 113180, total loss is 18.26 , inference loss is 10.61, weight deacy loss is 7.65, training accuracy is 0.406250, time 216.547 samples/sec
+epoch 7, total_step 113200, total loss is 23.18 , inference loss is 15.53, weight deacy loss is 7.65, training accuracy is 0.343750, time 213.991 samples/sec
+epoch 7, total_step 113220, total loss is 17.75 , inference loss is 10.10, weight deacy loss is 7.65, training accuracy is 0.250000, time 226.964 samples/sec
+epoch 7, total_step 113240, total loss is 19.03 , inference loss is 11.38, weight deacy loss is 7.65, training accuracy is 0.375000, time 218.369 samples/sec
+epoch 7, total_step 113260, total loss is 15.12 , inference loss is 7.47, weight deacy loss is 7.65, training accuracy is 0.468750, time 217.611 samples/sec
+epoch 7, total_step 113280, total loss is 15.49 , inference loss is 7.84, weight deacy loss is 7.65, training accuracy is 0.375000, time 213.345 samples/sec
+epoch 7, total_step 113300, total loss is 15.05 , inference loss is 7.41, weight deacy loss is 7.65, training accuracy is 0.343750, time 219.147 samples/sec
+epoch 7, total_step 113320, total loss is 14.23 , inference loss is 6.59, weight deacy loss is 7.64, training accuracy is 0.562500, time 301.571 samples/sec
+epoch 7, total_step 113340, total loss is 14.51 , inference loss is 6.86, weight deacy loss is 7.64, training accuracy is 0.437500, time 296.374 samples/sec
+epoch 7, total_step 113360, total loss is 18.73 , inference loss is 11.08, weight deacy loss is 7.64, training accuracy is 0.187500, time 212.311 samples/sec
+epoch 7, total_step 113380, total loss is 16.60 , inference loss is 8.96, weight deacy loss is 7.64, training accuracy is 0.406250, time 211.582 samples/sec
+epoch 7, total_step 113400, total loss is 21.21 , inference loss is 13.57, weight deacy loss is 7.64, training accuracy is 0.312500, time 11.992 samples/sec
+epoch 7, total_step 113420, total loss is 16.17 , inference loss is 8.53, weight deacy loss is 7.64, training accuracy is 0.312500, time 216.074 samples/sec
+epoch 7, total_step 113440, total loss is 14.05 , inference loss is 6.41, weight deacy loss is 7.64, training accuracy is 0.375000, time 217.083 samples/sec
+epoch 7, total_step 113460, total loss is 13.98 , inference loss is 6.33, weight deacy loss is 7.64, training accuracy is 0.562500, time 212.950 samples/sec
+epoch 7, total_step 113480, total loss is 15.45 , inference loss is 7.81, weight deacy loss is 7.64, training accuracy is 0.312500, time 205.443 samples/sec
+epoch 7, total_step 113500, total loss is 16.10 , inference loss is 8.46, weight deacy loss is 7.64, training accuracy is 0.343750, time 207.801 samples/sec
+epoch 7, total_step 113520, total loss is 16.76 , inference loss is 9.12, weight deacy loss is 7.64, training accuracy is 0.406250, time 199.605 samples/sec
+epoch 7, total_step 113540, total loss is 18.82 , inference loss is 11.18, weight deacy loss is 7.64, training accuracy is 0.250000, time 201.973 samples/sec
+epoch 7, total_step 113560, total loss is 15.83 , inference loss is 8.19, weight deacy loss is 7.64, training accuracy is 0.406250, time 214.860 samples/sec
+epoch 7, total_step 113580, total loss is 16.03 , inference loss is 8.39, weight deacy loss is 7.64, training accuracy is 0.437500, time 213.760 samples/sec
+epoch 7, total_step 113600, total loss is 14.79 , inference loss is 7.15, weight deacy loss is 7.64, training accuracy is 0.437500, time 298.834 samples/sec
+epoch 7, total_step 113620, total loss is 16.23 , inference loss is 8.60, weight deacy loss is 7.64, training accuracy is 0.281250, time 208.748 samples/sec
+epoch 7, total_step 113640, total loss is 14.20 , inference loss is 6.56, weight deacy loss is 7.64, training accuracy is 0.500000, time 211.659 samples/sec
+epoch 7, total_step 113660, total loss is 18.77 , inference loss is 11.14, weight deacy loss is 7.63, training accuracy is 0.312500, time 209.632 samples/sec
+epoch 7, total_step 113680, total loss is 12.54 , inference loss is 4.91, weight deacy loss is 7.63, training accuracy is 0.406250, time 208.642 samples/sec
+epoch 7, total_step 113700, total loss is 19.87 , inference loss is 12.23, weight deacy loss is 7.63, training accuracy is 0.281250, time 13.600 samples/sec
+epoch 7, total_step 113720, total loss is 17.71 , inference loss is 10.08, weight deacy loss is 7.63, training accuracy is 0.500000, time 211.513 samples/sec
+epoch 7, total_step 113740, total loss is 19.16 , inference loss is 11.53, weight deacy loss is 7.63, training accuracy is 0.187500, time 200.746 samples/sec
+epoch 7, total_step 113760, total loss is 14.40 , inference loss is 6.77, weight deacy loss is 7.63, training accuracy is 0.406250, time 197.314 samples/sec
+epoch 7, total_step 113780, total loss is 22.87 , inference loss is 15.24, weight deacy loss is 7.63, training accuracy is 0.218750, time 211.052 samples/sec
+epoch 7, total_step 113800, total loss is 15.42 , inference loss is 7.79, weight deacy loss is 7.63, training accuracy is 0.312500, time 210.556 samples/sec
+epoch 7, total_step 113820, total loss is 14.74 , inference loss is 7.11, weight deacy loss is 7.63, training accuracy is 0.375000, time 213.081 samples/sec
+epoch 7, total_step 113840, total loss is 21.81 , inference loss is 14.18, weight deacy loss is 7.63, training accuracy is 0.187500, time 213.930 samples/sec
+epoch 7, total_step 113860, total loss is 13.91 , inference loss is 6.29, weight deacy loss is 7.63, training accuracy is 0.406250, time 204.596 samples/sec
+epoch 7, total_step 113880, total loss is 20.39 , inference loss is 12.76, weight deacy loss is 7.63, training accuracy is 0.218750, time 298.674 samples/sec
+epoch 7, total_step 113900, total loss is 18.99 , inference loss is 11.37, weight deacy loss is 7.63, training accuracy is 0.218750, time 201.845 samples/sec
+epoch 7, total_step 113920, total loss is 16.81 , inference loss is 9.19, weight deacy loss is 7.63, training accuracy is 0.375000, time 217.728 samples/sec
+epoch 7, total_step 113940, total loss is 17.59 , inference loss is 9.96, weight deacy loss is 7.63, training accuracy is 0.343750, time 213.994 samples/sec
+epoch 7, total_step 113960, total loss is 20.08 , inference loss is 12.46, weight deacy loss is 7.63, training accuracy is 0.375000, time 209.190 samples/sec
+epoch 7, total_step 113980, total loss is 16.69 , inference loss is 9.07, weight deacy loss is 7.63, training accuracy is 0.281250, time 219.136 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.383555999999963
+best_threshold_index 112 0.9877777777777778
+best_threshold_index 112 0.9866666666666667
+best_threshold_index 112 0.987037037037037
+best_threshold_index 112 0.9881481481481481
+best_threshold_index 112 0.987037037037037
+best_threshold_index 112 0.9862962962962963
+best_threshold_index 112 0.9874074074074074
+best_threshold_index 112 0.9872222222222222
+best_threshold_index 112 0.9859259259259259
+best_threshold_index 112 0.9864814814814815
+[lfw][114000]XNorm: 22.818682
+[lfw][114000]Accuracy-Flip: 0.98700+-0.00581
+test accuracy is: 0.9870000000000001
+epoch 7, total_step 114000, total loss is 19.06 , inference loss is 11.44, weight deacy loss is 7.62, training accuracy is 0.343750, time 19.028 samples/sec
+epoch 7, total_step 114020, total loss is 16.49 , inference loss is 8.87, weight deacy loss is 7.62, training accuracy is 0.218750, time 209.691 samples/sec
+epoch 7, total_step 114040, total loss is 16.25 , inference loss is 8.62, weight deacy loss is 7.62, training accuracy is 0.468750, time 205.485 samples/sec
+epoch 7, total_step 114060, total loss is 24.02 , inference loss is 16.39, weight deacy loss is 7.62, training accuracy is 0.312500, time 199.494 samples/sec
+epoch 7, total_step 114080, total loss is 18.38 , inference loss is 10.76, weight deacy loss is 7.62, training accuracy is 0.375000, time 216.517 samples/sec
+epoch 7, total_step 114100, total loss is 17.47 , inference loss is 9.85, weight deacy loss is 7.62, training accuracy is 0.250000, time 206.081 samples/sec
+epoch 7, total_step 114120, total loss is 15.05 , inference loss is 7.43, weight deacy loss is 7.62, training accuracy is 0.375000, time 207.710 samples/sec
+epoch 7, total_step 114140, total loss is 16.39 , inference loss is 8.77, weight deacy loss is 7.62, training accuracy is 0.312500, time 207.177 samples/sec
+epoch 7, total_step 114160, total loss is 17.39 , inference loss is 9.77, weight deacy loss is 7.62, training accuracy is 0.281250, time 208.332 samples/sec
+epoch 7, total_step 114180, total loss is 13.47 , inference loss is 5.85, weight deacy loss is 7.62, training accuracy is 0.406250, time 213.840 samples/sec
+epoch 7, total_step 114200, total loss is 15.83 , inference loss is 8.21, weight deacy loss is 7.62, training accuracy is 0.281250, time 214.730 samples/sec
+epoch 7, total_step 114220, total loss is 12.54 , inference loss is 4.92, weight deacy loss is 7.62, training accuracy is 0.531250, time 212.466 samples/sec
+epoch 7, total_step 114240, total loss is 16.00 , inference loss is 8.38, weight deacy loss is 7.62, training accuracy is 0.343750, time 207.609 samples/sec
+epoch 7, total_step 114260, total loss is 16.53 , inference loss is 8.91, weight deacy loss is 7.62, training accuracy is 0.312500, time 204.398 samples/sec
+epoch 7, total_step 114280, total loss is 17.95 , inference loss is 10.34, weight deacy loss is 7.62, training accuracy is 0.281250, time 301.687 samples/sec
+epoch 7, total_step 114300, total loss is 18.27 , inference loss is 10.65, weight deacy loss is 7.62, training accuracy is 0.187500, time 13.559 samples/sec
+epoch 7, total_step 114320, total loss is 13.75 , inference loss is 6.13, weight deacy loss is 7.62, training accuracy is 0.406250, time 210.245 samples/sec
+epoch 7, total_step 114340, total loss is 14.76 , inference loss is 7.15, weight deacy loss is 7.61, training accuracy is 0.343750, time 220.181 samples/sec
+epoch 7, total_step 114360, total loss is 20.66 , inference loss is 13.04, weight deacy loss is 7.61, training accuracy is 0.156250, time 206.478 samples/sec
+epoch 7, total_step 114380, total loss is 18.09 , inference loss is 10.47, weight deacy loss is 7.61, training accuracy is 0.281250, time 207.681 samples/sec
+epoch 7, total_step 114400, total loss is 18.79 , inference loss is 11.17, weight deacy loss is 7.61, training accuracy is 0.312500, time 204.583 samples/sec
+epoch 7, total_step 114420, total loss is 22.88 , inference loss is 15.26, weight deacy loss is 7.61, training accuracy is 0.437500, time 212.297 samples/sec
+epoch 7, total_step 114440, total loss is 15.72 , inference loss is 8.10, weight deacy loss is 7.61, training accuracy is 0.437500, time 208.061 samples/sec
+epoch 7, total_step 114460, total loss is 16.69 , inference loss is 9.08, weight deacy loss is 7.61, training accuracy is 0.437500, time 207.985 samples/sec
+epoch 7, total_step 114480, total loss is 18.86 , inference loss is 11.25, weight deacy loss is 7.61, training accuracy is 0.281250, time 211.879 samples/sec
+epoch 7, total_step 114500, total loss is 16.84 , inference loss is 9.23, weight deacy loss is 7.61, training accuracy is 0.250000, time 217.523 samples/sec
+epoch 7, total_step 114520, total loss is 21.07 , inference loss is 13.47, weight deacy loss is 7.61, training accuracy is 0.312500, time 209.263 samples/sec
+epoch 7, total_step 114540, total loss is 18.76 , inference loss is 11.15, weight deacy loss is 7.61, training accuracy is 0.406250, time 301.734 samples/sec
+epoch 7, total_step 114560, total loss is 14.42 , inference loss is 6.81, weight deacy loss is 7.61, training accuracy is 0.281250, time 207.632 samples/sec
+epoch 7, total_step 114580, total loss is 17.55 , inference loss is 9.94, weight deacy loss is 7.61, training accuracy is 0.437500, time 213.137 samples/sec
+epoch 7, total_step 114600, total loss is 16.41 , inference loss is 8.80, weight deacy loss is 7.61, training accuracy is 0.343750, time 12.370 samples/sec
+epoch 7, total_step 114620, total loss is 15.10 , inference loss is 7.50, weight deacy loss is 7.61, training accuracy is 0.437500, time 211.095 samples/sec
+epoch 7, total_step 114640, total loss is 14.76 , inference loss is 7.16, weight deacy loss is 7.61, training accuracy is 0.343750, time 213.060 samples/sec
+epoch 7, total_step 114660, total loss is 17.40 , inference loss is 9.80, weight deacy loss is 7.61, training accuracy is 0.218750, time 222.721 samples/sec
+epoch 7, total_step 114680, total loss is 19.41 , inference loss is 11.81, weight deacy loss is 7.60, training accuracy is 0.406250, time 215.335 samples/sec
+epoch 7, total_step 114700, total loss is 14.35 , inference loss is 6.75, weight deacy loss is 7.60, training accuracy is 0.437500, time 220.829 samples/sec
+epoch 7, total_step 114720, total loss is 18.98 , inference loss is 11.37, weight deacy loss is 7.60, training accuracy is 0.343750, time 206.161 samples/sec
+epoch 7, total_step 114740, total loss is 18.99 , inference loss is 11.39, weight deacy loss is 7.60, training accuracy is 0.250000, time 240.387 samples/sec
+epoch 7, total_step 114760, total loss is 20.31 , inference loss is 12.70, weight deacy loss is 7.60, training accuracy is 0.218750, time 210.965 samples/sec
+epoch 7, total_step 114780, total loss is 18.68 , inference loss is 11.08, weight deacy loss is 7.60, training accuracy is 0.343750, time 211.715 samples/sec
+epoch 7, total_step 114800, total loss is 16.85 , inference loss is 9.25, weight deacy loss is 7.60, training accuracy is 0.281250, time 210.840 samples/sec
+epoch 7, total_step 114820, total loss is 16.87 , inference loss is 9.27, weight deacy loss is 7.60, training accuracy is 0.343750, time 194.909 samples/sec
+epoch 7, total_step 114840, total loss is 19.54 , inference loss is 11.94, weight deacy loss is 7.60, training accuracy is 0.218750, time 210.815 samples/sec
+epoch 7, total_step 114860, total loss is 17.83 , inference loss is 10.23, weight deacy loss is 7.60, training accuracy is 0.312500, time 209.072 samples/sec
+epoch 7, total_step 114880, total loss is 14.36 , inference loss is 6.76, weight deacy loss is 7.60, training accuracy is 0.375000, time 207.243 samples/sec
+epoch 7, total_step 114900, total loss is 15.43 , inference loss is 7.83, weight deacy loss is 7.60, training accuracy is 0.312500, time 13.798 samples/sec
+epoch 7, total_step 114920, total loss is 17.30 , inference loss is 9.70, weight deacy loss is 7.60, training accuracy is 0.406250, time 212.020 samples/sec
+epoch 7, total_step 114940, total loss is 16.12 , inference loss is 8.53, weight deacy loss is 7.60, training accuracy is 0.343750, time 210.782 samples/sec
+epoch 7, total_step 114960, total loss is 14.25 , inference loss is 6.65, weight deacy loss is 7.60, training accuracy is 0.281250, time 217.217 samples/sec
+epoch 7, total_step 114980, total loss is 17.05 , inference loss is 9.46, weight deacy loss is 7.60, training accuracy is 0.468750, time 213.819 samples/sec
+epoch 7, total_step 115000, total loss is 19.55 , inference loss is 11.95, weight deacy loss is 7.60, training accuracy is 0.406250, time 216.842 samples/sec
+epoch 7, total_step 115020, total loss is 18.32 , inference loss is 10.72, weight deacy loss is 7.59, training accuracy is 0.406250, time 219.574 samples/sec
+epoch 7, total_step 115040, total loss is 16.56 , inference loss is 8.96, weight deacy loss is 7.59, training accuracy is 0.343750, time 232.519 samples/sec
+epoch 7, total_step 115060, total loss is 14.15 , inference loss is 6.55, weight deacy loss is 7.59, training accuracy is 0.500000, time 215.613 samples/sec
+epoch 7, total_step 115080, total loss is 18.55 , inference loss is 10.96, weight deacy loss is 7.59, training accuracy is 0.375000, time 216.037 samples/sec
+epoch 7, total_step 115100, total loss is 16.96 , inference loss is 9.37, weight deacy loss is 7.59, training accuracy is 0.375000, time 223.820 samples/sec
+epoch 7, total_step 115120, total loss is 15.91 , inference loss is 8.32, weight deacy loss is 7.59, training accuracy is 0.437500, time 301.534 samples/sec
+epoch 7, total_step 115140, total loss is 14.69 , inference loss is 7.09, weight deacy loss is 7.59, training accuracy is 0.562500, time 300.711 samples/sec
+epoch 7, total_step 115160, total loss is 17.92 , inference loss is 10.33, weight deacy loss is 7.59, training accuracy is 0.312500, time 210.504 samples/sec
+epoch 7, total_step 115180, total loss is 16.39 , inference loss is 8.80, weight deacy loss is 7.59, training accuracy is 0.343750, time 212.472 samples/sec
+epoch 7, total_step 115200, total loss is 12.34 , inference loss is 4.75, weight deacy loss is 7.59, training accuracy is 0.656250, time 14.182 samples/sec
+epoch 7, total_step 115220, total loss is 14.77 , inference loss is 7.19, weight deacy loss is 7.59, training accuracy is 0.281250, time 210.654 samples/sec
+epoch 7, total_step 115240, total loss is 18.46 , inference loss is 10.87, weight deacy loss is 7.59, training accuracy is 0.343750, time 214.113 samples/sec
+epoch 7, total_step 115260, total loss is 18.02 , inference loss is 10.43, weight deacy loss is 7.59, training accuracy is 0.218750, time 223.634 samples/sec
+epoch 7, total_step 115280, total loss is 17.76 , inference loss is 10.18, weight deacy loss is 7.59, training accuracy is 0.375000, time 208.409 samples/sec
+epoch 7, total_step 115300, total loss is 14.17 , inference loss is 6.58, weight deacy loss is 7.59, training accuracy is 0.343750, time 207.913 samples/sec
+epoch 7, total_step 115320, total loss is 15.24 , inference loss is 7.66, weight deacy loss is 7.59, training accuracy is 0.406250, time 207.864 samples/sec
+epoch 7, total_step 115340, total loss is 16.09 , inference loss is 8.51, weight deacy loss is 7.58, training accuracy is 0.406250, time 213.883 samples/sec
+epoch 7, total_step 115360, total loss is 16.83 , inference loss is 9.25, weight deacy loss is 7.58, training accuracy is 0.406250, time 214.361 samples/sec
+epoch 7, total_step 115380, total loss is 18.41 , inference loss is 10.83, weight deacy loss is 7.58, training accuracy is 0.406250, time 215.545 samples/sec
+epoch 7, total_step 115400, total loss is 18.98 , inference loss is 11.40, weight deacy loss is 7.58, training accuracy is 0.406250, time 224.760 samples/sec
+epoch 7, total_step 115420, total loss is 17.59 , inference loss is 10.01, weight deacy loss is 7.58, training accuracy is 0.375000, time 214.249 samples/sec
+epoch 7, total_step 115440, total loss is 13.52 , inference loss is 5.94, weight deacy loss is 7.58, training accuracy is 0.406250, time 210.119 samples/sec
+epoch 7, total_step 115460, total loss is 13.03 , inference loss is 5.45, weight deacy loss is 7.58, training accuracy is 0.406250, time 214.980 samples/sec
+epoch 7, total_step 115480, total loss is 17.95 , inference loss is 10.37, weight deacy loss is 7.58, training accuracy is 0.218750, time 208.900 samples/sec
+epoch 7, total_step 115500, total loss is 15.55 , inference loss is 7.97, weight deacy loss is 7.58, training accuracy is 0.437500, time 14.267 samples/sec
+epoch 7, total_step 115520, total loss is 15.26 , inference loss is 7.68, weight deacy loss is 7.58, training accuracy is 0.406250, time 201.447 samples/sec
+epoch 7, total_step 115540, total loss is 17.05 , inference loss is 9.47, weight deacy loss is 7.58, training accuracy is 0.375000, time 213.427 samples/sec
+epoch 7, total_step 115560, total loss is 14.80 , inference loss is 7.22, weight deacy loss is 7.58, training accuracy is 0.375000, time 209.059 samples/sec
+epoch 7, total_step 115580, total loss is 20.54 , inference loss is 12.96, weight deacy loss is 7.58, training accuracy is 0.281250, time 224.106 samples/sec
+epoch 7, total_step 115600, total loss is 15.33 , inference loss is 7.75, weight deacy loss is 7.58, training accuracy is 0.406250, time 209.093 samples/sec
+epoch 7, total_step 115620, total loss is 20.21 , inference loss is 12.63, weight deacy loss is 7.58, training accuracy is 0.281250, time 211.165 samples/sec
+epoch 7, total_step 115640, total loss is 17.19 , inference loss is 9.62, weight deacy loss is 7.58, training accuracy is 0.281250, time 217.120 samples/sec
+epoch 7, total_step 115660, total loss is 20.59 , inference loss is 13.01, weight deacy loss is 7.58, training accuracy is 0.312500, time 209.940 samples/sec
+epoch 7, total_step 115680, total loss is 21.04 , inference loss is 13.47, weight deacy loss is 7.57, training accuracy is 0.218750, time 212.526 samples/sec
+epoch 7, total_step 115700, total loss is 15.35 , inference loss is 7.78, weight deacy loss is 7.57, training accuracy is 0.375000, time 205.310 samples/sec
+epoch 7, total_step 115720, total loss is 18.29 , inference loss is 10.71, weight deacy loss is 7.57, training accuracy is 0.250000, time 207.784 samples/sec
+epoch 7, total_step 115740, total loss is 16.07 , inference loss is 8.49, weight deacy loss is 7.57, training accuracy is 0.281250, time 211.748 samples/sec
+epoch 7, total_step 115760, total loss is 16.16 , inference loss is 8.59, weight deacy loss is 7.57, training accuracy is 0.500000, time 210.701 samples/sec
+epoch 7, total_step 115780, total loss is 16.11 , inference loss is 8.54, weight deacy loss is 7.57, training accuracy is 0.312500, time 208.440 samples/sec
+epoch 7, total_step 115800, total loss is 15.13 , inference loss is 7.55, weight deacy loss is 7.57, training accuracy is 0.375000, time 11.362 samples/sec
+epoch 7, total_step 115820, total loss is 15.52 , inference loss is 7.95, weight deacy loss is 7.57, training accuracy is 0.281250, time 212.877 samples/sec
+epoch 7, total_step 115840, total loss is 22.62 , inference loss is 15.05, weight deacy loss is 7.57, training accuracy is 0.187500, time 213.110 samples/sec
+epoch 7, total_step 115860, total loss is 19.22 , inference loss is 11.65, weight deacy loss is 7.57, training accuracy is 0.343750, time 216.622 samples/sec
+epoch 7, total_step 115880, total loss is 20.35 , inference loss is 12.78, weight deacy loss is 7.57, training accuracy is 0.312500, time 212.007 samples/sec
+epoch 7, total_step 115900, total loss is 17.08 , inference loss is 9.51, weight deacy loss is 7.57, training accuracy is 0.406250, time 214.135 samples/sec
+epoch 7, total_step 115920, total loss is 18.28 , inference loss is 10.71, weight deacy loss is 7.57, training accuracy is 0.281250, time 203.494 samples/sec
+epoch 7, total_step 115940, total loss is 17.87 , inference loss is 10.30, weight deacy loss is 7.57, training accuracy is 0.312500, time 203.882 samples/sec
+epoch 7, total_step 115960, total loss is 15.33 , inference loss is 7.77, weight deacy loss is 7.57, training accuracy is 0.375000, time 212.152 samples/sec
+epoch 7, total_step 115980, total loss is 10.90 , inference loss is 3.34, weight deacy loss is 7.57, training accuracy is 0.562500, time 212.267 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.656479
+best_threshold_index 117 0.9894444444444445
+best_threshold_index 117 0.9888888888888889
+best_threshold_index 117 0.9896296296296296
+best_threshold_index 117 0.9894444444444445
+best_threshold_index 117 0.9894444444444445
+best_threshold_index 117 0.9885185185185185
+best_threshold_index 117 0.9894444444444445
+best_threshold_index 117 0.9888888888888889
+best_threshold_index 117 0.9881481481481481
+best_threshold_index 117 0.9881481481481481
+[lfw][116000]XNorm: 22.080170
+[lfw][116000]Accuracy-Flip: 0.98900+-0.00484
+test accuracy is: 0.9889999999999999
+epoch 7, total_step 116000, total loss is 15.83 , inference loss is 8.26, weight deacy loss is 7.57, training accuracy is 0.375000, time 14.118 samples/sec
+epoch 7, total_step 116020, total loss is 17.79 , inference loss is 10.22, weight deacy loss is 7.56, training accuracy is 0.218750, time 215.135 samples/sec
+epoch 7, total_step 116040, total loss is 18.55 , inference loss is 10.99, weight deacy loss is 7.56, training accuracy is 0.281250, time 210.301 samples/sec
+epoch 7, total_step 116060, total loss is 16.53 , inference loss is 8.96, weight deacy loss is 7.56, training accuracy is 0.312500, time 203.669 samples/sec
+epoch 7, total_step 116080, total loss is 17.23 , inference loss is 9.67, weight deacy loss is 7.56, training accuracy is 0.281250, time 204.954 samples/sec
+epoch 7, total_step 116100, total loss is 18.10 , inference loss is 10.54, weight deacy loss is 7.56, training accuracy is 0.281250, time 14.063 samples/sec
+epoch 7, total_step 116120, total loss is 18.07 , inference loss is 10.51, weight deacy loss is 7.56, training accuracy is 0.343750, time 298.264 samples/sec
+epoch 7, total_step 116140, total loss is 17.79 , inference loss is 10.22, weight deacy loss is 7.56, training accuracy is 0.218750, time 209.853 samples/sec
+epoch 7, total_step 116160, total loss is 15.09 , inference loss is 7.53, weight deacy loss is 7.56, training accuracy is 0.500000, time 195.979 samples/sec
+epoch 7, total_step 116180, total loss is 13.49 , inference loss is 5.93, weight deacy loss is 7.56, training accuracy is 0.343750, time 216.942 samples/sec
+epoch 7, total_step 116200, total loss is 14.32 , inference loss is 6.76, weight deacy loss is 7.56, training accuracy is 0.531250, time 212.427 samples/sec
+epoch 7, total_step 116220, total loss is 18.95 , inference loss is 11.39, weight deacy loss is 7.56, training accuracy is 0.531250, time 226.031 samples/sec
+epoch 7, total_step 116240, total loss is 18.95 , inference loss is 11.39, weight deacy loss is 7.56, training accuracy is 0.250000, time 210.959 samples/sec
+epoch 7, total_step 116260, total loss is 15.56 , inference loss is 8.00, weight deacy loss is 7.56, training accuracy is 0.468750, time 218.907 samples/sec
+epoch 7, total_step 116280, total loss is 15.44 , inference loss is 7.88, weight deacy loss is 7.56, training accuracy is 0.437500, time 209.310 samples/sec
+epoch 7, total_step 116300, total loss is 14.47 , inference loss is 6.91, weight deacy loss is 7.56, training accuracy is 0.375000, time 207.020 samples/sec
+epoch 7, total_step 116320, total loss is 16.85 , inference loss is 9.29, weight deacy loss is 7.56, training accuracy is 0.375000, time 199.998 samples/sec
+epoch 7, total_step 116340, total loss is 15.41 , inference loss is 7.86, weight deacy loss is 7.56, training accuracy is 0.312500, time 210.169 samples/sec
+epoch 7, total_step 116360, total loss is 18.39 , inference loss is 10.84, weight deacy loss is 7.56, training accuracy is 0.312500, time 222.388 samples/sec
+epoch 7, total_step 116380, total loss is 16.56 , inference loss is 9.00, weight deacy loss is 7.55, training accuracy is 0.375000, time 210.566 samples/sec
+epoch 7, total_step 116400, total loss is 15.74 , inference loss is 8.18, weight deacy loss is 7.55, training accuracy is 0.406250, time 14.865 samples/sec
+epoch 7, total_step 116420, total loss is 15.35 , inference loss is 7.80, weight deacy loss is 7.55, training accuracy is 0.343750, time 215.253 samples/sec
+epoch 7, total_step 116440, total loss is 18.44 , inference loss is 10.89, weight deacy loss is 7.55, training accuracy is 0.343750, time 214.859 samples/sec
+epoch 7, total_step 116460, total loss is 19.71 , inference loss is 12.16, weight deacy loss is 7.55, training accuracy is 0.437500, time 209.240 samples/sec
+epoch 7, total_step 116480, total loss is 16.41 , inference loss is 8.85, weight deacy loss is 7.55, training accuracy is 0.312500, time 213.025 samples/sec
+epoch 7, total_step 116500, total loss is 13.71 , inference loss is 6.16, weight deacy loss is 7.55, training accuracy is 0.406250, time 211.705 samples/sec
+epoch 7, total_step 116520, total loss is 15.63 , inference loss is 8.08, weight deacy loss is 7.55, training accuracy is 0.437500, time 213.883 samples/sec
+epoch 7, total_step 116540, total loss is 16.89 , inference loss is 9.34, weight deacy loss is 7.55, training accuracy is 0.343750, time 203.757 samples/sec
+epoch 7, total_step 116560, total loss is 17.75 , inference loss is 10.20, weight deacy loss is 7.55, training accuracy is 0.406250, time 213.618 samples/sec
+epoch 7, total_step 116580, total loss is 16.00 , inference loss is 8.45, weight deacy loss is 7.55, training accuracy is 0.406250, time 222.336 samples/sec
+epoch 7, total_step 116600, total loss is 17.67 , inference loss is 10.12, weight deacy loss is 7.55, training accuracy is 0.437500, time 228.040 samples/sec
+epoch 7, total_step 116620, total loss is 13.90 , inference loss is 6.35, weight deacy loss is 7.55, training accuracy is 0.500000, time 208.571 samples/sec
+epoch 7, total_step 116640, total loss is 14.40 , inference loss is 6.86, weight deacy loss is 7.55, training accuracy is 0.437500, time 221.064 samples/sec
+epoch 7, total_step 116660, total loss is 19.01 , inference loss is 11.46, weight deacy loss is 7.55, training accuracy is 0.312500, time 213.788 samples/sec
+epoch 7, total_step 116680, total loss is 18.78 , inference loss is 11.24, weight deacy loss is 7.55, training accuracy is 0.250000, time 199.682 samples/sec
+epoch 7, total_step 116700, total loss is 18.69 , inference loss is 11.15, weight deacy loss is 7.55, training accuracy is 0.406250, time 13.297 samples/sec
+epoch 7, total_step 116720, total loss is 18.72 , inference loss is 11.17, weight deacy loss is 7.55, training accuracy is 0.281250, time 210.553 samples/sec
+epoch 7, total_step 116740, total loss is 13.05 , inference loss is 5.51, weight deacy loss is 7.54, training accuracy is 0.375000, time 218.014 samples/sec
+epoch 7, total_step 116760, total loss is 14.44 , inference loss is 6.90, weight deacy loss is 7.54, training accuracy is 0.343750, time 213.822 samples/sec
+epoch 7, total_step 116780, total loss is 15.77 , inference loss is 8.22, weight deacy loss is 7.54, training accuracy is 0.375000, time 218.957 samples/sec
+epoch 7, total_step 116800, total loss is 18.84 , inference loss is 11.30, weight deacy loss is 7.54, training accuracy is 0.437500, time 243.979 samples/sec
+epoch 7, total_step 116820, total loss is 13.35 , inference loss is 5.81, weight deacy loss is 7.54, training accuracy is 0.312500, time 215.973 samples/sec
+epoch 7, total_step 116840, total loss is 17.78 , inference loss is 10.24, weight deacy loss is 7.54, training accuracy is 0.312500, time 215.510 samples/sec
+epoch 7, total_step 116860, total loss is 13.51 , inference loss is 5.96, weight deacy loss is 7.54, training accuracy is 0.468750, time 216.166 samples/sec
+epoch 7, total_step 116880, total loss is 17.59 , inference loss is 10.05, weight deacy loss is 7.54, training accuracy is 0.375000, time 212.694 samples/sec
+epoch 7, total_step 116900, total loss is 16.35 , inference loss is 8.81, weight deacy loss is 7.54, training accuracy is 0.343750, time 226.028 samples/sec
+epoch 7, total_step 116920, total loss is 16.23 , inference loss is 8.69, weight deacy loss is 7.54, training accuracy is 0.406250, time 296.507 samples/sec
+epoch 7, total_step 116940, total loss is 19.46 , inference loss is 11.92, weight deacy loss is 7.54, training accuracy is 0.343750, time 226.715 samples/sec
+epoch 7, total_step 116960, total loss is 15.20 , inference loss is 7.66, weight deacy loss is 7.54, training accuracy is 0.375000, time 208.488 samples/sec
+epoch 7, total_step 116980, total loss is 23.24 , inference loss is 15.70, weight deacy loss is 7.54, training accuracy is 0.281250, time 210.868 samples/sec
+epoch 7, total_step 117000, total loss is 18.78 , inference loss is 11.24, weight deacy loss is 7.54, training accuracy is 0.375000, time 13.992 samples/sec
+epoch 7, total_step 117020, total loss is 15.83 , inference loss is 8.30, weight deacy loss is 7.54, training accuracy is 0.375000, time 225.366 samples/sec
+epoch 7, total_step 117040, total loss is 16.53 , inference loss is 8.99, weight deacy loss is 7.54, training accuracy is 0.468750, time 206.125 samples/sec
+epoch 7, total_step 117060, total loss is 16.78 , inference loss is 9.24, weight deacy loss is 7.54, training accuracy is 0.375000, time 205.053 samples/sec
+epoch 7, total_step 117080, total loss is 16.04 , inference loss is 8.50, weight deacy loss is 7.54, training accuracy is 0.437500, time 207.703 samples/sec
+epoch 7, total_step 117100, total loss is 16.42 , inference loss is 8.88, weight deacy loss is 7.53, training accuracy is 0.375000, time 299.012 samples/sec
+epoch 7, total_step 117120, total loss is 14.53 , inference loss is 7.00, weight deacy loss is 7.53, training accuracy is 0.343750, time 212.617 samples/sec
+epoch 7, total_step 117140, total loss is 19.81 , inference loss is 12.27, weight deacy loss is 7.53, training accuracy is 0.312500, time 194.375 samples/sec
+epoch 7, total_step 117160, total loss is 18.07 , inference loss is 10.54, weight deacy loss is 7.53, training accuracy is 0.375000, time 205.998 samples/sec
+epoch 7, total_step 117180, total loss is 13.78 , inference loss is 6.25, weight deacy loss is 7.53, training accuracy is 0.312500, time 212.570 samples/sec
+epoch 7, total_step 117200, total loss is 16.26 , inference loss is 8.73, weight deacy loss is 7.53, training accuracy is 0.343750, time 210.439 samples/sec
+epoch 7, total_step 117220, total loss is 13.90 , inference loss is 6.37, weight deacy loss is 7.53, training accuracy is 0.375000, time 216.639 samples/sec
+epoch 7, total_step 117240, total loss is 18.10 , inference loss is 10.57, weight deacy loss is 7.53, training accuracy is 0.343750, time 213.790 samples/sec
+epoch 7, total_step 117260, total loss is 24.46 , inference loss is 16.93, weight deacy loss is 7.53, training accuracy is 0.187500, time 227.131 samples/sec
+epoch 7, total_step 117280, total loss is 12.34 , inference loss is 4.81, weight deacy loss is 7.53, training accuracy is 0.593750, time 212.268 samples/sec
+epoch 7, total_step 117300, total loss is 18.55 , inference loss is 11.02, weight deacy loss is 7.53, training accuracy is 0.312500, time 12.791 samples/sec
+epoch 7, total_step 117320, total loss is 13.80 , inference loss is 6.27, weight deacy loss is 7.53, training accuracy is 0.468750, time 208.054 samples/sec
+epoch 7, total_step 117340, total loss is 13.33 , inference loss is 5.80, weight deacy loss is 7.53, training accuracy is 0.468750, time 211.154 samples/sec
+epoch 7, total_step 117360, total loss is 15.02 , inference loss is 7.49, weight deacy loss is 7.53, training accuracy is 0.437500, time 177.093 samples/sec
+epoch 7, total_step 117380, total loss is 17.08 , inference loss is 9.55, weight deacy loss is 7.53, training accuracy is 0.312500, time 207.001 samples/sec
+epoch 7, total_step 117400, total loss is 19.56 , inference loss is 12.03, weight deacy loss is 7.53, training accuracy is 0.343750, time 213.563 samples/sec
+epoch 7, total_step 117420, total loss is 15.83 , inference loss is 8.31, weight deacy loss is 7.53, training accuracy is 0.406250, time 206.288 samples/sec
+epoch 7, total_step 117440, total loss is 16.28 , inference loss is 8.75, weight deacy loss is 7.52, training accuracy is 0.281250, time 213.021 samples/sec
+epoch 7, total_step 117460, total loss is 17.21 , inference loss is 9.68, weight deacy loss is 7.52, training accuracy is 0.406250, time 212.518 samples/sec
+epoch 7, total_step 117480, total loss is 17.41 , inference loss is 9.88, weight deacy loss is 7.52, training accuracy is 0.343750, time 202.999 samples/sec
+epoch 7, total_step 117500, total loss is 18.90 , inference loss is 11.38, weight deacy loss is 7.52, training accuracy is 0.281250, time 217.367 samples/sec
+epoch 7, total_step 117520, total loss is 18.45 , inference loss is 10.93, weight deacy loss is 7.52, training accuracy is 0.375000, time 210.790 samples/sec
+epoch 7, total_step 117540, total loss is 15.34 , inference loss is 7.82, weight deacy loss is 7.52, training accuracy is 0.437500, time 215.366 samples/sec
+epoch 7, total_step 117560, total loss is 15.31 , inference loss is 7.79, weight deacy loss is 7.52, training accuracy is 0.500000, time 213.753 samples/sec
+epoch 7, total_step 117580, total loss is 21.34 , inference loss is 13.82, weight deacy loss is 7.52, training accuracy is 0.187500, time 226.527 samples/sec
+epoch 7, total_step 117600, total loss is 19.38 , inference loss is 11.86, weight deacy loss is 7.52, training accuracy is 0.281250, time 12.678 samples/sec
+epoch 7, total_step 117620, total loss is 17.16 , inference loss is 9.64, weight deacy loss is 7.52, training accuracy is 0.343750, time 196.448 samples/sec
+epoch 7, total_step 117640, total loss is 16.40 , inference loss is 8.88, weight deacy loss is 7.52, training accuracy is 0.468750, time 296.951 samples/sec
+epoch 7, total_step 117660, total loss is 16.53 , inference loss is 9.02, weight deacy loss is 7.52, training accuracy is 0.281250, time 213.348 samples/sec
+epoch 7, total_step 117680, total loss is 14.04 , inference loss is 6.52, weight deacy loss is 7.52, training accuracy is 0.437500, time 212.853 samples/sec
+epoch 7, total_step 117700, total loss is 15.53 , inference loss is 8.02, weight deacy loss is 7.52, training accuracy is 0.375000, time 212.166 samples/sec
+epoch 7, total_step 117720, total loss is 19.80 , inference loss is 12.28, weight deacy loss is 7.52, training accuracy is 0.343750, time 229.117 samples/sec
+epoch 7, total_step 117740, total loss is 13.67 , inference loss is 6.16, weight deacy loss is 7.52, training accuracy is 0.406250, time 209.515 samples/sec
+epoch 7, total_step 117760, total loss is 18.00 , inference loss is 10.49, weight deacy loss is 7.52, training accuracy is 0.281250, time 216.475 samples/sec
+epoch 7, total_step 117780, total loss is 19.62 , inference loss is 12.10, weight deacy loss is 7.51, training accuracy is 0.281250, time 212.716 samples/sec
+epoch 7, total_step 117800, total loss is 15.55 , inference loss is 8.04, weight deacy loss is 7.51, training accuracy is 0.468750, time 213.799 samples/sec
+epoch 7, total_step 117820, total loss is 16.76 , inference loss is 9.24, weight deacy loss is 7.51, training accuracy is 0.312500, time 212.050 samples/sec
+epoch 7, total_step 117840, total loss is 19.49 , inference loss is 11.98, weight deacy loss is 7.51, training accuracy is 0.375000, time 211.337 samples/sec
+epoch 7, total_step 117860, total loss is 16.48 , inference loss is 8.97, weight deacy loss is 7.51, training accuracy is 0.406250, time 207.124 samples/sec
+epoch 7, total_step 117880, total loss is 18.98 , inference loss is 11.46, weight deacy loss is 7.51, training accuracy is 0.218750, time 209.232 samples/sec
+epoch 7, total_step 117900, total loss is 13.98 , inference loss is 6.47, weight deacy loss is 7.51, training accuracy is 0.593750, time 14.066 samples/sec
+epoch 7, total_step 117920, total loss is 20.14 , inference loss is 12.63, weight deacy loss is 7.51, training accuracy is 0.375000, time 211.740 samples/sec
+epoch 7, total_step 117940, total loss is 18.19 , inference loss is 10.68, weight deacy loss is 7.51, training accuracy is 0.406250, time 206.385 samples/sec
+epoch 7, total_step 117960, total loss is 17.91 , inference loss is 10.40, weight deacy loss is 7.51, training accuracy is 0.437500, time 206.364 samples/sec
+epoch 7, total_step 117980, total loss is 16.66 , inference loss is 9.15, weight deacy loss is 7.51, training accuracy is 0.281250, time 209.195 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.620454000000024
+best_threshold_index 110 0.9887037037037038
+best_threshold_index 110 0.9875925925925926
+best_threshold_index 110 0.9874074074074074
+best_threshold_index 111 0.9883333333333333
+best_threshold_index 110 0.9879629629629629
+best_threshold_index 111 0.9874074074074074
+best_threshold_index 110 0.9879629629629629
+best_threshold_index 110 0.9866666666666667
+best_threshold_index 110 0.9862962962962963
+best_threshold_index 110 0.987037037037037
+[lfw][118000]XNorm: 21.525841
+[lfw][118000]Accuracy-Flip: 0.98717+-0.00633
+test accuracy is: 0.9871666666666667
+epoch 7, total_step 118000, total loss is 16.70 , inference loss is 9.19, weight deacy loss is 7.51, training accuracy is 0.250000, time 19.154 samples/sec
+epoch 7, total_step 118020, total loss is 15.70 , inference loss is 8.19, weight deacy loss is 7.51, training accuracy is 0.375000, time 299.254 samples/sec
+epoch 7, total_step 118040, total loss is 17.33 , inference loss is 9.82, weight deacy loss is 7.51, training accuracy is 0.343750, time 210.665 samples/sec
+epoch 7, total_step 118060, total loss is 15.93 , inference loss is 8.42, weight deacy loss is 7.51, training accuracy is 0.343750, time 206.796 samples/sec
+epoch 7, total_step 118080, total loss is 17.97 , inference loss is 10.47, weight deacy loss is 7.51, training accuracy is 0.343750, time 205.313 samples/sec
+epoch 7, total_step 118100, total loss is 18.85 , inference loss is 11.35, weight deacy loss is 7.51, training accuracy is 0.343750, time 217.543 samples/sec
+epoch 7, total_step 118120, total loss is 14.17 , inference loss is 6.67, weight deacy loss is 7.51, training accuracy is 0.375000, time 213.106 samples/sec
+epoch 7, total_step 118140, total loss is 20.42 , inference loss is 12.92, weight deacy loss is 7.50, training accuracy is 0.218750, time 191.831 samples/sec
+epoch 7, total_step 118160, total loss is 17.03 , inference loss is 9.53, weight deacy loss is 7.50, training accuracy is 0.343750, time 211.294 samples/sec
+epoch 7, total_step 118180, total loss is 15.50 , inference loss is 7.99, weight deacy loss is 7.50, training accuracy is 0.500000, time 223.569 samples/sec
+epoch 7, total_step 118200, total loss is 12.94 , inference loss is 5.44, weight deacy loss is 7.50, training accuracy is 0.437500, time 14.253 samples/sec
+epoch 7, total_step 118220, total loss is 18.74 , inference loss is 11.24, weight deacy loss is 7.50, training accuracy is 0.281250, time 222.903 samples/sec
+epoch 7, total_step 118240, total loss is 17.68 , inference loss is 10.18, weight deacy loss is 7.50, training accuracy is 0.343750, time 199.099 samples/sec
+epoch 7, total_step 118260, total loss is 14.80 , inference loss is 7.30, weight deacy loss is 7.50, training accuracy is 0.343750, time 207.616 samples/sec
+epoch 7, total_step 118280, total loss is 14.74 , inference loss is 7.24, weight deacy loss is 7.50, training accuracy is 0.343750, time 215.643 samples/sec
+epoch 7, total_step 118300, total loss is 15.85 , inference loss is 8.35, weight deacy loss is 7.50, training accuracy is 0.406250, time 297.478 samples/sec
+epoch 7, total_step 118320, total loss is 20.11 , inference loss is 12.61, weight deacy loss is 7.50, training accuracy is 0.218750, time 240.233 samples/sec
+epoch 7, total_step 118340, total loss is 17.69 , inference loss is 10.19, weight deacy loss is 7.50, training accuracy is 0.375000, time 212.745 samples/sec
+epoch 7, total_step 118360, total loss is 18.52 , inference loss is 11.02, weight deacy loss is 7.50, training accuracy is 0.218750, time 225.964 samples/sec
+epoch 7, total_step 118380, total loss is 14.32 , inference loss is 6.82, weight deacy loss is 7.50, training accuracy is 0.437500, time 221.290 samples/sec
+epoch 7, total_step 118400, total loss is 17.09 , inference loss is 9.59, weight deacy loss is 7.50, training accuracy is 0.343750, time 237.763 samples/sec
+epoch 7, total_step 118420, total loss is 15.60 , inference loss is 8.11, weight deacy loss is 7.50, training accuracy is 0.375000, time 216.344 samples/sec
+epoch 7, total_step 118440, total loss is 16.60 , inference loss is 9.10, weight deacy loss is 7.50, training accuracy is 0.343750, time 216.591 samples/sec
+epoch 7, total_step 118460, total loss is 18.85 , inference loss is 11.35, weight deacy loss is 7.50, training accuracy is 0.312500, time 212.361 samples/sec
+epoch 7, total_step 118480, total loss is 14.38 , inference loss is 6.88, weight deacy loss is 7.50, training accuracy is 0.468750, time 222.536 samples/sec
+epoch 7, total_step 118500, total loss is 22.17 , inference loss is 14.68, weight deacy loss is 7.49, training accuracy is 0.156250, time 14.425 samples/sec
+epoch 7, total_step 118520, total loss is 17.20 , inference loss is 9.70, weight deacy loss is 7.49, training accuracy is 0.312500, time 214.974 samples/sec
+epoch 7, total_step 118540, total loss is 19.81 , inference loss is 12.32, weight deacy loss is 7.49, training accuracy is 0.250000, time 191.363 samples/sec
+epoch 7, total_step 118560, total loss is 16.39 , inference loss is 8.90, weight deacy loss is 7.49, training accuracy is 0.281250, time 298.026 samples/sec
+epoch 7, total_step 118580, total loss is 13.79 , inference loss is 6.30, weight deacy loss is 7.49, training accuracy is 0.406250, time 213.898 samples/sec
+epoch 7, total_step 118600, total loss is 16.40 , inference loss is 8.91, weight deacy loss is 7.49, training accuracy is 0.375000, time 219.638 samples/sec
+epoch 7, total_step 118620, total loss is 15.86 , inference loss is 8.37, weight deacy loss is 7.49, training accuracy is 0.375000, time 214.637 samples/sec
+epoch 7, total_step 118640, total loss is 15.21 , inference loss is 7.71, weight deacy loss is 7.49, training accuracy is 0.500000, time 217.389 samples/sec
+epoch 7, total_step 118660, total loss is 13.03 , inference loss is 5.54, weight deacy loss is 7.49, training accuracy is 0.500000, time 221.266 samples/sec
+epoch 7, total_step 118680, total loss is 20.02 , inference loss is 12.53, weight deacy loss is 7.49, training accuracy is 0.218750, time 216.613 samples/sec
+epoch 7, total_step 118700, total loss is 13.03 , inference loss is 5.54, weight deacy loss is 7.49, training accuracy is 0.468750, time 217.341 samples/sec
+epoch 7, total_step 118720, total loss is 14.88 , inference loss is 7.39, weight deacy loss is 7.49, training accuracy is 0.531250, time 296.318 samples/sec
+epoch 7, total_step 118740, total loss is 17.49 , inference loss is 10.00, weight deacy loss is 7.49, training accuracy is 0.406250, time 299.998 samples/sec
+epoch 7, total_step 118760, total loss is 12.80 , inference loss is 5.31, weight deacy loss is 7.49, training accuracy is 0.312500, time 298.413 samples/sec
+epoch 7, total_step 118780, total loss is 15.47 , inference loss is 7.98, weight deacy loss is 7.49, training accuracy is 0.187500, time 214.985 samples/sec
+epoch 7, total_step 118800, total loss is 15.87 , inference loss is 8.38, weight deacy loss is 7.49, training accuracy is 0.312500, time 14.190 samples/sec
+epoch 7, total_step 118820, total loss is 14.58 , inference loss is 7.09, weight deacy loss is 7.49, training accuracy is 0.375000, time 223.303 samples/sec
+epoch 7, total_step 118840, total loss is 16.73 , inference loss is 9.25, weight deacy loss is 7.49, training accuracy is 0.406250, time 208.211 samples/sec
+epoch 7, total_step 118860, total loss is 18.62 , inference loss is 11.13, weight deacy loss is 7.48, training accuracy is 0.312500, time 216.248 samples/sec
+epoch 7, total_step 118880, total loss is 19.45 , inference loss is 11.96, weight deacy loss is 7.48, training accuracy is 0.281250, time 212.103 samples/sec
+epoch 7, total_step 118900, total loss is 17.26 , inference loss is 9.77, weight deacy loss is 7.48, training accuracy is 0.281250, time 208.710 samples/sec
+epoch 7, total_step 118920, total loss is 18.14 , inference loss is 10.66, weight deacy loss is 7.48, training accuracy is 0.250000, time 217.709 samples/sec
+epoch 7, total_step 118940, total loss is 14.55 , inference loss is 7.07, weight deacy loss is 7.48, training accuracy is 0.500000, time 202.376 samples/sec
+epoch 7, total_step 118960, total loss is 20.39 , inference loss is 12.91, weight deacy loss is 7.48, training accuracy is 0.187500, time 223.484 samples/sec
+epoch 7, total_step 118980, total loss is 16.17 , inference loss is 8.69, weight deacy loss is 7.48, training accuracy is 0.375000, time 204.894 samples/sec
+epoch 7, total_step 119000, total loss is 19.88 , inference loss is 12.40, weight deacy loss is 7.48, training accuracy is 0.343750, time 208.671 samples/sec
+epoch 7, total_step 119020, total loss is 18.99 , inference loss is 11.51, weight deacy loss is 7.48, training accuracy is 0.375000, time 297.925 samples/sec
+epoch 7, total_step 119040, total loss is 14.89 , inference loss is 7.41, weight deacy loss is 7.48, training accuracy is 0.375000, time 217.998 samples/sec
+epoch 7, total_step 119060, total loss is 14.89 , inference loss is 7.41, weight deacy loss is 7.48, training accuracy is 0.437500, time 216.817 samples/sec
+epoch 7, total_step 119080, total loss is 18.02 , inference loss is 10.54, weight deacy loss is 7.48, training accuracy is 0.343750, time 197.720 samples/sec
+epoch 7, total_step 119100, total loss is 15.67 , inference loss is 8.19, weight deacy loss is 7.48, training accuracy is 0.375000, time 13.998 samples/sec
+epoch 7, total_step 119120, total loss is 17.92 , inference loss is 10.44, weight deacy loss is 7.48, training accuracy is 0.187500, time 218.923 samples/sec
+epoch 7, total_step 119140, total loss is 20.45 , inference loss is 12.97, weight deacy loss is 7.48, training accuracy is 0.218750, time 213.382 samples/sec
+epoch 7, total_step 119160, total loss is 18.87 , inference loss is 11.39, weight deacy loss is 7.48, training accuracy is 0.281250, time 230.446 samples/sec
+epoch 7, total_step 119180, total loss is 17.86 , inference loss is 10.38, weight deacy loss is 7.48, training accuracy is 0.375000, time 210.632 samples/sec
+epoch 7, total_step 119200, total loss is 16.48 , inference loss is 9.01, weight deacy loss is 7.48, training accuracy is 0.343750, time 214.682 samples/sec
+epoch 7, total_step 119220, total loss is 19.74 , inference loss is 12.27, weight deacy loss is 7.47, training accuracy is 0.218750, time 211.336 samples/sec
+epoch 7, total_step 119240, total loss is 18.91 , inference loss is 11.43, weight deacy loss is 7.47, training accuracy is 0.250000, time 206.889 samples/sec
+epoch 7, total_step 119260, total loss is 17.69 , inference loss is 10.21, weight deacy loss is 7.47, training accuracy is 0.187500, time 206.476 samples/sec
+epoch 7, total_step 119280, total loss is 16.14 , inference loss is 8.67, weight deacy loss is 7.47, training accuracy is 0.343750, time 298.281 samples/sec
+epoch 7, total_step 119300, total loss is 15.30 , inference loss is 7.83, weight deacy loss is 7.47, training accuracy is 0.406250, time 208.768 samples/sec
+epoch 7, total_step 119320, total loss is 19.75 , inference loss is 12.28, weight deacy loss is 7.47, training accuracy is 0.250000, time 215.250 samples/sec
+epoch 7, total_step 119340, total loss is 17.68 , inference loss is 10.21, weight deacy loss is 7.47, training accuracy is 0.281250, time 224.615 samples/sec
+epoch 7, total_step 119360, total loss is 19.37 , inference loss is 11.90, weight deacy loss is 7.47, training accuracy is 0.343750, time 214.498 samples/sec
+epoch 7, total_step 119380, total loss is 15.37 , inference loss is 7.90, weight deacy loss is 7.47, training accuracy is 0.250000, time 211.148 samples/sec
+epoch 7, total_step 119400, total loss is 16.28 , inference loss is 8.81, weight deacy loss is 7.47, training accuracy is 0.406250, time 11.644 samples/sec
+epoch 7, total_step 119420, total loss is 17.62 , inference loss is 10.15, weight deacy loss is 7.47, training accuracy is 0.250000, time 207.535 samples/sec
+epoch 7, total_step 119440, total loss is 14.56 , inference loss is 7.09, weight deacy loss is 7.47, training accuracy is 0.375000, time 211.540 samples/sec
+epoch 7, total_step 119460, total loss is 21.30 , inference loss is 13.83, weight deacy loss is 7.47, training accuracy is 0.281250, time 206.183 samples/sec
+epoch 7, total_step 119480, total loss is 15.17 , inference loss is 7.70, weight deacy loss is 7.47, training accuracy is 0.375000, time 209.336 samples/sec
+epoch 7, total_step 119500, total loss is 19.42 , inference loss is 11.95, weight deacy loss is 7.47, training accuracy is 0.187500, time 217.782 samples/sec
+epoch 7, total_step 119520, total loss is 20.45 , inference loss is 12.99, weight deacy loss is 7.47, training accuracy is 0.218750, time 202.836 samples/sec
+epoch 7, total_step 119540, total loss is 20.13 , inference loss is 12.66, weight deacy loss is 7.47, training accuracy is 0.156250, time 298.173 samples/sec
+epoch 7, total_step 119560, total loss is 19.57 , inference loss is 12.11, weight deacy loss is 7.47, training accuracy is 0.281250, time 216.290 samples/sec
+epoch 7, total_step 119580, total loss is 15.60 , inference loss is 8.13, weight deacy loss is 7.46, training accuracy is 0.343750, time 205.425 samples/sec
+epoch 7, total_step 119600, total loss is 14.09 , inference loss is 6.63, weight deacy loss is 7.46, training accuracy is 0.437500, time 205.592 samples/sec
+epoch 7, total_step 119620, total loss is 17.39 , inference loss is 9.93, weight deacy loss is 7.46, training accuracy is 0.281250, time 210.495 samples/sec
+epoch 7, total_step 119640, total loss is 15.49 , inference loss is 8.03, weight deacy loss is 7.46, training accuracy is 0.312500, time 205.646 samples/sec
+epoch 7, total_step 119660, total loss is 18.90 , inference loss is 11.44, weight deacy loss is 7.46, training accuracy is 0.218750, time 236.134 samples/sec
+epoch 7, total_step 119680, total loss is 17.86 , inference loss is 10.40, weight deacy loss is 7.46, training accuracy is 0.281250, time 214.661 samples/sec
+epoch 7, total_step 119700, total loss is 15.03 , inference loss is 7.57, weight deacy loss is 7.46, training accuracy is 0.343750, time 14.550 samples/sec
+epoch 7, total_step 119720, total loss is 19.27 , inference loss is 11.81, weight deacy loss is 7.46, training accuracy is 0.250000, time 209.498 samples/sec
+epoch 7, total_step 119740, total loss is 16.81 , inference loss is 9.35, weight deacy loss is 7.46, training accuracy is 0.281250, time 197.362 samples/sec
+epoch 7, total_step 119760, total loss is 18.88 , inference loss is 11.42, weight deacy loss is 7.46, training accuracy is 0.187500, time 217.571 samples/sec
+epoch 7, total_step 119780, total loss is 16.79 , inference loss is 9.33, weight deacy loss is 7.46, training accuracy is 0.281250, time 213.380 samples/sec
+epoch 7, total_step 119800, total loss is 20.22 , inference loss is 12.76, weight deacy loss is 7.46, training accuracy is 0.156250, time 181.883 samples/sec
+epoch 7, total_step 119820, total loss is 18.51 , inference loss is 11.05, weight deacy loss is 7.46, training accuracy is 0.281250, time 210.340 samples/sec
+epoch 7, total_step 119840, total loss is 17.00 , inference loss is 9.54, weight deacy loss is 7.46, training accuracy is 0.343750, time 212.250 samples/sec
+epoch 7, total_step 119860, total loss is 16.40 , inference loss is 8.94, weight deacy loss is 7.46, training accuracy is 0.406250, time 201.029 samples/sec
+epoch 7, total_step 119880, total loss is 17.57 , inference loss is 10.11, weight deacy loss is 7.46, training accuracy is 0.187500, time 212.318 samples/sec
+epoch 7, total_step 119900, total loss is 18.02 , inference loss is 10.57, weight deacy loss is 7.46, training accuracy is 0.281250, time 215.782 samples/sec
+epoch 7, total_step 119920, total loss is 20.46 , inference loss is 13.01, weight deacy loss is 7.46, training accuracy is 0.250000, time 206.773 samples/sec
+epoch 7, total_step 119940, total loss is 16.25 , inference loss is 8.79, weight deacy loss is 7.45, training accuracy is 0.312500, time 202.238 samples/sec
+epoch 7, total_step 119960, total loss is 16.74 , inference loss is 9.29, weight deacy loss is 7.45, training accuracy is 0.343750, time 209.214 samples/sec
+epoch 7, total_step 119980, total loss is 18.19 , inference loss is 10.73, weight deacy loss is 7.45, training accuracy is 0.343750, time 207.300 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.04926099999999
+best_threshold_index 110 0.9855555555555555
+best_threshold_index 109 0.9848148148148148
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 110 0.9868518518518519
+best_threshold_index 110 0.9861111111111112
+best_threshold_index 110 0.9851851851851852
+best_threshold_index 109 0.9851851851851852
+best_threshold_index 109 0.9846296296296296
+best_threshold_index 110 0.9840740740740741
+best_threshold_index 110 0.9840740740740741
+[lfw][120000]XNorm: 21.798404
+[lfw][120000]Accuracy-Flip: 0.98450+-0.00723
+test accuracy is: 0.9844999999999999
+epoch 7, total_step 120000, total loss is 18.80 , inference loss is 11.34, weight deacy loss is 7.45, training accuracy is 0.375000, time 18.273 samples/sec
+epoch 7, total_step 120020, total loss is 15.33 , inference loss is 7.88, weight deacy loss is 7.45, training accuracy is 0.343750, time 230.071 samples/sec
+epoch 7, total_step 120040, total loss is 15.48 , inference loss is 8.03, weight deacy loss is 7.45, training accuracy is 0.343750, time 213.546 samples/sec
+epoch 7, total_step 120060, total loss is 15.92 , inference loss is 8.47, weight deacy loss is 7.45, training accuracy is 0.375000, time 211.756 samples/sec
+epoch 7, total_step 120080, total loss is 20.36 , inference loss is 12.91, weight deacy loss is 7.45, training accuracy is 0.375000, time 207.494 samples/sec
+epoch 7, total_step 120100, total loss is 17.12 , inference loss is 9.67, weight deacy loss is 7.45, training accuracy is 0.156250, time 209.027 samples/sec
+epoch 7, total_step 120120, total loss is 18.62 , inference loss is 11.17, weight deacy loss is 7.45, training accuracy is 0.281250, time 205.820 samples/sec
+epoch 7, total_step 120140, total loss is 17.60 , inference loss is 10.15, weight deacy loss is 7.45, training accuracy is 0.218750, time 214.863 samples/sec
+epoch 7, total_step 120160, total loss is 19.31 , inference loss is 11.87, weight deacy loss is 7.45, training accuracy is 0.218750, time 299.240 samples/sec
+epoch 7, total_step 120180, total loss is 20.53 , inference loss is 13.08, weight deacy loss is 7.45, training accuracy is 0.218750, time 216.381 samples/sec
+epoch 7, total_step 120200, total loss is 17.07 , inference loss is 9.62, weight deacy loss is 7.45, training accuracy is 0.343750, time 193.927 samples/sec
+epoch 7, total_step 120220, total loss is 13.07 , inference loss is 5.62, weight deacy loss is 7.45, training accuracy is 0.281250, time 213.023 samples/sec
+epoch 7, total_step 120240, total loss is 17.41 , inference loss is 9.96, weight deacy loss is 7.45, training accuracy is 0.250000, time 212.048 samples/sec
+epoch 7, total_step 120260, total loss is 19.75 , inference loss is 12.31, weight deacy loss is 7.45, training accuracy is 0.312500, time 213.659 samples/sec
+epoch 7, total_step 120280, total loss is 20.74 , inference loss is 13.30, weight deacy loss is 7.45, training accuracy is 0.093750, time 214.905 samples/sec
+epoch 7, total_step 120300, total loss is 17.91 , inference loss is 10.47, weight deacy loss is 7.44, training accuracy is 0.343750, time 13.985 samples/sec
+epoch 7, total_step 120320, total loss is 20.80 , inference loss is 13.35, weight deacy loss is 7.44, training accuracy is 0.250000, time 297.077 samples/sec
+epoch 7, total_step 120340, total loss is 13.19 , inference loss is 5.75, weight deacy loss is 7.44, training accuracy is 0.343750, time 298.644 samples/sec
+epoch 7, total_step 120360, total loss is 15.27 , inference loss is 7.83, weight deacy loss is 7.44, training accuracy is 0.312500, time 219.734 samples/sec
+epoch 7, total_step 120380, total loss is 14.17 , inference loss is 6.72, weight deacy loss is 7.44, training accuracy is 0.281250, time 217.846 samples/sec
+epoch 7, total_step 120400, total loss is 17.47 , inference loss is 10.03, weight deacy loss is 7.44, training accuracy is 0.281250, time 219.783 samples/sec
+epoch 7, total_step 120420, total loss is 18.93 , inference loss is 11.48, weight deacy loss is 7.44, training accuracy is 0.250000, time 215.810 samples/sec
+epoch 7, total_step 120440, total loss is 16.45 , inference loss is 9.01, weight deacy loss is 7.44, training accuracy is 0.187500, time 213.820 samples/sec
+epoch 7, total_step 120460, total loss is 21.26 , inference loss is 13.82, weight deacy loss is 7.44, training accuracy is 0.250000, time 214.218 samples/sec
+epoch 7, total_step 120480, total loss is 19.86 , inference loss is 12.42, weight deacy loss is 7.44, training accuracy is 0.218750, time 219.173 samples/sec
+epoch 7, total_step 120500, total loss is 21.72 , inference loss is 14.29, weight deacy loss is 7.44, training accuracy is 0.125000, time 297.615 samples/sec
+epoch 7, total_step 120520, total loss is 17.56 , inference loss is 10.12, weight deacy loss is 7.44, training accuracy is 0.281250, time 297.076 samples/sec
+epoch 7, total_step 120540, total loss is 20.96 , inference loss is 13.52, weight deacy loss is 7.44, training accuracy is 0.218750, time 208.390 samples/sec
+epoch 7, total_step 120560, total loss is 19.77 , inference loss is 12.33, weight deacy loss is 7.44, training accuracy is 0.281250, time 208.825 samples/sec
+epoch 7, total_step 120580, total loss is 16.98 , inference loss is 9.54, weight deacy loss is 7.44, training accuracy is 0.187500, time 215.024 samples/sec
+epoch 7, total_step 120600, total loss is 18.46 , inference loss is 11.02, weight deacy loss is 7.44, training accuracy is 0.281250, time 14.934 samples/sec
+epoch 7, total_step 120620, total loss is 20.06 , inference loss is 12.62, weight deacy loss is 7.44, training accuracy is 0.187500, time 201.433 samples/sec
+epoch 7, total_step 120640, total loss is 14.03 , inference loss is 6.60, weight deacy loss is 7.44, training accuracy is 0.312500, time 209.393 samples/sec
+epoch 7, total_step 120660, total loss is 21.68 , inference loss is 14.25, weight deacy loss is 7.43, training accuracy is 0.250000, time 209.551 samples/sec
+epoch 7, total_step 120680, total loss is 19.21 , inference loss is 11.78, weight deacy loss is 7.43, training accuracy is 0.187500, time 203.153 samples/sec
+epoch 7, total_step 120700, total loss is 18.50 , inference loss is 11.07, weight deacy loss is 7.43, training accuracy is 0.312500, time 215.364 samples/sec
+epoch 7, total_step 120720, total loss is 15.33 , inference loss is 7.89, weight deacy loss is 7.43, training accuracy is 0.281250, time 211.113 samples/sec
+epoch 7, total_step 120740, total loss is 16.99 , inference loss is 9.56, weight deacy loss is 7.43, training accuracy is 0.343750, time 206.347 samples/sec
+epoch 7, total_step 120760, total loss is 18.03 , inference loss is 10.60, weight deacy loss is 7.43, training accuracy is 0.281250, time 204.306 samples/sec
+epoch 7, total_step 120780, total loss is 19.41 , inference loss is 11.98, weight deacy loss is 7.43, training accuracy is 0.250000, time 206.131 samples/sec
+epoch 7, total_step 120800, total loss is 14.61 , inference loss is 7.18, weight deacy loss is 7.43, training accuracy is 0.281250, time 211.573 samples/sec
+epoch 7, total_step 120820, total loss is 15.66 , inference loss is 8.23, weight deacy loss is 7.43, training accuracy is 0.531250, time 213.423 samples/sec
+epoch 7, total_step 120840, total loss is 16.85 , inference loss is 9.42, weight deacy loss is 7.43, training accuracy is 0.281250, time 207.488 samples/sec
+epoch 7, total_step 120860, total loss is 16.80 , inference loss is 9.37, weight deacy loss is 7.43, training accuracy is 0.343750, time 209.213 samples/sec
+epoch 7, total_step 120880, total loss is 17.26 , inference loss is 9.83, weight deacy loss is 7.43, training accuracy is 0.250000, time 215.577 samples/sec
+epoch 7, total_step 120900, total loss is 18.14 , inference loss is 10.71, weight deacy loss is 7.43, training accuracy is 0.187500, time 13.678 samples/sec
+epoch 7, total_step 120920, total loss is 14.27 , inference loss is 6.84, weight deacy loss is 7.43, training accuracy is 0.250000, time 214.855 samples/sec
+epoch 7, total_step 120940, total loss is 16.23 , inference loss is 8.80, weight deacy loss is 7.43, training accuracy is 0.437500, time 218.796 samples/sec
+epoch 7, total_step 120960, total loss is 20.49 , inference loss is 13.07, weight deacy loss is 7.43, training accuracy is 0.218750, time 211.412 samples/sec
+epoch 7, total_step 120980, total loss is 19.84 , inference loss is 12.41, weight deacy loss is 7.43, training accuracy is 0.125000, time 235.312 samples/sec
+epoch 7, total_step 121000, total loss is 14.99 , inference loss is 7.56, weight deacy loss is 7.42, training accuracy is 0.375000, time 213.100 samples/sec
+epoch 7, total_step 121020, total loss is 20.58 , inference loss is 13.15, weight deacy loss is 7.42, training accuracy is 0.187500, time 209.561 samples/sec
+epoch 7, total_step 121040, total loss is 20.71 , inference loss is 13.28, weight deacy loss is 7.42, training accuracy is 0.125000, time 206.625 samples/sec
+epoch 7, total_step 121060, total loss is 19.02 , inference loss is 11.59, weight deacy loss is 7.42, training accuracy is 0.281250, time 218.322 samples/sec
+epoch 7, total_step 121080, total loss is 15.29 , inference loss is 7.87, weight deacy loss is 7.42, training accuracy is 0.406250, time 207.857 samples/sec
+epoch 7, total_step 121100, total loss is 20.09 , inference loss is 12.67, weight deacy loss is 7.42, training accuracy is 0.218750, time 208.907 samples/sec
+epoch 7, total_step 121120, total loss is 17.88 , inference loss is 10.46, weight deacy loss is 7.42, training accuracy is 0.250000, time 216.119 samples/sec
+epoch 7, total_step 121140, total loss is 14.49 , inference loss is 7.07, weight deacy loss is 7.42, training accuracy is 0.406250, time 217.163 samples/sec
+epoch 7, total_step 121160, total loss is 16.09 , inference loss is 8.67, weight deacy loss is 7.42, training accuracy is 0.468750, time 209.871 samples/sec
+epoch 7, total_step 121180, total loss is 16.36 , inference loss is 8.94, weight deacy loss is 7.42, training accuracy is 0.187500, time 297.633 samples/sec
+epoch 7, total_step 121200, total loss is 17.45 , inference loss is 10.03, weight deacy loss is 7.42, training accuracy is 0.250000, time 11.851 samples/sec
+epoch 7, total_step 121220, total loss is 21.85 , inference loss is 14.43, weight deacy loss is 7.42, training accuracy is 0.218750, time 230.963 samples/sec
+epoch 7, total_step 121240, total loss is 16.74 , inference loss is 9.32, weight deacy loss is 7.42, training accuracy is 0.281250, time 212.681 samples/sec
+epoch 7, total_step 121260, total loss is 14.49 , inference loss is 7.07, weight deacy loss is 7.42, training accuracy is 0.437500, time 210.789 samples/sec
+epoch 7, total_step 121280, total loss is 17.79 , inference loss is 10.37, weight deacy loss is 7.42, training accuracy is 0.343750, time 209.031 samples/sec
+epoch 7, total_step 121300, total loss is 19.37 , inference loss is 11.95, weight deacy loss is 7.42, training accuracy is 0.125000, time 210.911 samples/sec
+epoch 7, total_step 121320, total loss is 23.30 , inference loss is 15.89, weight deacy loss is 7.42, training accuracy is 0.125000, time 208.544 samples/sec
+epoch 7, total_step 121340, total loss is 18.69 , inference loss is 11.27, weight deacy loss is 7.42, training accuracy is 0.218750, time 211.007 samples/sec
+epoch 7, total_step 121360, total loss is 18.88 , inference loss is 11.46, weight deacy loss is 7.41, training accuracy is 0.343750, time 212.848 samples/sec
+epoch 7, total_step 121380, total loss is 15.24 , inference loss is 7.82, weight deacy loss is 7.41, training accuracy is 0.281250, time 211.992 samples/sec
+epoch 7, total_step 121400, total loss is 13.08 , inference loss is 5.67, weight deacy loss is 7.41, training accuracy is 0.375000, time 228.770 samples/sec
+epoch 7, total_step 121420, total loss is 16.64 , inference loss is 9.23, weight deacy loss is 7.41, training accuracy is 0.312500, time 219.451 samples/sec
+epoch 7, total_step 121440, total loss is 19.70 , inference loss is 12.28, weight deacy loss is 7.41, training accuracy is 0.281250, time 296.778 samples/sec
+epoch 7, total_step 121460, total loss is 21.77 , inference loss is 14.36, weight deacy loss is 7.41, training accuracy is 0.218750, time 215.105 samples/sec
+epoch 7, total_step 121480, total loss is 16.81 , inference loss is 9.40, weight deacy loss is 7.41, training accuracy is 0.312500, time 209.456 samples/sec
+epoch 7, total_step 121500, total loss is 19.77 , inference loss is 12.35, weight deacy loss is 7.41, training accuracy is 0.187500, time 14.257 samples/sec
+epoch 7, total_step 121520, total loss is 17.14 , inference loss is 9.73, weight deacy loss is 7.41, training accuracy is 0.187500, time 222.183 samples/sec
+epoch 7, total_step 121540, total loss is 18.43 , inference loss is 11.02, weight deacy loss is 7.41, training accuracy is 0.218750, time 213.878 samples/sec
+epoch 7, total_step 121560, total loss is 15.66 , inference loss is 8.25, weight deacy loss is 7.41, training accuracy is 0.250000, time 211.737 samples/sec
+epoch 7, total_step 121580, total loss is 16.23 , inference loss is 8.82, weight deacy loss is 7.41, training accuracy is 0.343750, time 220.241 samples/sec
+epoch 7, total_step 121600, total loss is 16.29 , inference loss is 8.89, weight deacy loss is 7.41, training accuracy is 0.187500, time 221.342 samples/sec
+epoch 7, total_step 121620, total loss is 16.70 , inference loss is 9.30, weight deacy loss is 7.41, training accuracy is 0.218750, time 209.667 samples/sec
+epoch 7, total_step 121640, total loss is 24.72 , inference loss is 17.31, weight deacy loss is 7.41, training accuracy is 0.156250, time 204.383 samples/sec
+epoch 7, total_step 121660, total loss is 23.16 , inference loss is 15.76, weight deacy loss is 7.41, training accuracy is 0.156250, time 208.388 samples/sec
+epoch 7, total_step 121680, total loss is 17.16 , inference loss is 9.75, weight deacy loss is 7.41, training accuracy is 0.343750, time 206.845 samples/sec
+epoch 7, total_step 121700, total loss is 19.06 , inference loss is 11.65, weight deacy loss is 7.41, training accuracy is 0.218750, time 217.467 samples/sec
+epoch 7, total_step 121720, total loss is 15.71 , inference loss is 8.30, weight deacy loss is 7.40, training accuracy is 0.281250, time 299.104 samples/sec
+epoch 7, total_step 121740, total loss is 14.31 , inference loss is 6.90, weight deacy loss is 7.40, training accuracy is 0.375000, time 205.215 samples/sec
+epoch 7, total_step 121760, total loss is 18.26 , inference loss is 10.85, weight deacy loss is 7.40, training accuracy is 0.187500, time 211.223 samples/sec
+epoch 7, total_step 121780, total loss is 15.34 , inference loss is 7.93, weight deacy loss is 7.40, training accuracy is 0.375000, time 203.054 samples/sec
+epoch 7, total_step 121800, total loss is 22.20 , inference loss is 14.80, weight deacy loss is 7.40, training accuracy is 0.125000, time 13.914 samples/sec
+epoch 7, total_step 121820, total loss is 21.86 , inference loss is 14.46, weight deacy loss is 7.40, training accuracy is 0.093750, time 218.496 samples/sec
+epoch 7, total_step 121840, total loss is 13.47 , inference loss is 6.07, weight deacy loss is 7.40, training accuracy is 0.312500, time 213.509 samples/sec
+epoch 7, total_step 121860, total loss is 17.31 , inference loss is 9.91, weight deacy loss is 7.40, training accuracy is 0.218750, time 213.570 samples/sec
+epoch 7, total_step 121880, total loss is 17.10 , inference loss is 9.70, weight deacy loss is 7.40, training accuracy is 0.375000, time 210.690 samples/sec
+epoch 7, total_step 121900, total loss is 18.17 , inference loss is 10.77, weight deacy loss is 7.40, training accuracy is 0.250000, time 200.753 samples/sec
+epoch 7, total_step 121920, total loss is 14.94 , inference loss is 7.54, weight deacy loss is 7.40, training accuracy is 0.250000, time 208.302 samples/sec
+epoch 7, total_step 121940, total loss is 16.99 , inference loss is 9.59, weight deacy loss is 7.40, training accuracy is 0.187500, time 226.663 samples/sec
+epoch 7, total_step 121960, total loss is 19.38 , inference loss is 11.98, weight deacy loss is 7.40, training accuracy is 0.218750, time 220.502 samples/sec
+epoch 7, total_step 121980, total loss is 19.14 , inference loss is 11.74, weight deacy loss is 7.40, training accuracy is 0.281250, time 299.189 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.963117000000018
+best_threshold_index 110 0.9862962962962963
+best_threshold_index 111 0.9848148148148148
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 107 0.9862962962962963
+best_threshold_index 107 0.9859259259259259
+best_threshold_index 110 0.9846296296296296
+best_threshold_index 110 0.9864814814814815
+best_threshold_index 110 0.9853703703703703
+best_threshold_index 110 0.9844444444444445
+best_threshold_index 110 0.9842592592592593
+[lfw][122000]XNorm: 22.335920
+[lfw][122000]Accuracy-Flip: 0.98450+-0.00742
+test accuracy is: 0.9845
+epoch 7, total_step 122000, total loss is 20.19 , inference loss is 12.79, weight deacy loss is 7.40, training accuracy is 0.125000, time 19.116 samples/sec
+epoch 7, total_step 122020, total loss is 23.67 , inference loss is 16.27, weight deacy loss is 7.40, training accuracy is 0.218750, time 212.665 samples/sec
+epoch 7, total_step 122040, total loss is 22.75 , inference loss is 15.35, weight deacy loss is 7.40, training accuracy is 0.218750, time 214.633 samples/sec
+epoch 7, total_step 122060, total loss is 20.47 , inference loss is 13.07, weight deacy loss is 7.40, training accuracy is 0.312500, time 209.743 samples/sec
+epoch 7, total_step 122080, total loss is 19.13 , inference loss is 11.74, weight deacy loss is 7.39, training accuracy is 0.187500, time 224.678 samples/sec
+epoch 7, total_step 122100, total loss is 20.36 , inference loss is 12.96, weight deacy loss is 7.39, training accuracy is 0.218750, time 11.748 samples/sec
+epoch 7, total_step 122120, total loss is 23.33 , inference loss is 15.93, weight deacy loss is 7.39, training accuracy is 0.187500, time 212.667 samples/sec
+epoch 7, total_step 122140, total loss is 16.72 , inference loss is 9.33, weight deacy loss is 7.39, training accuracy is 0.218750, time 206.926 samples/sec
+epoch 7, total_step 122160, total loss is 18.96 , inference loss is 11.57, weight deacy loss is 7.39, training accuracy is 0.218750, time 213.022 samples/sec
+epoch 7, total_step 122180, total loss is 21.14 , inference loss is 13.75, weight deacy loss is 7.39, training accuracy is 0.218750, time 217.303 samples/sec
+epoch 7, total_step 122200, total loss is 23.26 , inference loss is 15.87, weight deacy loss is 7.39, training accuracy is 0.218750, time 218.469 samples/sec
+epoch 7, total_step 122220, total loss is 20.25 , inference loss is 12.86, weight deacy loss is 7.39, training accuracy is 0.125000, time 214.792 samples/sec
+epoch 7, total_step 122240, total loss is 19.23 , inference loss is 11.84, weight deacy loss is 7.39, training accuracy is 0.156250, time 213.469 samples/sec
+epoch 7, total_step 122260, total loss is 16.33 , inference loss is 8.94, weight deacy loss is 7.39, training accuracy is 0.250000, time 215.907 samples/sec
+epoch 7, total_step 122280, total loss is 19.53 , inference loss is 12.14, weight deacy loss is 7.39, training accuracy is 0.250000, time 217.718 samples/sec
+epoch 7, total_step 122300, total loss is 17.30 , inference loss is 9.91, weight deacy loss is 7.39, training accuracy is 0.187500, time 297.464 samples/sec
+epoch 7, total_step 122320, total loss is 22.78 , inference loss is 15.39, weight deacy loss is 7.39, training accuracy is 0.125000, time 300.427 samples/sec
+epoch 7, total_step 122340, total loss is 16.67 , inference loss is 9.28, weight deacy loss is 7.39, training accuracy is 0.312500, time 196.780 samples/sec
+epoch 7, total_step 122360, total loss is 16.06 , inference loss is 8.67, weight deacy loss is 7.39, training accuracy is 0.281250, time 194.821 samples/sec
+epoch 7, total_step 122380, total loss is 18.19 , inference loss is 10.80, weight deacy loss is 7.39, training accuracy is 0.187500, time 209.884 samples/sec
+epoch 7, total_step 122400, total loss is 19.16 , inference loss is 11.78, weight deacy loss is 7.39, training accuracy is 0.156250, time 13.687 samples/sec
+epoch 7, total_step 122420, total loss is 19.06 , inference loss is 11.67, weight deacy loss is 7.39, training accuracy is 0.187500, time 210.084 samples/sec
+epoch 7, total_step 122440, total loss is 20.02 , inference loss is 12.64, weight deacy loss is 7.39, training accuracy is 0.343750, time 196.261 samples/sec
+epoch 7, total_step 122460, total loss is 18.96 , inference loss is 11.58, weight deacy loss is 7.38, training accuracy is 0.343750, time 218.332 samples/sec
+epoch 7, total_step 122480, total loss is 18.48 , inference loss is 11.09, weight deacy loss is 7.38, training accuracy is 0.187500, time 224.435 samples/sec
+epoch 7, total_step 122500, total loss is 16.31 , inference loss is 8.92, weight deacy loss is 7.38, training accuracy is 0.343750, time 196.233 samples/sec2022-12-01 08:17:22.536957: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 9046 of 10000
+2022-12-01 08:17:23.568423: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 7, total_step 122520, total loss is 20.41 , inference loss is 13.03, weight deacy loss is 7.38, training accuracy is 0.281250, time 189.402 samples/sec
+epoch 7, total_step 122540, total loss is 17.78 , inference loss is 10.40, weight deacy loss is 7.38, training accuracy is 0.281250, time 300.194 samples/sec
+epoch 7, total_step 122560, total loss is 16.35 , inference loss is 8.96, weight deacy loss is 7.38, training accuracy is 0.281250, time 229.672 samples/sec
+epoch 7, total_step 122580, total loss is 19.06 , inference loss is 11.68, weight deacy loss is 7.38, training accuracy is 0.250000, time 208.457 samples/sec
+epoch 7, total_step 122600, total loss is 19.42 , inference loss is 12.04, weight deacy loss is 7.38, training accuracy is 0.281250, time 209.413 samples/sec
+epoch 7, total_step 122620, total loss is 16.52 , inference loss is 9.14, weight deacy loss is 7.38, training accuracy is 0.250000, time 193.940 samples/sec
+epoch 7, total_step 122640, total loss is 14.84 , inference loss is 7.46, weight deacy loss is 7.38, training accuracy is 0.343750, time 227.172 samples/sec
+End of epoch 7
+epoch 8, total_step 122660, total loss is 20.51 , inference loss is 13.13, weight deacy loss is 7.38, training accuracy is 0.125000, time 202.326 samples/sec
+epoch 8, total_step 122680, total loss is 12.99 , inference loss is 5.62, weight deacy loss is 7.38, training accuracy is 0.593750, time 210.562 samples/sec
+epoch 8, total_step 122700, total loss is 13.47 , inference loss is 6.09, weight deacy loss is 7.38, training accuracy is 0.500000, time 14.430 samples/sec
+epoch 8, total_step 122720, total loss is 11.38 , inference loss is 4.00, weight deacy loss is 7.38, training accuracy is 0.593750, time 205.670 samples/sec
+epoch 8, total_step 122740, total loss is 11.26 , inference loss is 3.89, weight deacy loss is 7.38, training accuracy is 0.625000, time 214.173 samples/sec
+epoch 8, total_step 122760, total loss is 14.25 , inference loss is 6.87, weight deacy loss is 7.38, training accuracy is 0.562500, time 300.020 samples/sec
+epoch 8, total_step 122780, total loss is 12.50 , inference loss is 5.13, weight deacy loss is 7.38, training accuracy is 0.562500, time 219.517 samples/sec
+epoch 8, total_step 122800, total loss is 13.41 , inference loss is 6.04, weight deacy loss is 7.38, training accuracy is 0.656250, time 210.527 samples/sec
+epoch 8, total_step 122820, total loss is 9.79 , inference loss is 2.42, weight deacy loss is 7.37, training accuracy is 0.718750, time 213.831 samples/sec
+epoch 8, total_step 122840, total loss is 10.29 , inference loss is 2.91, weight deacy loss is 7.37, training accuracy is 0.656250, time 234.488 samples/sec
+epoch 8, total_step 122860, total loss is 11.19 , inference loss is 3.82, weight deacy loss is 7.37, training accuracy is 0.750000, time 208.949 samples/sec
+epoch 8, total_step 122880, total loss is 13.89 , inference loss is 6.52, weight deacy loss is 7.37, training accuracy is 0.593750, time 212.636 samples/sec
+epoch 8, total_step 122900, total loss is 11.93 , inference loss is 4.55, weight deacy loss is 7.37, training accuracy is 0.625000, time 209.905 samples/sec
+epoch 8, total_step 122920, total loss is 13.98 , inference loss is 6.61, weight deacy loss is 7.37, training accuracy is 0.437500, time 215.059 samples/sec
+epoch 8, total_step 122940, total loss is 14.38 , inference loss is 7.01, weight deacy loss is 7.37, training accuracy is 0.562500, time 210.012 samples/sec
+epoch 8, total_step 122960, total loss is 13.68 , inference loss is 6.31, weight deacy loss is 7.37, training accuracy is 0.531250, time 214.229 samples/sec
+epoch 8, total_step 122980, total loss is 15.07 , inference loss is 7.70, weight deacy loss is 7.37, training accuracy is 0.531250, time 209.925 samples/sec
+epoch 8, total_step 123000, total loss is 10.56 , inference loss is 3.19, weight deacy loss is 7.37, training accuracy is 0.625000, time 10.655 samples/sec
+epoch 8, total_step 123020, total loss is 13.80 , inference loss is 6.43, weight deacy loss is 7.37, training accuracy is 0.562500, time 297.028 samples/sec
+epoch 8, total_step 123040, total loss is 18.71 , inference loss is 11.35, weight deacy loss is 7.37, training accuracy is 0.468750, time 214.377 samples/sec
+epoch 8, total_step 123060, total loss is 12.80 , inference loss is 5.43, weight deacy loss is 7.37, training accuracy is 0.718750, time 205.147 samples/sec
+epoch 8, total_step 123080, total loss is 15.90 , inference loss is 8.53, weight deacy loss is 7.37, training accuracy is 0.562500, time 234.673 samples/sec
+epoch 8, total_step 123100, total loss is 11.16 , inference loss is 3.79, weight deacy loss is 7.37, training accuracy is 0.593750, time 211.263 samples/sec
+epoch 8, total_step 123120, total loss is 14.93 , inference loss is 7.56, weight deacy loss is 7.37, training accuracy is 0.593750, time 210.879 samples/sec
+epoch 8, total_step 123140, total loss is 12.92 , inference loss is 5.56, weight deacy loss is 7.37, training accuracy is 0.500000, time 209.859 samples/sec
+epoch 8, total_step 123160, total loss is 16.31 , inference loss is 8.95, weight deacy loss is 7.37, training accuracy is 0.437500, time 214.915 samples/sec
+epoch 8, total_step 123180, total loss is 15.03 , inference loss is 7.66, weight deacy loss is 7.37, training accuracy is 0.593750, time 212.966 samples/sec
+epoch 8, total_step 123200, total loss is 14.65 , inference loss is 7.29, weight deacy loss is 7.37, training accuracy is 0.500000, time 210.858 samples/sec
+epoch 8, total_step 123220, total loss is 14.80 , inference loss is 7.44, weight deacy loss is 7.36, training accuracy is 0.562500, time 202.038 samples/sec
+epoch 8, total_step 123240, total loss is 13.65 , inference loss is 6.28, weight deacy loss is 7.36, training accuracy is 0.437500, time 210.772 samples/sec
+epoch 8, total_step 123260, total loss is 14.60 , inference loss is 7.24, weight deacy loss is 7.36, training accuracy is 0.593750, time 211.044 samples/sec
+epoch 8, total_step 123280, total loss is 14.91 , inference loss is 7.55, weight deacy loss is 7.36, training accuracy is 0.500000, time 214.589 samples/sec
+epoch 8, total_step 123300, total loss is 14.75 , inference loss is 7.38, weight deacy loss is 7.36, training accuracy is 0.468750, time 15.279 samples/sec
+epoch 8, total_step 123320, total loss is 21.55 , inference loss is 14.19, weight deacy loss is 7.36, training accuracy is 0.375000, time 213.312 samples/sec
+epoch 8, total_step 123340, total loss is 12.83 , inference loss is 5.47, weight deacy loss is 7.36, training accuracy is 0.593750, time 206.897 samples/sec
+epoch 8, total_step 123360, total loss is 17.71 , inference loss is 10.34, weight deacy loss is 7.36, training accuracy is 0.343750, time 209.184 samples/sec
+epoch 8, total_step 123380, total loss is 13.23 , inference loss is 5.87, weight deacy loss is 7.36, training accuracy is 0.656250, time 225.952 samples/sec
+epoch 8, total_step 123400, total loss is 15.59 , inference loss is 8.23, weight deacy loss is 7.36, training accuracy is 0.531250, time 207.079 samples/sec
+epoch 8, total_step 123420, total loss is 17.33 , inference loss is 9.97, weight deacy loss is 7.36, training accuracy is 0.375000, time 211.529 samples/sec
+epoch 8, total_step 123440, total loss is 13.57 , inference loss is 6.21, weight deacy loss is 7.36, training accuracy is 0.500000, time 212.368 samples/sec
+epoch 8, total_step 123460, total loss is 18.52 , inference loss is 11.16, weight deacy loss is 7.36, training accuracy is 0.562500, time 211.216 samples/sec
+epoch 8, total_step 123480, total loss is 13.92 , inference loss is 6.56, weight deacy loss is 7.36, training accuracy is 0.656250, time 208.737 samples/sec
+epoch 8, total_step 123500, total loss is 12.74 , inference loss is 5.38, weight deacy loss is 7.36, training accuracy is 0.562500, time 204.970 samples/sec
+epoch 8, total_step 123520, total loss is 13.94 , inference loss is 6.58, weight deacy loss is 7.36, training accuracy is 0.437500, time 210.898 samples/sec
+epoch 8, total_step 123540, total loss is 13.34 , inference loss is 5.98, weight deacy loss is 7.36, training accuracy is 0.593750, time 213.112 samples/sec
+epoch 8, total_step 123560, total loss is 17.87 , inference loss is 10.51, weight deacy loss is 7.36, training accuracy is 0.437500, time 207.080 samples/sec
+epoch 8, total_step 123580, total loss is 15.65 , inference loss is 8.29, weight deacy loss is 7.36, training accuracy is 0.468750, time 221.340 samples/sec
+epoch 8, total_step 123600, total loss is 15.28 , inference loss is 7.92, weight deacy loss is 7.36, training accuracy is 0.468750, time 13.709 samples/sec
+epoch 8, total_step 123620, total loss is 14.07 , inference loss is 6.71, weight deacy loss is 7.36, training accuracy is 0.437500, time 215.771 samples/sec
+epoch 8, total_step 123640, total loss is 11.86 , inference loss is 4.50, weight deacy loss is 7.36, training accuracy is 0.406250, time 212.389 samples/sec
+epoch 8, total_step 123660, total loss is 15.14 , inference loss is 7.78, weight deacy loss is 7.36, training accuracy is 0.468750, time 219.810 samples/sec
+epoch 8, total_step 123680, total loss is 15.08 , inference loss is 7.73, weight deacy loss is 7.36, training accuracy is 0.437500, time 211.719 samples/sec
+epoch 8, total_step 123700, total loss is 15.22 , inference loss is 7.87, weight deacy loss is 7.36, training accuracy is 0.562500, time 207.387 samples/sec
+epoch 8, total_step 123720, total loss is 14.63 , inference loss is 7.27, weight deacy loss is 7.36, training accuracy is 0.500000, time 208.390 samples/sec
+epoch 8, total_step 123740, total loss is 14.17 , inference loss is 6.81, weight deacy loss is 7.36, training accuracy is 0.468750, time 213.901 samples/sec
+epoch 8, total_step 123760, total loss is 17.27 , inference loss is 9.92, weight deacy loss is 7.36, training accuracy is 0.531250, time 215.401 samples/sec
+epoch 8, total_step 123780, total loss is 14.92 , inference loss is 7.57, weight deacy loss is 7.36, training accuracy is 0.500000, time 216.970 samples/sec
+epoch 8, total_step 123800, total loss is 19.62 , inference loss is 12.26, weight deacy loss is 7.36, training accuracy is 0.281250, time 209.632 samples/sec
+epoch 8, total_step 123820, total loss is 18.28 , inference loss is 10.92, weight deacy loss is 7.36, training accuracy is 0.406250, time 207.134 samples/sec
+epoch 8, total_step 123840, total loss is 17.46 , inference loss is 10.10, weight deacy loss is 7.36, training accuracy is 0.343750, time 213.872 samples/sec
+epoch 8, total_step 123860, total loss is 19.37 , inference loss is 12.01, weight deacy loss is 7.36, training accuracy is 0.531250, time 210.864 samples/sec
+epoch 8, total_step 123880, total loss is 14.74 , inference loss is 7.39, weight deacy loss is 7.35, training accuracy is 0.562500, time 297.503 samples/sec
+epoch 8, total_step 123900, total loss is 14.01 , inference loss is 6.66, weight deacy loss is 7.35, training accuracy is 0.500000, time 14.007 samples/sec
+epoch 8, total_step 123920, total loss is 15.41 , inference loss is 8.06, weight deacy loss is 7.35, training accuracy is 0.531250, time 205.206 samples/sec
+epoch 8, total_step 123940, total loss is 14.28 , inference loss is 6.92, weight deacy loss is 7.35, training accuracy is 0.468750, time 210.971 samples/sec
+epoch 8, total_step 123960, total loss is 17.31 , inference loss is 9.95, weight deacy loss is 7.35, training accuracy is 0.468750, time 212.811 samples/sec
+epoch 8, total_step 123980, total loss is 15.83 , inference loss is 8.47, weight deacy loss is 7.35, training accuracy is 0.437500, time 217.854 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.891823
+best_threshold_index 122 0.9862962962962963
+best_threshold_index 122 0.9853703703703703
+best_threshold_index 122 0.9848148148148148
+best_threshold_index 122 0.9861111111111112
+best_threshold_index 119 0.9853703703703703
+best_threshold_index 122 0.9848148148148148
+best_threshold_index 122 0.9859259259259259
+best_threshold_index 122 0.9857407407407407
+best_threshold_index 122 0.9851851851851852
+best_threshold_index 122 0.9853703703703703
+[lfw][124000]XNorm: 20.415726
+[lfw][124000]Accuracy-Flip: 0.98483+-0.00462
+test accuracy is: 0.9848333333333332
+epoch 8, total_step 124000, total loss is 17.27 , inference loss is 9.92, weight deacy loss is 7.35, training accuracy is 0.437500, time 19.101 samples/sec
+epoch 8, total_step 124020, total loss is 13.91 , inference loss is 6.56, weight deacy loss is 7.35, training accuracy is 0.562500, time 219.810 samples/sec
+epoch 8, total_step 124040, total loss is 15.25 , inference loss is 7.90, weight deacy loss is 7.35, training accuracy is 0.468750, time 214.713 samples/sec
+epoch 8, total_step 124060, total loss is 17.83 , inference loss is 10.48, weight deacy loss is 7.35, training accuracy is 0.468750, time 213.563 samples/sec
+epoch 8, total_step 124080, total loss is 17.35 , inference loss is 10.00, weight deacy loss is 7.35, training accuracy is 0.406250, time 217.351 samples/sec
+epoch 8, total_step 124100, total loss is 14.42 , inference loss is 7.06, weight deacy loss is 7.35, training accuracy is 0.468750, time 299.935 samples/sec
+epoch 8, total_step 124120, total loss is 14.18 , inference loss is 6.83, weight deacy loss is 7.35, training accuracy is 0.375000, time 297.822 samples/sec
+epoch 8, total_step 124140, total loss is 18.49 , inference loss is 11.14, weight deacy loss is 7.35, training accuracy is 0.468750, time 220.515 samples/sec
+epoch 8, total_step 124160, total loss is 17.77 , inference loss is 10.42, weight deacy loss is 7.35, training accuracy is 0.437500, time 209.812 samples/sec
+epoch 8, total_step 124180, total loss is 14.34 , inference loss is 6.99, weight deacy loss is 7.35, training accuracy is 0.531250, time 210.433 samples/sec
+epoch 8, total_step 124200, total loss is 12.83 , inference loss is 5.48, weight deacy loss is 7.35, training accuracy is 0.687500, time 12.396 samples/sec
+epoch 8, total_step 124220, total loss is 15.99 , inference loss is 8.64, weight deacy loss is 7.35, training accuracy is 0.343750, time 219.352 samples/sec
+epoch 8, total_step 124240, total loss is 16.99 , inference loss is 9.64, weight deacy loss is 7.35, training accuracy is 0.468750, time 211.395 samples/sec
+epoch 8, total_step 124260, total loss is 14.09 , inference loss is 6.74, weight deacy loss is 7.35, training accuracy is 0.437500, time 208.112 samples/sec
+epoch 8, total_step 124280, total loss is 19.38 , inference loss is 12.03, weight deacy loss is 7.35, training accuracy is 0.437500, time 212.529 samples/sec
+epoch 8, total_step 124300, total loss is 14.49 , inference loss is 7.15, weight deacy loss is 7.35, training accuracy is 0.375000, time 213.301 samples/sec
+epoch 8, total_step 124320, total loss is 17.39 , inference loss is 10.05, weight deacy loss is 7.35, training accuracy is 0.593750, time 204.524 samples/sec
+epoch 8, total_step 124340, total loss is 19.86 , inference loss is 12.51, weight deacy loss is 7.35, training accuracy is 0.406250, time 226.577 samples/sec
+epoch 8, total_step 124360, total loss is 16.19 , inference loss is 8.84, weight deacy loss is 7.35, training accuracy is 0.468750, time 212.991 samples/sec
+epoch 8, total_step 124380, total loss is 18.29 , inference loss is 10.94, weight deacy loss is 7.35, training accuracy is 0.437500, time 298.653 samples/sec
+epoch 8, total_step 124400, total loss is 16.28 , inference loss is 8.94, weight deacy loss is 7.35, training accuracy is 0.468750, time 203.308 samples/sec
+epoch 8, total_step 124420, total loss is 12.40 , inference loss is 5.06, weight deacy loss is 7.35, training accuracy is 0.500000, time 204.008 samples/sec
+epoch 8, total_step 124440, total loss is 20.80 , inference loss is 13.46, weight deacy loss is 7.35, training accuracy is 0.218750, time 212.048 samples/sec
+epoch 8, total_step 124460, total loss is 15.01 , inference loss is 7.67, weight deacy loss is 7.34, training accuracy is 0.437500, time 204.880 samples/sec
+epoch 8, total_step 124480, total loss is 17.56 , inference loss is 10.22, weight deacy loss is 7.34, training accuracy is 0.500000, time 244.817 samples/sec
+epoch 8, total_step 124500, total loss is 15.20 , inference loss is 7.85, weight deacy loss is 7.34, training accuracy is 0.500000, time 13.349 samples/sec
+epoch 8, total_step 124520, total loss is 17.69 , inference loss is 10.34, weight deacy loss is 7.34, training accuracy is 0.531250, time 220.893 samples/sec
+epoch 8, total_step 124540, total loss is 19.15 , inference loss is 11.81, weight deacy loss is 7.34, training accuracy is 0.406250, time 210.198 samples/sec
+epoch 8, total_step 124560, total loss is 19.08 , inference loss is 11.73, weight deacy loss is 7.34, training accuracy is 0.375000, time 206.672 samples/sec
+epoch 8, total_step 124580, total loss is 14.43 , inference loss is 7.08, weight deacy loss is 7.34, training accuracy is 0.406250, time 209.942 samples/sec
+epoch 8, total_step 124600, total loss is 13.33 , inference loss is 5.99, weight deacy loss is 7.34, training accuracy is 0.625000, time 226.052 samples/sec
+epoch 8, total_step 124620, total loss is 17.44 , inference loss is 10.10, weight deacy loss is 7.34, training accuracy is 0.500000, time 217.247 samples/sec
+epoch 8, total_step 124640, total loss is 16.22 , inference loss is 8.88, weight deacy loss is 7.34, training accuracy is 0.437500, time 228.258 samples/sec
+epoch 8, total_step 124660, total loss is 12.42 , inference loss is 5.08, weight deacy loss is 7.34, training accuracy is 0.531250, time 299.126 samples/sec
+epoch 8, total_step 124680, total loss is 12.48 , inference loss is 5.14, weight deacy loss is 7.34, training accuracy is 0.500000, time 212.538 samples/sec
+epoch 8, total_step 124700, total loss is 13.20 , inference loss is 5.87, weight deacy loss is 7.34, training accuracy is 0.531250, time 218.441 samples/sec
+epoch 8, total_step 124720, total loss is 13.72 , inference loss is 6.38, weight deacy loss is 7.34, training accuracy is 0.468750, time 229.569 samples/sec
+epoch 8, total_step 124740, total loss is 15.28 , inference loss is 7.94, weight deacy loss is 7.34, training accuracy is 0.562500, time 212.489 samples/sec
+epoch 8, total_step 124760, total loss is 11.11 , inference loss is 3.77, weight deacy loss is 7.34, training accuracy is 0.625000, time 234.811 samples/sec
+epoch 8, total_step 124780, total loss is 12.02 , inference loss is 4.69, weight deacy loss is 7.34, training accuracy is 0.656250, time 209.586 samples/sec
+epoch 8, total_step 124800, total loss is 14.99 , inference loss is 7.65, weight deacy loss is 7.34, training accuracy is 0.531250, time 10.650 samples/sec
+epoch 8, total_step 124820, total loss is 12.37 , inference loss is 5.03, weight deacy loss is 7.34, training accuracy is 0.468750, time 210.082 samples/sec
+epoch 8, total_step 124840, total loss is 11.53 , inference loss is 4.19, weight deacy loss is 7.34, training accuracy is 0.625000, time 211.514 samples/sec
+epoch 8, total_step 124860, total loss is 15.58 , inference loss is 8.24, weight deacy loss is 7.34, training accuracy is 0.406250, time 214.540 samples/sec
+epoch 8, total_step 124880, total loss is 15.93 , inference loss is 8.59, weight deacy loss is 7.33, training accuracy is 0.468750, time 220.721 samples/sec
+epoch 8, total_step 124900, total loss is 13.78 , inference loss is 6.44, weight deacy loss is 7.33, training accuracy is 0.531250, time 222.711 samples/sec
+epoch 8, total_step 124920, total loss is 11.20 , inference loss is 3.87, weight deacy loss is 7.33, training accuracy is 0.625000, time 300.633 samples/sec
+epoch 8, total_step 124940, total loss is 13.87 , inference loss is 6.54, weight deacy loss is 7.33, training accuracy is 0.593750, time 211.193 samples/sec
+epoch 8, total_step 124960, total loss is 14.44 , inference loss is 7.10, weight deacy loss is 7.33, training accuracy is 0.500000, time 204.150 samples/sec
+epoch 8, total_step 124980, total loss is 12.01 , inference loss is 4.67, weight deacy loss is 7.33, training accuracy is 0.625000, time 206.731 samples/sec
+epoch 8, total_step 125000, total loss is 12.98 , inference loss is 5.65, weight deacy loss is 7.33, training accuracy is 0.593750, time 214.720 samples/sec
+epoch 8, total_step 125020, total loss is 18.23 , inference loss is 10.90, weight deacy loss is 7.33, training accuracy is 0.343750, time 204.860 samples/sec
+epoch 8, total_step 125040, total loss is 12.53 , inference loss is 5.20, weight deacy loss is 7.33, training accuracy is 0.468750, time 208.146 samples/sec
+epoch 8, total_step 125060, total loss is 11.65 , inference loss is 4.32, weight deacy loss is 7.33, training accuracy is 0.531250, time 214.517 samples/sec
+epoch 8, total_step 125080, total loss is 12.30 , inference loss is 4.97, weight deacy loss is 7.33, training accuracy is 0.500000, time 208.828 samples/sec
+epoch 8, total_step 125100, total loss is 15.44 , inference loss is 8.11, weight deacy loss is 7.33, training accuracy is 0.375000, time 14.372 samples/sec
+epoch 8, total_step 125120, total loss is 15.76 , inference loss is 8.43, weight deacy loss is 7.33, training accuracy is 0.406250, time 208.148 samples/sec
+epoch 8, total_step 125140, total loss is 16.02 , inference loss is 8.69, weight deacy loss is 7.33, training accuracy is 0.406250, time 216.686 samples/sec
+epoch 8, total_step 125160, total loss is 14.31 , inference loss is 6.98, weight deacy loss is 7.33, training accuracy is 0.500000, time 211.704 samples/sec
+epoch 8, total_step 125180, total loss is 15.33 , inference loss is 8.00, weight deacy loss is 7.33, training accuracy is 0.375000, time 299.310 samples/sec
+epoch 8, total_step 125200, total loss is 13.82 , inference loss is 6.49, weight deacy loss is 7.33, training accuracy is 0.500000, time 227.884 samples/sec
+epoch 8, total_step 125220, total loss is 12.46 , inference loss is 5.13, weight deacy loss is 7.33, training accuracy is 0.500000, time 211.038 samples/sec
+epoch 8, total_step 125240, total loss is 16.45 , inference loss is 9.12, weight deacy loss is 7.33, training accuracy is 0.281250, time 228.073 samples/sec
+epoch 8, total_step 125260, total loss is 13.90 , inference loss is 6.57, weight deacy loss is 7.33, training accuracy is 0.531250, time 202.489 samples/sec
+epoch 8, total_step 125280, total loss is 12.82 , inference loss is 5.49, weight deacy loss is 7.32, training accuracy is 0.531250, time 213.595 samples/sec
+epoch 8, total_step 125300, total loss is 12.78 , inference loss is 5.46, weight deacy loss is 7.32, training accuracy is 0.562500, time 212.350 samples/sec
+epoch 8, total_step 125320, total loss is 13.03 , inference loss is 5.71, weight deacy loss is 7.32, training accuracy is 0.593750, time 193.924 samples/sec
+epoch 8, total_step 125340, total loss is 12.28 , inference loss is 4.96, weight deacy loss is 7.32, training accuracy is 0.500000, time 210.140 samples/sec
+epoch 8, total_step 125360, total loss is 13.22 , inference loss is 5.90, weight deacy loss is 7.32, training accuracy is 0.406250, time 223.247 samples/sec
+epoch 8, total_step 125380, total loss is 17.90 , inference loss is 10.58, weight deacy loss is 7.32, training accuracy is 0.406250, time 217.763 samples/sec
+epoch 8, total_step 125400, total loss is 11.93 , inference loss is 4.61, weight deacy loss is 7.32, training accuracy is 0.593750, time 12.850 samples/sec
+epoch 8, total_step 125420, total loss is 17.47 , inference loss is 10.14, weight deacy loss is 7.32, training accuracy is 0.312500, time 200.075 samples/sec
+epoch 8, total_step 125440, total loss is 15.00 , inference loss is 7.68, weight deacy loss is 7.32, training accuracy is 0.500000, time 209.017 samples/sec
+epoch 8, total_step 125460, total loss is 15.91 , inference loss is 8.59, weight deacy loss is 7.32, training accuracy is 0.343750, time 297.248 samples/sec
+epoch 8, total_step 125480, total loss is 10.13 , inference loss is 2.81, weight deacy loss is 7.32, training accuracy is 0.750000, time 215.637 samples/sec
+epoch 8, total_step 125500, total loss is 14.24 , inference loss is 6.92, weight deacy loss is 7.32, training accuracy is 0.562500, time 207.348 samples/sec
+epoch 8, total_step 125520, total loss is 11.17 , inference loss is 3.85, weight deacy loss is 7.32, training accuracy is 0.718750, time 215.719 samples/sec
+epoch 8, total_step 125540, total loss is 12.39 , inference loss is 5.07, weight deacy loss is 7.32, training accuracy is 0.531250, time 213.820 samples/sec
+epoch 8, total_step 125560, total loss is 14.21 , inference loss is 6.89, weight deacy loss is 7.32, training accuracy is 0.687500, time 218.657 samples/sec
+epoch 8, total_step 125580, total loss is 9.28 , inference loss is 1.96, weight deacy loss is 7.32, training accuracy is 0.625000, time 209.366 samples/sec
+epoch 8, total_step 125600, total loss is 11.17 , inference loss is 3.85, weight deacy loss is 7.32, training accuracy is 0.750000, time 210.426 samples/sec
+epoch 8, total_step 125620, total loss is 14.19 , inference loss is 6.87, weight deacy loss is 7.32, training accuracy is 0.500000, time 219.198 samples/sec
+epoch 8, total_step 125640, total loss is 17.70 , inference loss is 10.39, weight deacy loss is 7.32, training accuracy is 0.437500, time 221.904 samples/sec
+epoch 8, total_step 125660, total loss is 16.14 , inference loss is 8.83, weight deacy loss is 7.32, training accuracy is 0.468750, time 212.129 samples/sec
+epoch 8, total_step 125680, total loss is 12.44 , inference loss is 5.13, weight deacy loss is 7.32, training accuracy is 0.750000, time 205.844 samples/sec
+epoch 8, total_step 125700, total loss is 16.46 , inference loss is 9.15, weight deacy loss is 7.31, training accuracy is 0.437500, time 14.233 samples/sec
+epoch 8, total_step 125720, total loss is 12.81 , inference loss is 5.50, weight deacy loss is 7.31, training accuracy is 0.562500, time 298.762 samples/sec
+epoch 8, total_step 125740, total loss is 15.51 , inference loss is 8.19, weight deacy loss is 7.31, training accuracy is 0.437500, time 206.625 samples/sec
+epoch 8, total_step 125760, total loss is 13.66 , inference loss is 6.35, weight deacy loss is 7.31, training accuracy is 0.593750, time 225.182 samples/sec
+epoch 8, total_step 125780, total loss is 16.64 , inference loss is 9.33, weight deacy loss is 7.31, training accuracy is 0.500000, time 212.078 samples/sec
+epoch 8, total_step 125800, total loss is 10.94 , inference loss is 3.62, weight deacy loss is 7.31, training accuracy is 0.625000, time 210.400 samples/sec
+epoch 8, total_step 125820, total loss is 14.35 , inference loss is 7.04, weight deacy loss is 7.31, training accuracy is 0.468750, time 217.966 samples/sec
+epoch 8, total_step 125840, total loss is 15.60 , inference loss is 8.28, weight deacy loss is 7.31, training accuracy is 0.406250, time 218.977 samples/sec
+epoch 8, total_step 125860, total loss is 14.18 , inference loss is 6.87, weight deacy loss is 7.31, training accuracy is 0.468750, time 217.283 samples/sec
+epoch 8, total_step 125880, total loss is 11.77 , inference loss is 4.46, weight deacy loss is 7.31, training accuracy is 0.562500, time 216.525 samples/sec
+epoch 8, total_step 125900, total loss is 15.39 , inference loss is 8.08, weight deacy loss is 7.31, training accuracy is 0.562500, time 207.056 samples/sec
+epoch 8, total_step 125920, total loss is 15.80 , inference loss is 8.49, weight deacy loss is 7.31, training accuracy is 0.406250, time 212.543 samples/sec
+epoch 8, total_step 125940, total loss is 19.92 , inference loss is 12.61, weight deacy loss is 7.31, training accuracy is 0.406250, time 217.509 samples/sec
+epoch 8, total_step 125960, total loss is 19.24 , inference loss is 11.93, weight deacy loss is 7.31, training accuracy is 0.406250, time 214.484 samples/sec
+epoch 8, total_step 125980, total loss is 17.10 , inference loss is 9.79, weight deacy loss is 7.31, training accuracy is 0.343750, time 216.500 samples/sec
+testing verification..
+(12000, 512)
+infer time 22.384251999999975
+best_threshold_index 119 0.9877777777777778
+best_threshold_index 117 0.9866666666666667
+best_threshold_index 119 0.9872222222222222
+best_threshold_index 113 0.9879629629629629
+best_threshold_index 117 0.987037037037037
+best_threshold_index 119 0.987037037037037
+best_threshold_index 117 0.9877777777777778
+best_threshold_index 117 0.9868518518518519
+best_threshold_index 117 0.9859259259259259
+best_threshold_index 117 0.9864814814814815
+[lfw][126000]XNorm: 22.383718
+[lfw][126000]Accuracy-Flip: 0.98583+-0.00634
+test accuracy is: 0.9858333333333335
+epoch 8, total_step 126000, total loss is 14.85 , inference loss is 7.54, weight deacy loss is 7.31, training accuracy is 0.531250, time 19.458 samples/sec
+epoch 8, total_step 126020, total loss is 12.09 , inference loss is 4.79, weight deacy loss is 7.31, training accuracy is 0.562500, time 207.461 samples/sec
+epoch 8, total_step 126040, total loss is 14.93 , inference loss is 7.63, weight deacy loss is 7.31, training accuracy is 0.500000, time 222.626 samples/sec
+epoch 8, total_step 126060, total loss is 13.98 , inference loss is 6.67, weight deacy loss is 7.31, training accuracy is 0.406250, time 221.506 samples/sec
+epoch 8, total_step 126080, total loss is 16.86 , inference loss is 9.55, weight deacy loss is 7.31, training accuracy is 0.500000, time 212.583 samples/sec
+epoch 8, total_step 126100, total loss is 17.72 , inference loss is 10.41, weight deacy loss is 7.31, training accuracy is 0.343750, time 222.888 samples/sec
+epoch 8, total_step 126120, total loss is 15.37 , inference loss is 8.07, weight deacy loss is 7.31, training accuracy is 0.531250, time 231.696 samples/sec
+epoch 8, total_step 126140, total loss is 13.04 , inference loss is 5.74, weight deacy loss is 7.31, training accuracy is 0.531250, time 212.438 samples/sec
+epoch 8, total_step 126160, total loss is 13.86 , inference loss is 6.55, weight deacy loss is 7.30, training accuracy is 0.281250, time 211.232 samples/sec
+epoch 8, total_step 126180, total loss is 16.69 , inference loss is 9.38, weight deacy loss is 7.30, training accuracy is 0.531250, time 212.723 samples/sec
+epoch 8, total_step 126200, total loss is 11.48 , inference loss is 4.17, weight deacy loss is 7.30, training accuracy is 0.531250, time 221.634 samples/sec
+epoch 8, total_step 126220, total loss is 14.57 , inference loss is 7.27, weight deacy loss is 7.30, training accuracy is 0.468750, time 218.682 samples/sec
+epoch 8, total_step 126240, total loss is 17.32 , inference loss is 10.01, weight deacy loss is 7.30, training accuracy is 0.375000, time 237.769 samples/sec
+epoch 8, total_step 126260, total loss is 16.30 , inference loss is 8.99, weight deacy loss is 7.30, training accuracy is 0.500000, time 216.719 samples/sec
+epoch 8, total_step 126280, total loss is 14.19 , inference loss is 6.89, weight deacy loss is 7.30, training accuracy is 0.531250, time 211.509 samples/sec
+epoch 8, total_step 126300, total loss is 20.86 , inference loss is 13.56, weight deacy loss is 7.30, training accuracy is 0.312500, time 14.079 samples/sec
+epoch 8, total_step 126320, total loss is 21.58 , inference loss is 14.28, weight deacy loss is 7.30, training accuracy is 0.406250, time 209.542 samples/sec
+epoch 8, total_step 126340, total loss is 16.81 , inference loss is 9.50, weight deacy loss is 7.30, training accuracy is 0.343750, time 219.102 samples/sec
+epoch 8, total_step 126360, total loss is 14.81 , inference loss is 7.51, weight deacy loss is 7.30, training accuracy is 0.500000, time 221.030 samples/sec
+epoch 8, total_step 126380, total loss is 17.41 , inference loss is 10.11, weight deacy loss is 7.30, training accuracy is 0.343750, time 212.215 samples/sec
+epoch 8, total_step 126400, total loss is 15.26 , inference loss is 7.96, weight deacy loss is 7.30, training accuracy is 0.406250, time 214.075 samples/sec
+epoch 8, total_step 126420, total loss is 17.09 , inference loss is 9.79, weight deacy loss is 7.30, training accuracy is 0.312500, time 211.772 samples/sec
+epoch 8, total_step 126440, total loss is 20.22 , inference loss is 12.92, weight deacy loss is 7.30, training accuracy is 0.281250, time 205.090 samples/sec
+epoch 8, total_step 126460, total loss is 19.17 , inference loss is 11.87, weight deacy loss is 7.30, training accuracy is 0.406250, time 216.161 samples/sec
+epoch 8, total_step 126480, total loss is 13.69 , inference loss is 6.39, weight deacy loss is 7.30, training accuracy is 0.468750, time 210.819 samples/sec
+epoch 8, total_step 126500, total loss is 16.08 , inference loss is 8.78, weight deacy loss is 7.30, training accuracy is 0.375000, time 216.453 samples/sec
+epoch 8, total_step 126520, total loss is 15.56 , inference loss is 8.26, weight deacy loss is 7.30, training accuracy is 0.375000, time 215.702 samples/sec
+epoch 8, total_step 126540, total loss is 15.07 , inference loss is 7.78, weight deacy loss is 7.30, training accuracy is 0.468750, time 211.878 samples/sec
+epoch 8, total_step 126560, total loss is 17.66 , inference loss is 10.37, weight deacy loss is 7.30, training accuracy is 0.406250, time 299.012 samples/sec
+epoch 8, total_step 126580, total loss is 14.07 , inference loss is 6.77, weight deacy loss is 7.30, training accuracy is 0.531250, time 203.627 samples/sec
+epoch 8, total_step 126600, total loss is 15.08 , inference loss is 7.79, weight deacy loss is 7.30, training accuracy is 0.406250, time 12.132 samples/sec
+epoch 8, total_step 126620, total loss is 16.79 , inference loss is 9.49, weight deacy loss is 7.30, training accuracy is 0.468750, time 204.737 samples/sec
+epoch 8, total_step 126640, total loss is 14.72 , inference loss is 7.42, weight deacy loss is 7.30, training accuracy is 0.406250, time 213.797 samples/sec
+epoch 8, total_step 126660, total loss is 13.30 , inference loss is 6.00, weight deacy loss is 7.30, training accuracy is 0.531250, time 194.347 samples/sec
+epoch 8, total_step 126680, total loss is 18.19 , inference loss is 10.89, weight deacy loss is 7.29, training accuracy is 0.312500, time 204.822 samples/sec
+epoch 8, total_step 126700, total loss is 21.39 , inference loss is 14.09, weight deacy loss is 7.29, training accuracy is 0.250000, time 204.257 samples/sec
+epoch 8, total_step 126720, total loss is 15.84 , inference loss is 8.54, weight deacy loss is 7.29, training accuracy is 0.437500, time 196.243 samples/sec
+epoch 8, total_step 126740, total loss is 13.35 , inference loss is 6.06, weight deacy loss is 7.29, training accuracy is 0.437500, time 210.916 samples/sec
+epoch 8, total_step 126760, total loss is 19.41 , inference loss is 12.12, weight deacy loss is 7.29, training accuracy is 0.343750, time 208.750 samples/sec
+epoch 8, total_step 126780, total loss is 20.65 , inference loss is 13.36, weight deacy loss is 7.29, training accuracy is 0.218750, time 217.153 samples/sec
+epoch 8, total_step 126800, total loss is 14.73 , inference loss is 7.44, weight deacy loss is 7.29, training accuracy is 0.531250, time 213.741 samples/sec
+epoch 8, total_step 126820, total loss is 18.96 , inference loss is 11.67, weight deacy loss is 7.29, training accuracy is 0.312500, time 297.924 samples/sec
+epoch 8, total_step 126840, total loss is 19.11 , inference loss is 11.82, weight deacy loss is 7.29, training accuracy is 0.250000, time 227.617 samples/sec
+epoch 8, total_step 126860, total loss is 14.70 , inference loss is 7.41, weight deacy loss is 7.29, training accuracy is 0.375000, time 233.894 samples/sec
+epoch 8, total_step 126880, total loss is 13.85 , inference loss is 6.56, weight deacy loss is 7.29, training accuracy is 0.468750, time 207.692 samples/sec
+epoch 8, total_step 126900, total loss is 16.97 , inference loss is 9.68, weight deacy loss is 7.29, training accuracy is 0.437500, time 13.339 samples/sec
+epoch 8, total_step 126920, total loss is 15.81 , inference loss is 8.52, weight deacy loss is 7.29, training accuracy is 0.406250, time 217.937 samples/sec
+epoch 8, total_step 126940, total loss is 15.30 , inference loss is 8.01, weight deacy loss is 7.29, training accuracy is 0.625000, time 212.372 samples/sec
+epoch 8, total_step 126960, total loss is 17.62 , inference loss is 10.34, weight deacy loss is 7.29, training accuracy is 0.250000, time 215.674 samples/sec
+epoch 8, total_step 126980, total loss is 13.39 , inference loss is 6.10, weight deacy loss is 7.29, training accuracy is 0.468750, time 206.939 samples/sec
+epoch 8, total_step 127000, total loss is 17.88 , inference loss is 10.60, weight deacy loss is 7.29, training accuracy is 0.343750, time 209.527 samples/sec
+epoch 8, total_step 127020, total loss is 11.29 , inference loss is 4.01, weight deacy loss is 7.29, training accuracy is 0.625000, time 216.214 samples/sec
+epoch 8, total_step 127040, total loss is 14.97 , inference loss is 7.68, weight deacy loss is 7.29, training accuracy is 0.687500, time 206.711 samples/sec
+epoch 8, total_step 127060, total loss is 17.67 , inference loss is 10.39, weight deacy loss is 7.29, training accuracy is 0.375000, time 196.679 samples/sec
+epoch 8, total_step 127080, total loss is 13.34 , inference loss is 6.06, weight deacy loss is 7.29, training accuracy is 0.375000, time 226.908 samples/sec
+epoch 8, total_step 127100, total loss is 13.67 , inference loss is 6.39, weight deacy loss is 7.29, training accuracy is 0.437500, time 224.518 samples/sec
+epoch 8, total_step 127120, total loss is 15.84 , inference loss is 8.56, weight deacy loss is 7.29, training accuracy is 0.218750, time 207.651 samples/sec
+epoch 8, total_step 127140, total loss is 20.01 , inference loss is 12.72, weight deacy loss is 7.28, training accuracy is 0.250000, time 211.296 samples/sec
+epoch 8, total_step 127160, total loss is 18.74 , inference loss is 11.45, weight deacy loss is 7.28, training accuracy is 0.500000, time 208.764 samples/sec
+epoch 8, total_step 127180, total loss is 15.51 , inference loss is 8.23, weight deacy loss is 7.28, training accuracy is 0.437500, time 216.201 samples/sec
+epoch 8, total_step 127200, total loss is 18.75 , inference loss is 11.46, weight deacy loss is 7.28, training accuracy is 0.406250, time 13.738 samples/sec
+epoch 8, total_step 127220, total loss is 17.46 , inference loss is 10.18, weight deacy loss is 7.28, training accuracy is 0.437500, time 212.099 samples/sec
+epoch 8, total_step 127240, total loss is 18.07 , inference loss is 10.79, weight deacy loss is 7.28, training accuracy is 0.343750, time 226.340 samples/sec
+epoch 8, total_step 127260, total loss is 20.89 , inference loss is 13.61, weight deacy loss is 7.28, training accuracy is 0.406250, time 219.382 samples/sec
+epoch 8, total_step 127280, total loss is 17.23 , inference loss is 9.95, weight deacy loss is 7.28, training accuracy is 0.406250, time 210.389 samples/sec
+epoch 8, total_step 127300, total loss is 15.87 , inference loss is 8.59, weight deacy loss is 7.28, training accuracy is 0.375000, time 204.757 samples/sec
+epoch 8, total_step 127320, total loss is 15.75 , inference loss is 8.47, weight deacy loss is 7.28, training accuracy is 0.437500, time 208.048 samples/sec
+epoch 8, total_step 127340, total loss is 12.71 , inference loss is 5.43, weight deacy loss is 7.28, training accuracy is 0.437500, time 211.190 samples/sec
+epoch 8, total_step 127360, total loss is 17.82 , inference loss is 10.54, weight deacy loss is 7.28, training accuracy is 0.312500, time 299.843 samples/sec
+epoch 8, total_step 127380, total loss is 16.44 , inference loss is 9.16, weight deacy loss is 7.28, training accuracy is 0.375000, time 216.035 samples/sec
+epoch 8, total_step 127400, total loss is 16.65 , inference loss is 9.38, weight deacy loss is 7.28, training accuracy is 0.343750, time 211.334 samples/sec
+epoch 8, total_step 127420, total loss is 16.88 , inference loss is 9.60, weight deacy loss is 7.28, training accuracy is 0.375000, time 204.324 samples/sec
+epoch 8, total_step 127440, total loss is 19.91 , inference loss is 12.63, weight deacy loss is 7.28, training accuracy is 0.406250, time 210.482 samples/sec
+epoch 8, total_step 127460, total loss is 18.86 , inference loss is 11.58, weight deacy loss is 7.28, training accuracy is 0.312500, time 226.996 samples/sec
+epoch 8, total_step 127480, total loss is 16.24 , inference loss is 8.96, weight deacy loss is 7.28, training accuracy is 0.437500, time 226.889 samples/sec
+epoch 8, total_step 127500, total loss is 17.81 , inference loss is 10.54, weight deacy loss is 7.28, training accuracy is 0.343750, time 12.458 samples/sec
+epoch 8, total_step 127520, total loss is 19.80 , inference loss is 12.53, weight deacy loss is 7.28, training accuracy is 0.343750, time 214.565 samples/sec
+epoch 8, total_step 127540, total loss is 20.26 , inference loss is 12.99, weight deacy loss is 7.28, training accuracy is 0.406250, time 215.278 samples/sec
+epoch 8, total_step 127560, total loss is 15.84 , inference loss is 8.57, weight deacy loss is 7.28, training accuracy is 0.375000, time 208.060 samples/sec
+epoch 8, total_step 127580, total loss is 13.32 , inference loss is 6.04, weight deacy loss is 7.28, training accuracy is 0.500000, time 200.625 samples/sec
+epoch 8, total_step 127600, total loss is 17.24 , inference loss is 9.96, weight deacy loss is 7.27, training accuracy is 0.375000, time 213.690 samples/sec
+epoch 8, total_step 127620, total loss is 12.85 , inference loss is 5.58, weight deacy loss is 7.27, training accuracy is 0.562500, time 288.907 samples/sec
+epoch 8, total_step 127640, total loss is 15.54 , inference loss is 8.27, weight deacy loss is 7.27, training accuracy is 0.343750, time 234.120 samples/sec
+epoch 8, total_step 127660, total loss is 13.33 , inference loss is 6.05, weight deacy loss is 7.27, training accuracy is 0.437500, time 215.033 samples/sec
+epoch 8, total_step 127680, total loss is 13.03 , inference loss is 5.75, weight deacy loss is 7.27, training accuracy is 0.562500, time 206.757 samples/sec
+epoch 8, total_step 127700, total loss is 17.15 , inference loss is 9.88, weight deacy loss is 7.27, training accuracy is 0.250000, time 213.650 samples/sec
+epoch 8, total_step 127720, total loss is 16.83 , inference loss is 9.55, weight deacy loss is 7.27, training accuracy is 0.468750, time 209.492 samples/sec
+epoch 8, total_step 127740, total loss is 16.73 , inference loss is 9.46, weight deacy loss is 7.27, training accuracy is 0.406250, time 218.320 samples/sec
+epoch 8, total_step 127760, total loss is 14.74 , inference loss is 7.47, weight deacy loss is 7.27, training accuracy is 0.343750, time 229.834 samples/sec
+epoch 8, total_step 127780, total loss is 13.81 , inference loss is 6.54, weight deacy loss is 7.27, training accuracy is 0.500000, time 213.526 samples/sec
+epoch 8, total_step 127800, total loss is 18.34 , inference loss is 11.07, weight deacy loss is 7.27, training accuracy is 0.468750, time 14.039 samples/sec
+epoch 8, total_step 127820, total loss is 13.96 , inference loss is 6.69, weight deacy loss is 7.27, training accuracy is 0.406250, time 226.091 samples/sec
+epoch 8, total_step 127840, total loss is 15.17 , inference loss is 7.90, weight deacy loss is 7.27, training accuracy is 0.437500, time 298.622 samples/sec
+epoch 8, total_step 127860, total loss is 13.66 , inference loss is 6.39, weight deacy loss is 7.27, training accuracy is 0.500000, time 298.215 samples/sec
+epoch 8, total_step 127880, total loss is 13.78 , inference loss is 6.51, weight deacy loss is 7.27, training accuracy is 0.437500, time 246.135 samples/sec
+epoch 8, total_step 127900, total loss is 17.09 , inference loss is 9.82, weight deacy loss is 7.27, training accuracy is 0.468750, time 203.456 samples/sec
+epoch 8, total_step 127920, total loss is 15.07 , inference loss is 7.80, weight deacy loss is 7.27, training accuracy is 0.406250, time 208.866 samples/sec
+epoch 8, total_step 127940, total loss is 16.60 , inference loss is 9.33, weight deacy loss is 7.27, training accuracy is 0.437500, time 220.253 samples/sec
+epoch 8, total_step 127960, total loss is 15.90 , inference loss is 8.64, weight deacy loss is 7.27, training accuracy is 0.312500, time 211.766 samples/sec
+epoch 8, total_step 127980, total loss is 10.17 , inference loss is 2.90, weight deacy loss is 7.27, training accuracy is 0.625000, time 214.529 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.348349000000045
+best_threshold_index 115 0.9879629629629629
+best_threshold_index 114 0.987037037037037
+best_threshold_index 114 0.9872222222222222
+best_threshold_index 115 0.9885185185185185
+best_threshold_index 115 0.9883333333333333
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 114 0.9875925925925926
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9866666666666667
+best_threshold_index 113 0.9864814814814815
+[lfw][128000]XNorm: 22.221234
+[lfw][128000]Accuracy-Flip: 0.98667+-0.00511
+test accuracy is: 0.9866666666666667
+epoch 8, total_step 128000, total loss is 16.12 , inference loss is 8.86, weight deacy loss is 7.27, training accuracy is 0.468750, time 19.523 samples/sec
+epoch 8, total_step 128020, total loss is 16.27 , inference loss is 9.00, weight deacy loss is 7.27, training accuracy is 0.531250, time 213.415 samples/sec
+epoch 8, total_step 128040, total loss is 17.74 , inference loss is 10.48, weight deacy loss is 7.27, training accuracy is 0.406250, time 211.830 samples/sec
+epoch 8, total_step 128060, total loss is 16.16 , inference loss is 8.90, weight deacy loss is 7.26, training accuracy is 0.531250, time 210.618 samples/sec
+epoch 8, total_step 128080, total loss is 15.76 , inference loss is 8.49, weight deacy loss is 7.26, training accuracy is 0.437500, time 205.585 samples/sec
+epoch 8, total_step 128100, total loss is 14.90 , inference loss is 7.64, weight deacy loss is 7.26, training accuracy is 0.343750, time 13.910 samples/sec
+epoch 8, total_step 128120, total loss is 14.33 , inference loss is 7.07, weight deacy loss is 7.26, training accuracy is 0.468750, time 213.877 samples/sec
+epoch 8, total_step 128140, total loss is 20.95 , inference loss is 13.68, weight deacy loss is 7.26, training accuracy is 0.281250, time 300.606 samples/sec
+epoch 8, total_step 128160, total loss is 17.27 , inference loss is 10.01, weight deacy loss is 7.26, training accuracy is 0.468750, time 205.800 samples/sec
+epoch 8, total_step 128180, total loss is 16.21 , inference loss is 8.95, weight deacy loss is 7.26, training accuracy is 0.437500, time 218.253 samples/sec
+epoch 8, total_step 128200, total loss is 15.34 , inference loss is 8.08, weight deacy loss is 7.26, training accuracy is 0.281250, time 203.888 samples/sec
+epoch 8, total_step 128220, total loss is 14.74 , inference loss is 7.48, weight deacy loss is 7.26, training accuracy is 0.437500, time 200.056 samples/sec
+epoch 8, total_step 128240, total loss is 13.80 , inference loss is 6.54, weight deacy loss is 7.26, training accuracy is 0.562500, time 210.374 samples/sec
+epoch 8, total_step 128260, total loss is 13.80 , inference loss is 6.54, weight deacy loss is 7.26, training accuracy is 0.468750, time 215.868 samples/sec
+epoch 8, total_step 128280, total loss is 14.80 , inference loss is 7.54, weight deacy loss is 7.26, training accuracy is 0.531250, time 212.463 samples/sec
+epoch 8, total_step 128300, total loss is 15.95 , inference loss is 8.69, weight deacy loss is 7.26, training accuracy is 0.406250, time 217.197 samples/sec
+epoch 8, total_step 128320, total loss is 12.20 , inference loss is 4.94, weight deacy loss is 7.26, training accuracy is 0.437500, time 209.138 samples/sec
+epoch 8, total_step 128340, total loss is 15.12 , inference loss is 7.87, weight deacy loss is 7.26, training accuracy is 0.312500, time 236.205 samples/sec
+epoch 8, total_step 128360, total loss is 16.73 , inference loss is 9.47, weight deacy loss is 7.26, training accuracy is 0.375000, time 211.457 samples/sec
+epoch 8, total_step 128380, total loss is 11.48 , inference loss is 4.22, weight deacy loss is 7.26, training accuracy is 0.531250, time 205.922 samples/sec
+epoch 8, total_step 128400, total loss is 16.56 , inference loss is 9.30, weight deacy loss is 7.26, training accuracy is 0.312500, time 11.951 samples/sec
+epoch 8, total_step 128420, total loss is 13.19 , inference loss is 5.93, weight deacy loss is 7.26, training accuracy is 0.468750, time 207.428 samples/sec
+epoch 8, total_step 128440, total loss is 22.17 , inference loss is 14.91, weight deacy loss is 7.26, training accuracy is 0.281250, time 208.028 samples/sec
+epoch 8, total_step 128460, total loss is 16.53 , inference loss is 9.27, weight deacy loss is 7.26, training accuracy is 0.437500, time 211.591 samples/sec
+epoch 8, total_step 128480, total loss is 15.31 , inference loss is 8.06, weight deacy loss is 7.25, training accuracy is 0.500000, time 206.599 samples/sec
+epoch 8, total_step 128500, total loss is 14.61 , inference loss is 7.36, weight deacy loss is 7.25, training accuracy is 0.531250, time 211.378 samples/sec
+epoch 8, total_step 128520, total loss is 18.83 , inference loss is 11.57, weight deacy loss is 7.25, training accuracy is 0.281250, time 209.563 samples/sec
+epoch 8, total_step 128540, total loss is 14.09 , inference loss is 6.84, weight deacy loss is 7.25, training accuracy is 0.500000, time 208.259 samples/sec
+epoch 8, total_step 128560, total loss is 18.64 , inference loss is 11.39, weight deacy loss is 7.25, training accuracy is 0.281250, time 207.098 samples/sec
+epoch 8, total_step 128580, total loss is 17.51 , inference loss is 10.26, weight deacy loss is 7.25, training accuracy is 0.343750, time 205.826 samples/sec
+epoch 8, total_step 128600, total loss is 14.77 , inference loss is 7.51, weight deacy loss is 7.25, training accuracy is 0.468750, time 211.571 samples/sec
+epoch 8, total_step 128620, total loss is 15.44 , inference loss is 8.19, weight deacy loss is 7.25, training accuracy is 0.468750, time 213.192 samples/sec
+epoch 8, total_step 128640, total loss is 16.28 , inference loss is 9.03, weight deacy loss is 7.25, training accuracy is 0.406250, time 216.015 samples/sec
+epoch 8, total_step 128660, total loss is 16.55 , inference loss is 9.30, weight deacy loss is 7.25, training accuracy is 0.281250, time 226.203 samples/sec
+epoch 8, total_step 128680, total loss is 13.15 , inference loss is 5.90, weight deacy loss is 7.25, training accuracy is 0.468750, time 200.804 samples/sec
+epoch 8, total_step 128700, total loss is 20.24 , inference loss is 12.99, weight deacy loss is 7.25, training accuracy is 0.281250, time 13.937 samples/sec
+epoch 8, total_step 128720, total loss is 20.39 , inference loss is 13.14, weight deacy loss is 7.25, training accuracy is 0.406250, time 222.228 samples/sec
+epoch 8, total_step 128740, total loss is 16.10 , inference loss is 8.85, weight deacy loss is 7.25, training accuracy is 0.312500, time 213.589 samples/sec
+epoch 8, total_step 128760, total loss is 13.86 , inference loss is 6.61, weight deacy loss is 7.25, training accuracy is 0.437500, time 212.664 samples/sec
+epoch 8, total_step 128780, total loss is 16.74 , inference loss is 9.50, weight deacy loss is 7.25, training accuracy is 0.281250, time 206.227 samples/sec
+epoch 8, total_step 128800, total loss is 16.13 , inference loss is 8.89, weight deacy loss is 7.25, training accuracy is 0.375000, time 206.654 samples/sec
+epoch 8, total_step 128820, total loss is 14.08 , inference loss is 6.83, weight deacy loss is 7.25, training accuracy is 0.343750, time 208.759 samples/sec
+epoch 8, total_step 128840, total loss is 13.41 , inference loss is 6.16, weight deacy loss is 7.25, training accuracy is 0.500000, time 210.132 samples/sec
+epoch 8, total_step 128860, total loss is 13.96 , inference loss is 6.71, weight deacy loss is 7.25, training accuracy is 0.437500, time 218.794 samples/sec
+epoch 8, total_step 128880, total loss is 16.63 , inference loss is 9.39, weight deacy loss is 7.25, training accuracy is 0.375000, time 200.156 samples/sec
+epoch 8, total_step 128900, total loss is 17.43 , inference loss is 10.18, weight deacy loss is 7.25, training accuracy is 0.375000, time 206.872 samples/sec
+epoch 8, total_step 128920, total loss is 16.00 , inference loss is 8.76, weight deacy loss is 7.24, training accuracy is 0.375000, time 212.674 samples/sec
+epoch 8, total_step 128940, total loss is 14.90 , inference loss is 7.66, weight deacy loss is 7.24, training accuracy is 0.437500, time 207.593 samples/sec
+epoch 8, total_step 128960, total loss is 15.44 , inference loss is 8.20, weight deacy loss is 7.24, training accuracy is 0.343750, time 210.131 samples/sec
+epoch 8, total_step 128980, total loss is 13.59 , inference loss is 6.35, weight deacy loss is 7.24, training accuracy is 0.500000, time 206.844 samples/sec
+epoch 8, total_step 129000, total loss is 19.29 , inference loss is 12.05, weight deacy loss is 7.24, training accuracy is 0.062500, time 14.273 samples/sec
+epoch 8, total_step 129020, total loss is 15.20 , inference loss is 7.95, weight deacy loss is 7.24, training accuracy is 0.468750, time 219.891 samples/sec
+epoch 8, total_step 129040, total loss is 14.37 , inference loss is 7.13, weight deacy loss is 7.24, training accuracy is 0.375000, time 215.152 samples/sec
+epoch 8, total_step 129060, total loss is 17.59 , inference loss is 10.35, weight deacy loss is 7.24, training accuracy is 0.375000, time 205.265 samples/sec
+epoch 8, total_step 129080, total loss is 13.46 , inference loss is 6.21, weight deacy loss is 7.24, training accuracy is 0.531250, time 209.653 samples/sec
+epoch 8, total_step 129100, total loss is 18.59 , inference loss is 11.35, weight deacy loss is 7.24, training accuracy is 0.281250, time 207.983 samples/sec
+epoch 8, total_step 129120, total loss is 14.73 , inference loss is 7.49, weight deacy loss is 7.24, training accuracy is 0.312500, time 209.504 samples/sec
+epoch 8, total_step 129140, total loss is 13.04 , inference loss is 5.80, weight deacy loss is 7.24, training accuracy is 0.437500, time 217.677 samples/sec
+epoch 8, total_step 129160, total loss is 14.02 , inference loss is 6.78, weight deacy loss is 7.24, training accuracy is 0.406250, time 202.150 samples/sec
+epoch 8, total_step 129180, total loss is 18.78 , inference loss is 11.54, weight deacy loss is 7.24, training accuracy is 0.281250, time 213.486 samples/sec
+epoch 8, total_step 129200, total loss is 18.40 , inference loss is 11.16, weight deacy loss is 7.24, training accuracy is 0.343750, time 216.148 samples/sec
+epoch 8, total_step 129220, total loss is 15.08 , inference loss is 7.85, weight deacy loss is 7.24, training accuracy is 0.312500, time 209.527 samples/sec
+epoch 8, total_step 129240, total loss is 13.80 , inference loss is 6.56, weight deacy loss is 7.24, training accuracy is 0.312500, time 211.075 samples/sec
+epoch 8, total_step 129260, total loss is 18.51 , inference loss is 11.27, weight deacy loss is 7.24, training accuracy is 0.375000, time 300.461 samples/sec
+epoch 8, total_step 129280, total loss is 15.65 , inference loss is 8.41, weight deacy loss is 7.24, training accuracy is 0.312500, time 222.441 samples/sec
+epoch 8, total_step 129300, total loss is 15.99 , inference loss is 8.75, weight deacy loss is 7.24, training accuracy is 0.312500, time 13.747 samples/sec
+epoch 8, total_step 129320, total loss is 15.82 , inference loss is 8.59, weight deacy loss is 7.24, training accuracy is 0.437500, time 209.194 samples/sec
+epoch 8, total_step 129340, total loss is 13.28 , inference loss is 6.04, weight deacy loss is 7.24, training accuracy is 0.531250, time 223.026 samples/sec
+epoch 8, total_step 129360, total loss is 12.42 , inference loss is 5.18, weight deacy loss is 7.23, training accuracy is 0.500000, time 223.288 samples/sec
+epoch 8, total_step 129380, total loss is 17.86 , inference loss is 10.62, weight deacy loss is 7.23, training accuracy is 0.343750, time 219.274 samples/sec
+epoch 8, total_step 129400, total loss is 13.93 , inference loss is 6.70, weight deacy loss is 7.23, training accuracy is 0.375000, time 205.216 samples/sec
+epoch 8, total_step 129420, total loss is 18.56 , inference loss is 11.33, weight deacy loss is 7.23, training accuracy is 0.375000, time 213.087 samples/sec
+epoch 8, total_step 129440, total loss is 18.57 , inference loss is 11.34, weight deacy loss is 7.23, training accuracy is 0.406250, time 218.741 samples/sec
+epoch 8, total_step 129460, total loss is 19.60 , inference loss is 12.37, weight deacy loss is 7.23, training accuracy is 0.312500, time 210.254 samples/sec
+epoch 8, total_step 129480, total loss is 12.32 , inference loss is 5.09, weight deacy loss is 7.23, training accuracy is 0.500000, time 211.191 samples/sec
+epoch 8, total_step 129500, total loss is 15.97 , inference loss is 8.74, weight deacy loss is 7.23, training accuracy is 0.343750, time 217.624 samples/sec
+epoch 8, total_step 129520, total loss is 17.83 , inference loss is 10.60, weight deacy loss is 7.23, training accuracy is 0.281250, time 300.165 samples/sec
+epoch 8, total_step 129540, total loss is 17.24 , inference loss is 10.01, weight deacy loss is 7.23, training accuracy is 0.312500, time 297.720 samples/sec
+epoch 8, total_step 129560, total loss is 15.52 , inference loss is 8.29, weight deacy loss is 7.23, training accuracy is 0.500000, time 297.670 samples/sec
+epoch 8, total_step 129580, total loss is 17.18 , inference loss is 9.95, weight deacy loss is 7.23, training accuracy is 0.343750, time 215.358 samples/sec
+epoch 8, total_step 129600, total loss is 17.08 , inference loss is 9.85, weight deacy loss is 7.23, training accuracy is 0.406250, time 14.175 samples/sec
+epoch 8, total_step 129620, total loss is 10.92 , inference loss is 3.69, weight deacy loss is 7.23, training accuracy is 0.687500, time 216.194 samples/sec
+epoch 8, total_step 129640, total loss is 16.37 , inference loss is 9.14, weight deacy loss is 7.23, training accuracy is 0.468750, time 213.663 samples/sec
+epoch 8, total_step 129660, total loss is 13.12 , inference loss is 5.89, weight deacy loss is 7.23, training accuracy is 0.468750, time 298.818 samples/sec
+epoch 8, total_step 129680, total loss is 17.44 , inference loss is 10.21, weight deacy loss is 7.23, training accuracy is 0.281250, time 299.307 samples/sec
+epoch 8, total_step 129700, total loss is 13.71 , inference loss is 6.48, weight deacy loss is 7.23, training accuracy is 0.437500, time 208.271 samples/sec
+epoch 8, total_step 129720, total loss is 18.03 , inference loss is 10.80, weight deacy loss is 7.23, training accuracy is 0.406250, time 214.437 samples/sec
+epoch 8, total_step 129740, total loss is 17.61 , inference loss is 10.39, weight deacy loss is 7.23, training accuracy is 0.281250, time 197.046 samples/sec
+epoch 8, total_step 129760, total loss is 15.00 , inference loss is 7.77, weight deacy loss is 7.23, training accuracy is 0.343750, time 211.435 samples/sec
+epoch 8, total_step 129780, total loss is 12.27 , inference loss is 5.04, weight deacy loss is 7.23, training accuracy is 0.500000, time 211.660 samples/sec
+epoch 8, total_step 129800, total loss is 17.80 , inference loss is 10.57, weight deacy loss is 7.23, training accuracy is 0.312500, time 226.954 samples/sec
+epoch 8, total_step 129820, total loss is 15.13 , inference loss is 7.91, weight deacy loss is 7.22, training accuracy is 0.312500, time 213.151 samples/sec
+epoch 8, total_step 129840, total loss is 18.14 , inference loss is 10.92, weight deacy loss is 7.22, training accuracy is 0.312500, time 215.384 samples/sec
+epoch 8, total_step 129860, total loss is 17.31 , inference loss is 10.08, weight deacy loss is 7.22, training accuracy is 0.312500, time 213.703 samples/sec
+epoch 8, total_step 129880, total loss is 14.38 , inference loss is 7.15, weight deacy loss is 7.22, training accuracy is 0.312500, time 211.731 samples/sec
+epoch 8, total_step 129900, total loss is 13.94 , inference loss is 6.71, weight deacy loss is 7.22, training accuracy is 0.406250, time 14.026 samples/sec
+epoch 8, total_step 129920, total loss is 14.94 , inference loss is 7.72, weight deacy loss is 7.22, training accuracy is 0.375000, time 214.200 samples/sec
+epoch 8, total_step 129940, total loss is 13.62 , inference loss is 6.40, weight deacy loss is 7.22, training accuracy is 0.468750, time 212.274 samples/sec
+epoch 8, total_step 129960, total loss is 15.54 , inference loss is 8.32, weight deacy loss is 7.22, training accuracy is 0.500000, time 225.375 samples/sec
+epoch 8, total_step 129980, total loss is 13.53 , inference loss is 6.31, weight deacy loss is 7.22, training accuracy is 0.500000, time 212.259 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.55036099999998
+best_threshold_index 117 0.9887037037037038
+best_threshold_index 117 0.9881481481481481
+best_threshold_index 117 0.9890740740740741
+best_threshold_index 116 0.9888888888888889
+best_threshold_index 117 0.9888888888888889
+best_threshold_index 117 0.9887037037037038
+best_threshold_index 117 0.9887037037037038
+best_threshold_index 116 0.9883333333333333
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 117 0.9879629629629629
+[lfw][130000]XNorm: 21.137521
+[lfw][130000]Accuracy-Flip: 0.98817+-0.00418
+test accuracy is: 0.9881666666666667
+epoch 8, total_step 130000, total loss is 14.96 , inference loss is 7.74, weight deacy loss is 7.22, training accuracy is 0.312500, time 17.076 samples/sec
+epoch 8, total_step 130020, total loss is 15.70 , inference loss is 8.48, weight deacy loss is 7.22, training accuracy is 0.531250, time 296.703 samples/sec
+epoch 8, total_step 130040, total loss is 14.27 , inference loss is 7.05, weight deacy loss is 7.22, training accuracy is 0.406250, time 210.947 samples/sec
+epoch 8, total_step 130060, total loss is 16.11 , inference loss is 8.89, weight deacy loss is 7.22, training accuracy is 0.343750, time 212.420 samples/sec
+epoch 8, total_step 130080, total loss is 15.96 , inference loss is 8.75, weight deacy loss is 7.22, training accuracy is 0.406250, time 210.696 samples/sec
+epoch 8, total_step 130100, total loss is 17.63 , inference loss is 10.41, weight deacy loss is 7.22, training accuracy is 0.281250, time 211.190 samples/sec
+epoch 8, total_step 130120, total loss is 17.71 , inference loss is 10.50, weight deacy loss is 7.22, training accuracy is 0.312500, time 211.808 samples/sec
+epoch 8, total_step 130140, total loss is 15.39 , inference loss is 8.17, weight deacy loss is 7.22, training accuracy is 0.343750, time 211.721 samples/sec
+epoch 8, total_step 130160, total loss is 18.14 , inference loss is 10.93, weight deacy loss is 7.22, training accuracy is 0.281250, time 215.986 samples/sec
+epoch 8, total_step 130180, total loss is 16.64 , inference loss is 9.42, weight deacy loss is 7.22, training accuracy is 0.343750, time 211.518 samples/sec
+epoch 8, total_step 130200, total loss is 13.54 , inference loss is 6.32, weight deacy loss is 7.22, training accuracy is 0.406250, time 14.601 samples/sec
+epoch 8, total_step 130220, total loss is 14.20 , inference loss is 6.99, weight deacy loss is 7.22, training accuracy is 0.500000, time 211.434 samples/sec
+epoch 8, total_step 130240, total loss is 14.60 , inference loss is 7.39, weight deacy loss is 7.21, training accuracy is 0.562500, time 211.118 samples/sec
+epoch 8, total_step 130260, total loss is 16.33 , inference loss is 9.12, weight deacy loss is 7.21, training accuracy is 0.437500, time 216.186 samples/sec
+epoch 8, total_step 130280, total loss is 14.94 , inference loss is 7.73, weight deacy loss is 7.21, training accuracy is 0.437500, time 297.354 samples/sec
+epoch 8, total_step 130300, total loss is 16.29 , inference loss is 9.07, weight deacy loss is 7.21, training accuracy is 0.312500, time 215.007 samples/sec
+epoch 8, total_step 130320, total loss is 16.21 , inference loss is 9.00, weight deacy loss is 7.21, training accuracy is 0.312500, time 217.796 samples/sec
+epoch 8, total_step 130340, total loss is 16.98 , inference loss is 9.77, weight deacy loss is 7.21, training accuracy is 0.437500, time 221.753 samples/sec
+epoch 8, total_step 130360, total loss is 13.72 , inference loss is 6.51, weight deacy loss is 7.21, training accuracy is 0.375000, time 205.618 samples/sec
+epoch 8, total_step 130380, total loss is 13.98 , inference loss is 6.77, weight deacy loss is 7.21, training accuracy is 0.500000, time 220.228 samples/sec
+epoch 8, total_step 130400, total loss is 15.59 , inference loss is 8.38, weight deacy loss is 7.21, training accuracy is 0.468750, time 216.487 samples/sec
+epoch 8, total_step 130420, total loss is 14.35 , inference loss is 7.14, weight deacy loss is 7.21, training accuracy is 0.343750, time 215.359 samples/sec
+epoch 8, total_step 130440, total loss is 20.18 , inference loss is 12.97, weight deacy loss is 7.21, training accuracy is 0.312500, time 212.673 samples/sec
+epoch 8, total_step 130460, total loss is 19.04 , inference loss is 11.83, weight deacy loss is 7.21, training accuracy is 0.312500, time 236.882 samples/sec
+epoch 8, total_step 130480, total loss is 16.24 , inference loss is 9.03, weight deacy loss is 7.21, training accuracy is 0.468750, time 221.060 samples/sec
+epoch 8, total_step 130500, total loss is 16.93 , inference loss is 9.72, weight deacy loss is 7.21, training accuracy is 0.468750, time 13.687 samples/sec
+epoch 8, total_step 130520, total loss is 14.58 , inference loss is 7.37, weight deacy loss is 7.21, training accuracy is 0.468750, time 219.701 samples/sec
+epoch 8, total_step 130540, total loss is 15.01 , inference loss is 7.80, weight deacy loss is 7.21, training accuracy is 0.468750, time 225.271 samples/sec
+epoch 8, total_step 130560, total loss is 13.14 , inference loss is 5.94, weight deacy loss is 7.21, training accuracy is 0.468750, time 197.516 samples/sec
+epoch 8, total_step 130580, total loss is 19.41 , inference loss is 12.21, weight deacy loss is 7.21, training accuracy is 0.281250, time 212.346 samples/sec
+epoch 8, total_step 130600, total loss is 16.03 , inference loss is 8.83, weight deacy loss is 7.21, training accuracy is 0.468750, time 222.028 samples/sec
+epoch 8, total_step 130620, total loss is 20.48 , inference loss is 13.27, weight deacy loss is 7.21, training accuracy is 0.281250, time 243.473 samples/sec
+epoch 8, total_step 130640, total loss is 16.51 , inference loss is 9.31, weight deacy loss is 7.21, training accuracy is 0.281250, time 201.961 samples/sec
+epoch 8, total_step 130660, total loss is 20.32 , inference loss is 13.12, weight deacy loss is 7.20, training accuracy is 0.312500, time 206.936 samples/sec
+epoch 8, total_step 130680, total loss is 14.05 , inference loss is 6.85, weight deacy loss is 7.20, training accuracy is 0.437500, time 217.155 samples/sec
+epoch 8, total_step 130700, total loss is 15.82 , inference loss is 8.61, weight deacy loss is 7.20, training accuracy is 0.343750, time 214.713 samples/sec
+epoch 8, total_step 130720, total loss is 16.55 , inference loss is 9.35, weight deacy loss is 7.20, training accuracy is 0.375000, time 206.331 samples/sec
+epoch 8, total_step 130740, total loss is 20.40 , inference loss is 13.20, weight deacy loss is 7.20, training accuracy is 0.312500, time 211.856 samples/sec
+epoch 8, total_step 130760, total loss is 16.49 , inference loss is 9.29, weight deacy loss is 7.20, training accuracy is 0.343750, time 216.485 samples/sec
+epoch 8, total_step 130780, total loss is 16.13 , inference loss is 8.93, weight deacy loss is 7.20, training accuracy is 0.343750, time 205.566 samples/sec
+epoch 8, total_step 130800, total loss is 14.71 , inference loss is 7.51, weight deacy loss is 7.20, training accuracy is 0.375000, time 14.089 samples/sec
+epoch 8, total_step 130820, total loss is 18.45 , inference loss is 11.25, weight deacy loss is 7.20, training accuracy is 0.312500, time 297.968 samples/sec
+epoch 8, total_step 130840, total loss is 14.43 , inference loss is 7.23, weight deacy loss is 7.20, training accuracy is 0.343750, time 206.516 samples/sec
+epoch 8, total_step 130860, total loss is 17.78 , inference loss is 10.58, weight deacy loss is 7.20, training accuracy is 0.281250, time 216.051 samples/sec
+epoch 8, total_step 130880, total loss is 14.79 , inference loss is 7.59, weight deacy loss is 7.20, training accuracy is 0.406250, time 214.113 samples/sec
+epoch 8, total_step 130900, total loss is 16.65 , inference loss is 9.45, weight deacy loss is 7.20, training accuracy is 0.406250, time 208.777 samples/sec
+epoch 8, total_step 130920, total loss is 18.02 , inference loss is 10.82, weight deacy loss is 7.20, training accuracy is 0.343750, time 211.666 samples/sec
+epoch 8, total_step 130940, total loss is 18.11 , inference loss is 10.91, weight deacy loss is 7.20, training accuracy is 0.281250, time 207.390 samples/sec
+epoch 8, total_step 130960, total loss is 15.08 , inference loss is 7.88, weight deacy loss is 7.20, training accuracy is 0.375000, time 219.227 samples/sec
+epoch 8, total_step 130980, total loss is 17.11 , inference loss is 9.92, weight deacy loss is 7.20, training accuracy is 0.406250, time 197.908 samples/sec
+epoch 8, total_step 131000, total loss is 18.99 , inference loss is 11.79, weight deacy loss is 7.20, training accuracy is 0.312500, time 221.377 samples/sec
+epoch 8, total_step 131020, total loss is 12.92 , inference loss is 5.73, weight deacy loss is 7.20, training accuracy is 0.593750, time 203.925 samples/sec
+epoch 8, total_step 131040, total loss is 18.79 , inference loss is 11.59, weight deacy loss is 7.20, training accuracy is 0.375000, time 217.186 samples/sec
+epoch 8, total_step 131060, total loss is 17.88 , inference loss is 10.68, weight deacy loss is 7.20, training accuracy is 0.500000, time 207.255 samples/sec
+epoch 8, total_step 131080, total loss is 19.48 , inference loss is 12.28, weight deacy loss is 7.20, training accuracy is 0.468750, time 206.770 samples/sec
+epoch 8, total_step 131100, total loss is 18.68 , inference loss is 11.49, weight deacy loss is 7.20, training accuracy is 0.406250, time 15.132 samples/sec
+epoch 8, total_step 131120, total loss is 16.30 , inference loss is 9.11, weight deacy loss is 7.19, training accuracy is 0.375000, time 209.518 samples/sec
+epoch 8, total_step 131140, total loss is 16.29 , inference loss is 9.09, weight deacy loss is 7.19, training accuracy is 0.375000, time 222.773 samples/sec
+epoch 8, total_step 131160, total loss is 17.47 , inference loss is 10.28, weight deacy loss is 7.19, training accuracy is 0.281250, time 203.361 samples/sec
+epoch 8, total_step 131180, total loss is 18.17 , inference loss is 10.98, weight deacy loss is 7.19, training accuracy is 0.375000, time 205.303 samples/sec
+epoch 8, total_step 131200, total loss is 20.32 , inference loss is 13.13, weight deacy loss is 7.19, training accuracy is 0.187500, time 214.241 samples/sec
+epoch 8, total_step 131220, total loss is 18.30 , inference loss is 11.11, weight deacy loss is 7.19, training accuracy is 0.343750, time 215.661 samples/sec
+epoch 8, total_step 131240, total loss is 16.27 , inference loss is 9.08, weight deacy loss is 7.19, training accuracy is 0.343750, time 219.125 samples/sec
+epoch 8, total_step 131260, total loss is 14.80 , inference loss is 7.61, weight deacy loss is 7.19, training accuracy is 0.343750, time 210.748 samples/sec
+epoch 8, total_step 131280, total loss is 16.43 , inference loss is 9.24, weight deacy loss is 7.19, training accuracy is 0.218750, time 203.274 samples/sec
+epoch 8, total_step 131300, total loss is 20.81 , inference loss is 13.62, weight deacy loss is 7.19, training accuracy is 0.218750, time 297.353 samples/sec
+epoch 8, total_step 131320, total loss is 18.67 , inference loss is 11.48, weight deacy loss is 7.19, training accuracy is 0.250000, time 215.555 samples/sec
+epoch 8, total_step 131340, total loss is 17.61 , inference loss is 10.42, weight deacy loss is 7.19, training accuracy is 0.468750, time 217.925 samples/sec
+epoch 8, total_step 131360, total loss is 14.73 , inference loss is 7.54, weight deacy loss is 7.19, training accuracy is 0.437500, time 224.274 samples/sec
+epoch 8, total_step 131380, total loss is 12.90 , inference loss is 5.72, weight deacy loss is 7.19, training accuracy is 0.562500, time 219.038 samples/sec
+epoch 8, total_step 131400, total loss is 16.25 , inference loss is 9.06, weight deacy loss is 7.19, training accuracy is 0.531250, time 14.150 samples/sec
+epoch 8, total_step 131420, total loss is 13.80 , inference loss is 6.61, weight deacy loss is 7.19, training accuracy is 0.375000, time 296.506 samples/sec
+epoch 8, total_step 131440, total loss is 14.35 , inference loss is 7.16, weight deacy loss is 7.19, training accuracy is 0.406250, time 220.284 samples/sec
+epoch 8, total_step 131460, total loss is 16.75 , inference loss is 9.56, weight deacy loss is 7.19, training accuracy is 0.312500, time 204.303 samples/sec
+epoch 8, total_step 131480, total loss is 15.36 , inference loss is 8.17, weight deacy loss is 7.19, training accuracy is 0.375000, time 210.708 samples/sec
+epoch 8, total_step 131500, total loss is 15.82 , inference loss is 8.63, weight deacy loss is 7.19, training accuracy is 0.312500, time 207.970 samples/sec
+epoch 8, total_step 131520, total loss is 15.41 , inference loss is 8.22, weight deacy loss is 7.19, training accuracy is 0.250000, time 210.777 samples/sec
+epoch 8, total_step 131540, total loss is 17.04 , inference loss is 9.85, weight deacy loss is 7.19, training accuracy is 0.437500, time 297.244 samples/sec
+epoch 8, total_step 131560, total loss is 15.71 , inference loss is 8.53, weight deacy loss is 7.19, training accuracy is 0.562500, time 210.860 samples/sec
+epoch 8, total_step 131580, total loss is 11.08 , inference loss is 3.89, weight deacy loss is 7.18, training accuracy is 0.625000, time 218.887 samples/sec
+epoch 8, total_step 131600, total loss is 14.95 , inference loss is 7.77, weight deacy loss is 7.18, training accuracy is 0.437500, time 205.481 samples/sec
+epoch 8, total_step 131620, total loss is 20.76 , inference loss is 13.58, weight deacy loss is 7.18, training accuracy is 0.375000, time 219.404 samples/sec
+epoch 8, total_step 131640, total loss is 16.92 , inference loss is 9.74, weight deacy loss is 7.18, training accuracy is 0.312500, time 201.406 samples/sec
+epoch 8, total_step 131660, total loss is 13.96 , inference loss is 6.78, weight deacy loss is 7.18, training accuracy is 0.437500, time 210.814 samples/sec
+epoch 8, total_step 131680, total loss is 15.34 , inference loss is 8.16, weight deacy loss is 7.18, training accuracy is 0.468750, time 207.250 samples/sec
+epoch 8, total_step 131700, total loss is 15.81 , inference loss is 8.63, weight deacy loss is 7.18, training accuracy is 0.312500, time 12.984 samples/sec
+epoch 8, total_step 131720, total loss is 15.66 , inference loss is 8.48, weight deacy loss is 7.18, training accuracy is 0.281250, time 219.697 samples/sec
+epoch 8, total_step 131740, total loss is 15.35 , inference loss is 8.16, weight deacy loss is 7.18, training accuracy is 0.406250, time 211.873 samples/sec
+epoch 8, total_step 131760, total loss is 16.09 , inference loss is 8.91, weight deacy loss is 7.18, training accuracy is 0.500000, time 216.517 samples/sec
+epoch 8, total_step 131780, total loss is 16.95 , inference loss is 9.77, weight deacy loss is 7.18, training accuracy is 0.343750, time 216.652 samples/sec
+epoch 8, total_step 131800, total loss is 21.00 , inference loss is 13.82, weight deacy loss is 7.18, training accuracy is 0.312500, time 300.018 samples/sec
+epoch 8, total_step 131820, total loss is 14.80 , inference loss is 7.62, weight deacy loss is 7.18, training accuracy is 0.406250, time 212.650 samples/sec
+epoch 8, total_step 131840, total loss is 14.35 , inference loss is 7.17, weight deacy loss is 7.18, training accuracy is 0.312500, time 201.429 samples/sec
+epoch 8, total_step 131860, total loss is 16.10 , inference loss is 8.92, weight deacy loss is 7.18, training accuracy is 0.437500, time 209.538 samples/sec
+epoch 8, total_step 131880, total loss is 17.87 , inference loss is 10.69, weight deacy loss is 7.18, training accuracy is 0.343750, time 210.110 samples/sec
+epoch 8, total_step 131900, total loss is 10.66 , inference loss is 3.48, weight deacy loss is 7.18, training accuracy is 0.625000, time 213.827 samples/sec
+epoch 8, total_step 131920, total loss is 12.43 , inference loss is 5.25, weight deacy loss is 7.18, training accuracy is 0.500000, time 210.899 samples/sec
+epoch 8, total_step 131940, total loss is 19.46 , inference loss is 12.28, weight deacy loss is 7.18, training accuracy is 0.281250, time 213.313 samples/sec
+epoch 8, total_step 131960, total loss is 16.66 , inference loss is 9.48, weight deacy loss is 7.18, training accuracy is 0.437500, time 210.537 samples/sec
+epoch 8, total_step 131980, total loss is 13.66 , inference loss is 6.48, weight deacy loss is 7.18, training accuracy is 0.312500, time 219.238 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.54797099999998
+best_threshold_index 116 0.9879629629629629
+best_threshold_index 113 0.9866666666666667
+best_threshold_index 113 0.9874074074074074
+best_threshold_index 113 0.9874074074074074
+best_threshold_index 113 0.9875925925925926
+best_threshold_index 116 0.9874074074074074
+best_threshold_index 113 0.987037037037037
+best_threshold_index 113 0.9866666666666667
+best_threshold_index 116 0.9861111111111112
+best_threshold_index 113 0.9864814814814815
+[lfw][132000]XNorm: 22.230508
+[lfw][132000]Accuracy-Flip: 0.98633+-0.00493
+test accuracy is: 0.9863333333333333
+epoch 8, total_step 132000, total loss is 15.91 , inference loss is 8.73, weight deacy loss is 7.18, training accuracy is 0.468750, time 17.359 samples/sec
+epoch 8, total_step 132020, total loss is 17.43 , inference loss is 10.26, weight deacy loss is 7.18, training accuracy is 0.312500, time 207.132 samples/sec
+epoch 8, total_step 132040, total loss is 16.74 , inference loss is 9.56, weight deacy loss is 7.17, training accuracy is 0.437500, time 210.690 samples/sec
+epoch 8, total_step 132060, total loss is 13.29 , inference loss is 6.12, weight deacy loss is 7.17, training accuracy is 0.343750, time 218.791 samples/sec
+epoch 8, total_step 132080, total loss is 13.65 , inference loss is 6.47, weight deacy loss is 7.17, training accuracy is 0.406250, time 214.177 samples/sec
+epoch 8, total_step 132100, total loss is 18.55 , inference loss is 11.38, weight deacy loss is 7.17, training accuracy is 0.343750, time 210.475 samples/sec
+epoch 8, total_step 132120, total loss is 15.28 , inference loss is 8.11, weight deacy loss is 7.17, training accuracy is 0.406250, time 213.502 samples/sec
+epoch 8, total_step 132140, total loss is 14.01 , inference loss is 6.84, weight deacy loss is 7.17, training accuracy is 0.437500, time 211.277 samples/sec
+epoch 8, total_step 132160, total loss is 16.86 , inference loss is 9.69, weight deacy loss is 7.17, training accuracy is 0.343750, time 224.093 samples/sec
+epoch 8, total_step 132180, total loss is 15.02 , inference loss is 7.85, weight deacy loss is 7.17, training accuracy is 0.437500, time 213.761 samples/sec
+epoch 8, total_step 132200, total loss is 14.16 , inference loss is 6.99, weight deacy loss is 7.17, training accuracy is 0.437500, time 298.948 samples/sec
+epoch 8, total_step 132220, total loss is 15.05 , inference loss is 7.88, weight deacy loss is 7.17, training accuracy is 0.375000, time 216.522 samples/sec
+epoch 8, total_step 132240, total loss is 15.45 , inference loss is 8.28, weight deacy loss is 7.17, training accuracy is 0.406250, time 209.407 samples/sec
+epoch 8, total_step 132260, total loss is 20.23 , inference loss is 13.06, weight deacy loss is 7.17, training accuracy is 0.125000, time 210.814 samples/sec
+epoch 8, total_step 132280, total loss is 18.20 , inference loss is 11.03, weight deacy loss is 7.17, training accuracy is 0.375000, time 218.602 samples/sec
+epoch 8, total_step 132300, total loss is 14.15 , inference loss is 6.98, weight deacy loss is 7.17, training accuracy is 0.312500, time 13.780 samples/sec
+epoch 8, total_step 132320, total loss is 13.54 , inference loss is 6.38, weight deacy loss is 7.17, training accuracy is 0.312500, time 215.958 samples/sec
+epoch 8, total_step 132340, total loss is 14.86 , inference loss is 7.69, weight deacy loss is 7.17, training accuracy is 0.500000, time 213.840 samples/sec
+epoch 8, total_step 132360, total loss is 12.82 , inference loss is 5.65, weight deacy loss is 7.17, training accuracy is 0.531250, time 211.615 samples/sec
+epoch 8, total_step 132380, total loss is 15.03 , inference loss is 7.86, weight deacy loss is 7.17, training accuracy is 0.406250, time 203.876 samples/sec
+epoch 8, total_step 132400, total loss is 13.33 , inference loss is 6.16, weight deacy loss is 7.17, training accuracy is 0.500000, time 212.035 samples/sec
+epoch 8, total_step 132420, total loss is 20.34 , inference loss is 13.18, weight deacy loss is 7.17, training accuracy is 0.437500, time 215.235 samples/sec
+epoch 8, total_step 132440, total loss is 17.59 , inference loss is 10.42, weight deacy loss is 7.17, training accuracy is 0.312500, time 215.433 samples/sec
+epoch 8, total_step 132460, total loss is 18.39 , inference loss is 11.22, weight deacy loss is 7.17, training accuracy is 0.343750, time 300.473 samples/sec
+epoch 8, total_step 132480, total loss is 18.06 , inference loss is 10.89, weight deacy loss is 7.17, training accuracy is 0.281250, time 208.963 samples/sec
+epoch 8, total_step 132500, total loss is 16.73 , inference loss is 9.57, weight deacy loss is 7.16, training accuracy is 0.375000, time 211.896 samples/sec
+epoch 8, total_step 132520, total loss is 15.30 , inference loss is 8.13, weight deacy loss is 7.16, training accuracy is 0.406250, time 207.577 samples/sec
+epoch 8, total_step 132540, total loss is 15.02 , inference loss is 7.85, weight deacy loss is 7.16, training accuracy is 0.468750, time 219.961 samples/sec
+epoch 8, total_step 132560, total loss is 22.64 , inference loss is 15.48, weight deacy loss is 7.16, training accuracy is 0.156250, time 211.084 samples/sec
+epoch 8, total_step 132580, total loss is 12.04 , inference loss is 4.87, weight deacy loss is 7.16, training accuracy is 0.562500, time 207.997 samples/sec
+epoch 8, total_step 132600, total loss is 17.72 , inference loss is 10.56, weight deacy loss is 7.16, training accuracy is 0.312500, time 13.394 samples/sec
+epoch 8, total_step 132620, total loss is 14.91 , inference loss is 7.75, weight deacy loss is 7.16, training accuracy is 0.343750, time 215.753 samples/sec
+epoch 8, total_step 132640, total loss is 14.89 , inference loss is 7.73, weight deacy loss is 7.16, training accuracy is 0.375000, time 210.393 samples/sec
+epoch 8, total_step 132660, total loss is 15.46 , inference loss is 8.30, weight deacy loss is 7.16, training accuracy is 0.468750, time 203.676 samples/sec
+epoch 8, total_step 132680, total loss is 14.94 , inference loss is 7.78, weight deacy loss is 7.16, training accuracy is 0.375000, time 209.416 samples/sec
+epoch 8, total_step 132700, total loss is 13.13 , inference loss is 5.97, weight deacy loss is 7.16, training accuracy is 0.468750, time 208.863 samples/sec
+epoch 8, total_step 132720, total loss is 13.43 , inference loss is 6.27, weight deacy loss is 7.16, training accuracy is 0.531250, time 216.451 samples/sec
+epoch 8, total_step 132740, total loss is 15.58 , inference loss is 8.42, weight deacy loss is 7.16, training accuracy is 0.375000, time 225.669 samples/sec
+epoch 8, total_step 132760, total loss is 15.58 , inference loss is 8.42, weight deacy loss is 7.16, training accuracy is 0.375000, time 206.012 samples/sec
+epoch 8, total_step 132780, total loss is 16.75 , inference loss is 9.59, weight deacy loss is 7.16, training accuracy is 0.406250, time 196.557 samples/sec
+epoch 8, total_step 132800, total loss is 16.81 , inference loss is 9.65, weight deacy loss is 7.16, training accuracy is 0.375000, time 224.623 samples/sec
+epoch 8, total_step 132820, total loss is 15.33 , inference loss is 8.17, weight deacy loss is 7.16, training accuracy is 0.343750, time 215.814 samples/sec
+epoch 8, total_step 132840, total loss is 14.00 , inference loss is 6.84, weight deacy loss is 7.16, training accuracy is 0.406250, time 214.644 samples/sec
+epoch 8, total_step 132860, total loss is 21.93 , inference loss is 14.77, weight deacy loss is 7.16, training accuracy is 0.250000, time 212.649 samples/sec
+epoch 8, total_step 132880, total loss is 11.66 , inference loss is 4.50, weight deacy loss is 7.16, training accuracy is 0.593750, time 209.704 samples/sec
+epoch 8, total_step 132900, total loss is 11.28 , inference loss is 4.12, weight deacy loss is 7.16, training accuracy is 0.500000, time 14.356 samples/sec
+epoch 8, total_step 132920, total loss is 15.00 , inference loss is 7.84, weight deacy loss is 7.16, training accuracy is 0.375000, time 213.810 samples/sec
+epoch 8, total_step 132940, total loss is 16.58 , inference loss is 9.42, weight deacy loss is 7.15, training accuracy is 0.468750, time 219.287 samples/sec
+epoch 8, total_step 132960, total loss is 22.28 , inference loss is 15.13, weight deacy loss is 7.15, training accuracy is 0.375000, time 210.007 samples/sec
+epoch 8, total_step 132980, total loss is 18.33 , inference loss is 11.18, weight deacy loss is 7.15, training accuracy is 0.250000, time 212.421 samples/sec
+epoch 8, total_step 133000, total loss is 14.21 , inference loss is 7.06, weight deacy loss is 7.15, training accuracy is 0.375000, time 298.423 samples/sec
+epoch 8, total_step 133020, total loss is 15.70 , inference loss is 8.55, weight deacy loss is 7.15, training accuracy is 0.531250, time 198.849 samples/sec
+epoch 8, total_step 133040, total loss is 13.43 , inference loss is 6.28, weight deacy loss is 7.15, training accuracy is 0.562500, time 218.281 samples/sec
+epoch 8, total_step 133060, total loss is 16.02 , inference loss is 8.87, weight deacy loss is 7.15, training accuracy is 0.468750, time 220.110 samples/sec
+epoch 8, total_step 133080, total loss is 15.63 , inference loss is 8.48, weight deacy loss is 7.15, training accuracy is 0.406250, time 222.151 samples/sec
+epoch 8, total_step 133100, total loss is 11.74 , inference loss is 4.59, weight deacy loss is 7.15, training accuracy is 0.500000, time 216.324 samples/sec
+epoch 8, total_step 133120, total loss is 18.18 , inference loss is 11.03, weight deacy loss is 7.15, training accuracy is 0.281250, time 216.680 samples/sec
+epoch 8, total_step 133140, total loss is 13.15 , inference loss is 6.00, weight deacy loss is 7.15, training accuracy is 0.531250, time 218.797 samples/sec
+epoch 8, total_step 133160, total loss is 13.93 , inference loss is 6.78, weight deacy loss is 7.15, training accuracy is 0.437500, time 215.587 samples/sec
+epoch 8, total_step 133180, total loss is 17.56 , inference loss is 10.41, weight deacy loss is 7.15, training accuracy is 0.343750, time 214.121 samples/sec
+epoch 8, total_step 133200, total loss is 14.34 , inference loss is 7.19, weight deacy loss is 7.15, training accuracy is 0.468750, time 14.448 samples/sec
+epoch 8, total_step 133220, total loss is 14.48 , inference loss is 7.33, weight deacy loss is 7.15, training accuracy is 0.468750, time 300.562 samples/sec
+epoch 8, total_step 133240, total loss is 15.74 , inference loss is 8.59, weight deacy loss is 7.15, training accuracy is 0.468750, time 300.689 samples/sec
+epoch 8, total_step 133260, total loss is 17.09 , inference loss is 9.94, weight deacy loss is 7.15, training accuracy is 0.437500, time 212.847 samples/sec
+epoch 8, total_step 133280, total loss is 17.72 , inference loss is 10.57, weight deacy loss is 7.15, training accuracy is 0.312500, time 206.862 samples/sec
+epoch 8, total_step 133300, total loss is 17.01 , inference loss is 9.86, weight deacy loss is 7.15, training accuracy is 0.343750, time 213.533 samples/sec
+epoch 8, total_step 133320, total loss is 17.07 , inference loss is 9.93, weight deacy loss is 7.15, training accuracy is 0.437500, time 209.939 samples/sec
+epoch 8, total_step 133340, total loss is 12.52 , inference loss is 5.37, weight deacy loss is 7.15, training accuracy is 0.343750, time 204.501 samples/sec
+epoch 8, total_step 133360, total loss is 16.46 , inference loss is 9.31, weight deacy loss is 7.15, training accuracy is 0.312500, time 218.491 samples/sec
+epoch 8, total_step 133380, total loss is 15.87 , inference loss is 8.72, weight deacy loss is 7.14, training accuracy is 0.406250, time 204.640 samples/sec
+epoch 8, total_step 133400, total loss is 16.51 , inference loss is 9.36, weight deacy loss is 7.14, training accuracy is 0.406250, time 213.693 samples/sec
+epoch 8, total_step 133420, total loss is 17.57 , inference loss is 10.42, weight deacy loss is 7.14, training accuracy is 0.437500, time 217.571 samples/sec
+epoch 8, total_step 133440, total loss is 15.24 , inference loss is 8.10, weight deacy loss is 7.14, training accuracy is 0.531250, time 215.264 samples/sec
+epoch 8, total_step 133460, total loss is 18.13 , inference loss is 10.99, weight deacy loss is 7.14, training accuracy is 0.281250, time 300.757 samples/sec
+epoch 8, total_step 133480, total loss is 15.97 , inference loss is 8.83, weight deacy loss is 7.14, training accuracy is 0.281250, time 207.825 samples/sec
+epoch 8, total_step 133500, total loss is 16.71 , inference loss is 9.57, weight deacy loss is 7.14, training accuracy is 0.375000, time 13.302 samples/sec
+epoch 8, total_step 133520, total loss is 17.38 , inference loss is 10.24, weight deacy loss is 7.14, training accuracy is 0.281250, time 217.988 samples/sec
+epoch 8, total_step 133540, total loss is 13.33 , inference loss is 6.19, weight deacy loss is 7.14, training accuracy is 0.375000, time 220.686 samples/sec
+epoch 8, total_step 133560, total loss is 18.90 , inference loss is 11.75, weight deacy loss is 7.14, training accuracy is 0.281250, time 212.589 samples/sec
+epoch 8, total_step 133580, total loss is 15.21 , inference loss is 8.07, weight deacy loss is 7.14, training accuracy is 0.531250, time 211.740 samples/sec
+epoch 8, total_step 133600, total loss is 12.49 , inference loss is 5.35, weight deacy loss is 7.14, training accuracy is 0.375000, time 228.459 samples/sec
+epoch 8, total_step 133620, total loss is 14.55 , inference loss is 7.41, weight deacy loss is 7.14, training accuracy is 0.375000, time 213.389 samples/sec
+epoch 8, total_step 133640, total loss is 15.66 , inference loss is 8.52, weight deacy loss is 7.14, training accuracy is 0.218750, time 209.993 samples/sec
+epoch 8, total_step 133660, total loss is 20.64 , inference loss is 13.50, weight deacy loss is 7.14, training accuracy is 0.281250, time 213.031 samples/sec
+epoch 8, total_step 133680, total loss is 14.86 , inference loss is 7.72, weight deacy loss is 7.14, training accuracy is 0.500000, time 212.843 samples/sec
+epoch 8, total_step 133700, total loss is 17.34 , inference loss is 10.20, weight deacy loss is 7.14, training accuracy is 0.375000, time 219.549 samples/sec
+epoch 8, total_step 133720, total loss is 14.40 , inference loss is 7.27, weight deacy loss is 7.14, training accuracy is 0.343750, time 298.614 samples/sec
+epoch 8, total_step 133740, total loss is 16.17 , inference loss is 9.03, weight deacy loss is 7.14, training accuracy is 0.343750, time 204.940 samples/sec
+epoch 8, total_step 133760, total loss is 16.46 , inference loss is 9.32, weight deacy loss is 7.14, training accuracy is 0.281250, time 213.212 samples/sec
+epoch 8, total_step 133780, total loss is 15.60 , inference loss is 8.46, weight deacy loss is 7.14, training accuracy is 0.437500, time 214.807 samples/sec
+epoch 8, total_step 133800, total loss is 18.59 , inference loss is 11.46, weight deacy loss is 7.14, training accuracy is 0.375000, time 13.692 samples/sec
+epoch 8, total_step 133820, total loss is 19.34 , inference loss is 12.20, weight deacy loss is 7.14, training accuracy is 0.250000, time 224.687 samples/sec
+epoch 8, total_step 133840, total loss is 15.42 , inference loss is 8.28, weight deacy loss is 7.14, training accuracy is 0.500000, time 206.442 samples/sec
+epoch 8, total_step 133860, total loss is 17.74 , inference loss is 10.60, weight deacy loss is 7.14, training accuracy is 0.312500, time 227.016 samples/sec
+epoch 8, total_step 133880, total loss is 16.22 , inference loss is 9.09, weight deacy loss is 7.13, training accuracy is 0.281250, time 212.766 samples/sec
+epoch 8, total_step 133900, total loss is 16.09 , inference loss is 8.96, weight deacy loss is 7.13, training accuracy is 0.343750, time 217.514 samples/sec
+epoch 8, total_step 133920, total loss is 17.59 , inference loss is 10.46, weight deacy loss is 7.13, training accuracy is 0.281250, time 211.531 samples/sec
+epoch 8, total_step 133940, total loss is 15.94 , inference loss is 8.80, weight deacy loss is 7.13, training accuracy is 0.437500, time 212.481 samples/sec
+epoch 8, total_step 133960, total loss is 19.69 , inference loss is 12.56, weight deacy loss is 7.13, training accuracy is 0.343750, time 216.530 samples/sec
+epoch 8, total_step 133980, total loss is 16.76 , inference loss is 9.63, weight deacy loss is 7.13, training accuracy is 0.343750, time 208.575 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.568871000000026
+best_threshold_index 109 0.985
+best_threshold_index 109 0.9846296296296296
+best_threshold_index 108 0.9842592592592593
+best_threshold_index 107 0.9855555555555555
+best_threshold_index 109 0.9859259259259259
+best_threshold_index 109 0.9842592592592593
+best_threshold_index 109 0.985
+best_threshold_index 109 0.9842592592592593
+best_threshold_index 109 0.9835185185185186
+best_threshold_index 109 0.9844444444444445
+[lfw][134000]XNorm: 19.811025
+[lfw][134000]Accuracy-Flip: 0.98367+-0.00666
+test accuracy is: 0.9836666666666666
+epoch 8, total_step 134000, total loss is 13.59 , inference loss is 6.46, weight deacy loss is 7.13, training accuracy is 0.343750, time 19.375 samples/sec
+epoch 8, total_step 134020, total loss is 12.60 , inference loss is 5.47, weight deacy loss is 7.13, training accuracy is 0.406250, time 207.870 samples/sec
+epoch 8, total_step 134040, total loss is 16.17 , inference loss is 9.04, weight deacy loss is 7.13, training accuracy is 0.312500, time 211.149 samples/sec
+epoch 8, total_step 134060, total loss is 17.27 , inference loss is 10.13, weight deacy loss is 7.13, training accuracy is 0.312500, time 213.956 samples/sec
+epoch 8, total_step 134080, total loss is 16.60 , inference loss is 9.47, weight deacy loss is 7.13, training accuracy is 0.406250, time 216.693 samples/sec
+epoch 8, total_step 134100, total loss is 13.60 , inference loss is 6.47, weight deacy loss is 7.13, training accuracy is 0.500000, time 11.955 samples/sec
+epoch 8, total_step 134120, total loss is 14.97 , inference loss is 7.84, weight deacy loss is 7.13, training accuracy is 0.281250, time 206.932 samples/sec
+epoch 8, total_step 134140, total loss is 18.94 , inference loss is 11.81, weight deacy loss is 7.13, training accuracy is 0.312500, time 230.804 samples/sec
+epoch 8, total_step 134160, total loss is 19.18 , inference loss is 12.05, weight deacy loss is 7.13, training accuracy is 0.406250, time 224.605 samples/sec
+epoch 8, total_step 134180, total loss is 20.68 , inference loss is 13.56, weight deacy loss is 7.13, training accuracy is 0.218750, time 224.083 samples/sec
+epoch 8, total_step 134200, total loss is 19.22 , inference loss is 12.09, weight deacy loss is 7.13, training accuracy is 0.187500, time 226.270 samples/sec
+epoch 8, total_step 134220, total loss is 17.23 , inference loss is 10.10, weight deacy loss is 7.13, training accuracy is 0.375000, time 242.133 samples/sec
+epoch 8, total_step 134240, total loss is 19.65 , inference loss is 12.52, weight deacy loss is 7.13, training accuracy is 0.281250, time 213.541 samples/sec
+epoch 8, total_step 134260, total loss is 14.32 , inference loss is 7.19, weight deacy loss is 7.13, training accuracy is 0.406250, time 203.584 samples/sec
+epoch 8, total_step 134280, total loss is 15.46 , inference loss is 8.33, weight deacy loss is 7.13, training accuracy is 0.500000, time 210.655 samples/sec
+epoch 8, total_step 134300, total loss is 16.52 , inference loss is 9.39, weight deacy loss is 7.13, training accuracy is 0.250000, time 214.942 samples/sec
+epoch 8, total_step 134320, total loss is 18.08 , inference loss is 10.95, weight deacy loss is 7.13, training accuracy is 0.437500, time 209.714 samples/sec
+epoch 8, total_step 134340, total loss is 18.55 , inference loss is 11.43, weight deacy loss is 7.13, training accuracy is 0.281250, time 211.124 samples/sec
+epoch 8, total_step 134360, total loss is 14.53 , inference loss is 7.41, weight deacy loss is 7.12, training accuracy is 0.437500, time 213.076 samples/sec
+epoch 8, total_step 134380, total loss is 14.64 , inference loss is 7.52, weight deacy loss is 7.12, training accuracy is 0.531250, time 301.109 samples/sec
+epoch 8, total_step 134400, total loss is 13.85 , inference loss is 6.72, weight deacy loss is 7.12, training accuracy is 0.250000, time 13.062 samples/sec
+epoch 8, total_step 134420, total loss is 15.18 , inference loss is 8.06, weight deacy loss is 7.12, training accuracy is 0.343750, time 239.904 samples/sec
+epoch 8, total_step 134440, total loss is 21.63 , inference loss is 14.50, weight deacy loss is 7.12, training accuracy is 0.312500, time 227.356 samples/sec
+epoch 8, total_step 134460, total loss is 16.66 , inference loss is 9.54, weight deacy loss is 7.12, training accuracy is 0.406250, time 221.622 samples/sec
+epoch 8, total_step 134480, total loss is 15.03 , inference loss is 7.91, weight deacy loss is 7.12, training accuracy is 0.375000, time 207.992 samples/sec
+epoch 8, total_step 134500, total loss is 16.53 , inference loss is 9.41, weight deacy loss is 7.12, training accuracy is 0.406250, time 207.513 samples/sec
+epoch 8, total_step 134520, total loss is 16.59 , inference loss is 9.47, weight deacy loss is 7.12, training accuracy is 0.218750, time 230.428 samples/sec
+epoch 8, total_step 134540, total loss is 16.86 , inference loss is 9.74, weight deacy loss is 7.12, training accuracy is 0.500000, time 211.221 samples/sec
+epoch 8, total_step 134560, total loss is 17.36 , inference loss is 10.24, weight deacy loss is 7.12, training accuracy is 0.343750, time 210.925 samples/sec
+epoch 8, total_step 134580, total loss is 15.45 , inference loss is 8.33, weight deacy loss is 7.12, training accuracy is 0.312500, time 208.026 samples/sec
+epoch 8, total_step 134600, total loss is 17.32 , inference loss is 10.20, weight deacy loss is 7.12, training accuracy is 0.375000, time 202.092 samples/sec
+epoch 8, total_step 134620, total loss is 20.48 , inference loss is 13.36, weight deacy loss is 7.12, training accuracy is 0.187500, time 214.707 samples/sec
+epoch 8, total_step 134640, total loss is 17.79 , inference loss is 10.67, weight deacy loss is 7.12, training accuracy is 0.218750, time 297.186 samples/sec
+epoch 8, total_step 134660, total loss is 14.57 , inference loss is 7.46, weight deacy loss is 7.12, training accuracy is 0.343750, time 218.786 samples/sec
+epoch 8, total_step 134680, total loss is 16.97 , inference loss is 9.86, weight deacy loss is 7.12, training accuracy is 0.250000, time 203.955 samples/sec
+epoch 8, total_step 134700, total loss is 18.60 , inference loss is 11.49, weight deacy loss is 7.12, training accuracy is 0.312500, time 11.753 samples/sec
+epoch 8, total_step 134720, total loss is 17.37 , inference loss is 10.26, weight deacy loss is 7.12, training accuracy is 0.250000, time 207.651 samples/sec
+epoch 8, total_step 134740, total loss is 15.27 , inference loss is 8.15, weight deacy loss is 7.12, training accuracy is 0.406250, time 218.497 samples/sec
+epoch 8, total_step 134760, total loss is 21.98 , inference loss is 14.87, weight deacy loss is 7.12, training accuracy is 0.250000, time 210.665 samples/sec
+epoch 8, total_step 134780, total loss is 17.55 , inference loss is 10.43, weight deacy loss is 7.12, training accuracy is 0.281250, time 214.536 samples/sec
+epoch 8, total_step 134800, total loss is 14.50 , inference loss is 7.39, weight deacy loss is 7.12, training accuracy is 0.406250, time 210.072 samples/sec
+epoch 8, total_step 134820, total loss is 18.80 , inference loss is 11.68, weight deacy loss is 7.11, training accuracy is 0.250000, time 208.660 samples/sec
+epoch 8, total_step 134840, total loss is 14.48 , inference loss is 7.36, weight deacy loss is 7.11, training accuracy is 0.281250, time 214.219 samples/sec
+epoch 8, total_step 134860, total loss is 11.49 , inference loss is 4.37, weight deacy loss is 7.11, training accuracy is 0.593750, time 234.324 samples/sec
+epoch 8, total_step 134880, total loss is 17.14 , inference loss is 10.03, weight deacy loss is 7.11, training accuracy is 0.343750, time 214.439 samples/sec
+epoch 8, total_step 134900, total loss is 15.91 , inference loss is 8.80, weight deacy loss is 7.11, training accuracy is 0.312500, time 183.664 samples/sec
+epoch 8, total_step 134920, total loss is 18.63 , inference loss is 11.52, weight deacy loss is 7.11, training accuracy is 0.281250, time 229.607 samples/sec
+epoch 8, total_step 134940, total loss is 15.70 , inference loss is 8.59, weight deacy loss is 7.11, training accuracy is 0.250000, time 218.746 samples/sec
+epoch 8, total_step 134960, total loss is 11.39 , inference loss is 4.28, weight deacy loss is 7.11, training accuracy is 0.500000, time 216.805 samples/sec
+epoch 8, total_step 134980, total loss is 16.09 , inference loss is 8.98, weight deacy loss is 7.11, training accuracy is 0.375000, time 224.506 samples/sec
+epoch 8, total_step 135000, total loss is 16.85 , inference loss is 9.73, weight deacy loss is 7.11, training accuracy is 0.218750, time 14.336 samples/sec
+epoch 8, total_step 135020, total loss is 16.75 , inference loss is 9.64, weight deacy loss is 7.11, training accuracy is 0.343750, time 299.328 samples/sec
+epoch 8, total_step 135040, total loss is 18.21 , inference loss is 11.10, weight deacy loss is 7.11, training accuracy is 0.343750, time 297.730 samples/sec
+epoch 8, total_step 135060, total loss is 17.04 , inference loss is 9.93, weight deacy loss is 7.11, training accuracy is 0.218750, time 296.414 samples/sec
+epoch 8, total_step 135080, total loss is 17.90 , inference loss is 10.79, weight deacy loss is 7.11, training accuracy is 0.187500, time 299.379 samples/sec
+epoch 8, total_step 135100, total loss is 14.35 , inference loss is 7.24, weight deacy loss is 7.11, training accuracy is 0.281250, time 217.537 samples/sec
+epoch 8, total_step 135120, total loss is 16.11 , inference loss is 9.00, weight deacy loss is 7.11, training accuracy is 0.468750, time 217.173 samples/sec
+epoch 8, total_step 135140, total loss is 16.60 , inference loss is 9.49, weight deacy loss is 7.11, training accuracy is 0.281250, time 207.115 samples/sec
+epoch 8, total_step 135160, total loss is 15.74 , inference loss is 8.63, weight deacy loss is 7.11, training accuracy is 0.281250, time 221.204 samples/sec
+epoch 8, total_step 135180, total loss is 13.58 , inference loss is 6.48, weight deacy loss is 7.11, training accuracy is 0.406250, time 203.716 samples/sec
+epoch 8, total_step 135200, total loss is 15.95 , inference loss is 8.84, weight deacy loss is 7.11, training accuracy is 0.406250, time 209.944 samples/sec
+epoch 8, total_step 135220, total loss is 15.38 , inference loss is 8.28, weight deacy loss is 7.11, training accuracy is 0.343750, time 216.903 samples/sec
+epoch 8, total_step 135240, total loss is 19.49 , inference loss is 12.39, weight deacy loss is 7.11, training accuracy is 0.187500, time 225.341 samples/sec
+epoch 8, total_step 135260, total loss is 14.13 , inference loss is 7.02, weight deacy loss is 7.10, training accuracy is 0.593750, time 195.231 samples/sec
+epoch 8, total_step 135280, total loss is 14.97 , inference loss is 7.87, weight deacy loss is 7.10, training accuracy is 0.250000, time 209.099 samples/sec
+epoch 8, total_step 135300, total loss is 15.47 , inference loss is 8.36, weight deacy loss is 7.10, training accuracy is 0.406250, time 6.406 samples/sec
+epoch 8, total_step 135320, total loss is 20.14 , inference loss is 13.03, weight deacy loss is 7.10, training accuracy is 0.187500, time 225.313 samples/sec
+epoch 8, total_step 135340, total loss is 20.24 , inference loss is 13.14, weight deacy loss is 7.10, training accuracy is 0.250000, time 219.647 samples/sec
+epoch 8, total_step 135360, total loss is 16.15 , inference loss is 9.05, weight deacy loss is 7.10, training accuracy is 0.406250, time 209.292 samples/sec
+epoch 8, total_step 135380, total loss is 14.27 , inference loss is 7.17, weight deacy loss is 7.10, training accuracy is 0.468750, time 209.745 samples/sec
+epoch 8, total_step 135400, total loss is 18.38 , inference loss is 11.28, weight deacy loss is 7.10, training accuracy is 0.250000, time 220.097 samples/sec
+epoch 8, total_step 135420, total loss is 18.48 , inference loss is 11.38, weight deacy loss is 7.10, training accuracy is 0.125000, time 226.628 samples/sec
+epoch 8, total_step 135440, total loss is 18.01 , inference loss is 10.91, weight deacy loss is 7.10, training accuracy is 0.156250, time 203.018 samples/sec
+epoch 8, total_step 135460, total loss is 20.30 , inference loss is 13.20, weight deacy loss is 7.10, training accuracy is 0.375000, time 206.636 samples/sec
+epoch 8, total_step 135480, total loss is 15.43 , inference loss is 8.33, weight deacy loss is 7.10, training accuracy is 0.187500, time 213.685 samples/sec
+epoch 8, total_step 135500, total loss is 15.65 , inference loss is 8.55, weight deacy loss is 7.10, training accuracy is 0.312500, time 216.916 samples/sec
+epoch 8, total_step 135520, total loss is 14.25 , inference loss is 7.15, weight deacy loss is 7.10, training accuracy is 0.281250, time 228.550 samples/sec
+epoch 8, total_step 135540, total loss is 16.89 , inference loss is 9.79, weight deacy loss is 7.10, training accuracy is 0.281250, time 215.887 samples/sec
+epoch 8, total_step 135560, total loss is 13.81 , inference loss is 6.72, weight deacy loss is 7.10, training accuracy is 0.406250, time 219.093 samples/sec
+epoch 8, total_step 135580, total loss is 17.14 , inference loss is 10.04, weight deacy loss is 7.10, training accuracy is 0.281250, time 209.943 samples/sec
+epoch 8, total_step 135600, total loss is 16.28 , inference loss is 9.18, weight deacy loss is 7.10, training accuracy is 0.281250, time 9.805 samples/sec
+epoch 8, total_step 135620, total loss is 20.09 , inference loss is 13.00, weight deacy loss is 7.10, training accuracy is 0.281250, time 220.406 samples/sec
+epoch 8, total_step 135640, total loss is 15.53 , inference loss is 8.44, weight deacy loss is 7.10, training accuracy is 0.375000, time 210.469 samples/sec
+epoch 8, total_step 135660, total loss is 17.93 , inference loss is 10.83, weight deacy loss is 7.10, training accuracy is 0.250000, time 208.368 samples/sec
+epoch 8, total_step 135680, total loss is 14.73 , inference loss is 7.63, weight deacy loss is 7.10, training accuracy is 0.375000, time 218.696 samples/sec
+epoch 8, total_step 135700, total loss is 16.61 , inference loss is 9.51, weight deacy loss is 7.10, training accuracy is 0.375000, time 211.744 samples/sec
+epoch 8, total_step 135720, total loss is 17.98 , inference loss is 10.89, weight deacy loss is 7.10, training accuracy is 0.250000, time 237.808 samples/sec
+epoch 8, total_step 135740, total loss is 14.07 , inference loss is 6.97, weight deacy loss is 7.09, training accuracy is 0.468750, time 197.448 samples/sec
+epoch 8, total_step 135760, total loss is 16.02 , inference loss is 8.93, weight deacy loss is 7.09, training accuracy is 0.406250, time 207.186 samples/sec
+epoch 8, total_step 135780, total loss is 15.73 , inference loss is 8.63, weight deacy loss is 7.09, training accuracy is 0.218750, time 219.006 samples/sec
+epoch 8, total_step 135800, total loss is 18.91 , inference loss is 11.82, weight deacy loss is 7.09, training accuracy is 0.156250, time 218.455 samples/sec
+epoch 8, total_step 135820, total loss is 13.85 , inference loss is 6.76, weight deacy loss is 7.09, training accuracy is 0.406250, time 209.249 samples/sec
+epoch 8, total_step 135840, total loss is 16.15 , inference loss is 9.06, weight deacy loss is 7.09, training accuracy is 0.312500, time 298.611 samples/sec
+epoch 8, total_step 135860, total loss is 15.52 , inference loss is 8.43, weight deacy loss is 7.09, training accuracy is 0.312500, time 299.635 samples/sec
+epoch 8, total_step 135880, total loss is 14.02 , inference loss is 6.92, weight deacy loss is 7.09, training accuracy is 0.343750, time 299.417 samples/sec
+epoch 8, total_step 135900, total loss is 14.51 , inference loss is 7.42, weight deacy loss is 7.09, training accuracy is 0.406250, time 3.223 samples/sec
+epoch 8, total_step 135920, total loss is 18.16 , inference loss is 11.07, weight deacy loss is 7.09, training accuracy is 0.406250, time 203.267 samples/sec
+epoch 8, total_step 135940, total loss is 17.64 , inference loss is 10.55, weight deacy loss is 7.09, training accuracy is 0.281250, time 298.731 samples/sec
+epoch 8, total_step 135960, total loss is 19.86 , inference loss is 12.77, weight deacy loss is 7.09, training accuracy is 0.250000, time 299.484 samples/sec
+epoch 8, total_step 135980, total loss is 16.51 , inference loss is 9.42, weight deacy loss is 7.09, training accuracy is 0.406250, time 298.677 samples/sec
+testing verification..
+(12000, 512)
+infer time 31.388446
+best_threshold_index 105 0.9851851851851852
+best_threshold_index 105 0.9835185185185186
+best_threshold_index 103 0.9844444444444445
+best_threshold_index 105 0.9853703703703703
+best_threshold_index 105 0.9853703703703703
+best_threshold_index 105 0.9840740740740741
+best_threshold_index 105 0.9846296296296296
+best_threshold_index 101 0.9840740740740741
+best_threshold_index 105 0.9829629629629629
+best_threshold_index 105 0.9829629629629629
+[lfw][136000]XNorm: 23.026528
+[lfw][136000]Accuracy-Flip: 0.98300+-0.00781
+test accuracy is: 0.9829999999999999
+epoch 8, total_step 136000, total loss is 18.21 , inference loss is 11.12, weight deacy loss is 7.09, training accuracy is 0.187500, time 7.456 samples/sec
+epoch 8, total_step 136020, total loss is 17.97 , inference loss is 10.88, weight deacy loss is 7.09, training accuracy is 0.375000, time 222.716 samples/sec
+epoch 8, total_step 136040, total loss is 15.43 , inference loss is 8.34, weight deacy loss is 7.09, training accuracy is 0.312500, time 299.211 samples/sec
+epoch 8, total_step 136060, total loss is 14.15 , inference loss is 7.06, weight deacy loss is 7.09, training accuracy is 0.468750, time 298.187 samples/sec
+epoch 8, total_step 136080, total loss is 16.57 , inference loss is 9.48, weight deacy loss is 7.09, training accuracy is 0.312500, time 298.820 samples/sec
+epoch 8, total_step 136100, total loss is 15.72 , inference loss is 8.64, weight deacy loss is 7.09, training accuracy is 0.406250, time 299.046 samples/sec
+epoch 8, total_step 136120, total loss is 18.53 , inference loss is 11.45, weight deacy loss is 7.09, training accuracy is 0.218750, time 299.493 samples/sec
+epoch 8, total_step 136140, total loss is 18.69 , inference loss is 11.60, weight deacy loss is 7.09, training accuracy is 0.218750, time 179.871 samples/sec
+epoch 8, total_step 136160, total loss is 19.09 , inference loss is 12.01, weight deacy loss is 7.09, training accuracy is 0.281250, time 192.883 samples/sec
+epoch 8, total_step 136180, total loss is 20.91 , inference loss is 13.83, weight deacy loss is 7.08, training accuracy is 0.156250, time 214.663 samples/sec
+epoch 8, total_step 136200, total loss is 17.73 , inference loss is 10.65, weight deacy loss is 7.08, training accuracy is 0.343750, time 2.855 samples/sec
+epoch 8, total_step 136220, total loss is 16.09 , inference loss is 9.00, weight deacy loss is 7.08, training accuracy is 0.281250, time 299.953 samples/sec
+epoch 8, total_step 136240, total loss is 19.99 , inference loss is 12.91, weight deacy loss is 7.08, training accuracy is 0.187500, time 216.356 samples/sec
+epoch 8, total_step 136260, total loss is 14.38 , inference loss is 7.29, weight deacy loss is 7.08, training accuracy is 0.406250, time 217.060 samples/sec
+epoch 8, total_step 136280, total loss is 16.05 , inference loss is 8.96, weight deacy loss is 7.08, training accuracy is 0.250000, time 208.088 samples/sec
+epoch 8, total_step 136300, total loss is 14.00 , inference loss is 6.92, weight deacy loss is 7.08, training accuracy is 0.312500, time 215.741 samples/sec
+epoch 8, total_step 136320, total loss is 14.11 , inference loss is 7.03, weight deacy loss is 7.08, training accuracy is 0.468750, time 223.530 samples/sec
+epoch 8, total_step 136340, total loss is 17.06 , inference loss is 9.98, weight deacy loss is 7.08, training accuracy is 0.250000, time 219.332 samples/sec
+epoch 8, total_step 136360, total loss is 15.08 , inference loss is 8.00, weight deacy loss is 7.08, training accuracy is 0.281250, time 210.525 samples/sec
+epoch 8, total_step 136380, total loss is 20.37 , inference loss is 13.29, weight deacy loss is 7.08, training accuracy is 0.156250, time 210.455 samples/sec
+epoch 8, total_step 136400, total loss is 13.55 , inference loss is 6.47, weight deacy loss is 7.08, training accuracy is 0.343750, time 216.299 samples/sec
+epoch 8, total_step 136420, total loss is 14.72 , inference loss is 7.64, weight deacy loss is 7.08, training accuracy is 0.375000, time 232.436 samples/sec
+epoch 8, total_step 136440, total loss is 12.42 , inference loss is 5.34, weight deacy loss is 7.08, training accuracy is 0.593750, time 230.280 samples/sec
+epoch 8, total_step 136460, total loss is 15.96 , inference loss is 8.88, weight deacy loss is 7.08, training accuracy is 0.312500, time 218.138 samples/sec
+epoch 8, total_step 136480, total loss is 14.37 , inference loss is 7.29, weight deacy loss is 7.08, training accuracy is 0.250000, time 217.243 samples/sec
+epoch 8, total_step 136500, total loss is 16.12 , inference loss is 9.04, weight deacy loss is 7.08, training accuracy is 0.281250, time 3.511 samples/sec
+epoch 8, total_step 136520, total loss is 18.52 , inference loss is 11.45, weight deacy loss is 7.08, training accuracy is 0.281250, time 290.954 samples/sec
+epoch 8, total_step 136540, total loss is 16.72 , inference loss is 9.64, weight deacy loss is 7.08, training accuracy is 0.218750, time 299.592 samples/sec
+epoch 8, total_step 136560, total loss is 18.24 , inference loss is 11.17, weight deacy loss is 7.08, training accuracy is 0.281250, time 277.824 samples/sec
+epoch 8, total_step 136580, total loss is 13.91 , inference loss is 6.84, weight deacy loss is 7.08, training accuracy is 0.343750, time 299.550 samples/sec
+epoch 8, total_step 136600, total loss is 17.78 , inference loss is 10.71, weight deacy loss is 7.08, training accuracy is 0.437500, time 298.834 samples/sec
+epoch 8, total_step 136620, total loss is 15.20 , inference loss is 8.12, weight deacy loss is 7.07, training accuracy is 0.281250, time 300.109 samples/sec
+epoch 8, total_step 136640, total loss is 18.95 , inference loss is 11.88, weight deacy loss is 7.07, training accuracy is 0.375000, time 296.839 samples/sec
+epoch 8, total_step 136660, total loss is 16.95 , inference loss is 9.88, weight deacy loss is 7.07, training accuracy is 0.375000, time 298.496 samples/sec
+epoch 8, total_step 136680, total loss is 17.42 , inference loss is 10.35, weight deacy loss is 7.07, training accuracy is 0.312500, time 298.589 samples/sec
+epoch 8, total_step 136700, total loss is 16.29 , inference loss is 9.22, weight deacy loss is 7.07, training accuracy is 0.343750, time 298.488 samples/sec
+epoch 8, total_step 136720, total loss is 20.92 , inference loss is 13.85, weight deacy loss is 7.07, training accuracy is 0.156250, time 299.704 samples/sec
+epoch 8, total_step 136740, total loss is 21.46 , inference loss is 14.39, weight deacy loss is 7.07, training accuracy is 0.218750, time 299.242 samples/sec
+epoch 8, total_step 136760, total loss is 17.97 , inference loss is 10.89, weight deacy loss is 7.07, training accuracy is 0.125000, time 299.260 samples/sec
+epoch 8, total_step 136780, total loss is 16.35 , inference loss is 9.28, weight deacy loss is 7.07, training accuracy is 0.218750, time 300.234 samples/sec
+epoch 8, total_step 136800, total loss is 19.11 , inference loss is 12.04, weight deacy loss is 7.07, training accuracy is 0.281250, time 6.091 samples/sec
+epoch 8, total_step 136820, total loss is 19.20 , inference loss is 12.13, weight deacy loss is 7.07, training accuracy is 0.312500, time 216.157 samples/sec
+epoch 8, total_step 136840, total loss is 14.12 , inference loss is 7.05, weight deacy loss is 7.07, training accuracy is 0.468750, time 224.892 samples/sec
+epoch 8, total_step 136860, total loss is 18.83 , inference loss is 11.76, weight deacy loss is 7.07, training accuracy is 0.250000, time 216.799 samples/sec
+epoch 8, total_step 136880, total loss is 16.96 , inference loss is 9.89, weight deacy loss is 7.07, training accuracy is 0.218750, time 214.351 samples/sec
+epoch 8, total_step 136900, total loss is 17.15 , inference loss is 10.08, weight deacy loss is 7.07, training accuracy is 0.281250, time 210.066 samples/sec
+epoch 8, total_step 136920, total loss is 22.72 , inference loss is 15.66, weight deacy loss is 7.07, training accuracy is 0.218750, time 220.655 samples/sec
+epoch 8, total_step 136940, total loss is 18.46 , inference loss is 11.40, weight deacy loss is 7.07, training accuracy is 0.250000, time 219.624 samples/sec
+epoch 8, total_step 136960, total loss is 16.12 , inference loss is 9.05, weight deacy loss is 7.07, training accuracy is 0.312500, time 212.675 samples/sec
+epoch 8, total_step 136980, total loss is 19.71 , inference loss is 12.64, weight deacy loss is 7.07, training accuracy is 0.406250, time 235.819 samples/sec
+epoch 8, total_step 137000, total loss is 16.29 , inference loss is 9.22, weight deacy loss is 7.07, training accuracy is 0.406250, time 211.767 samples/sec
+epoch 8, total_step 137020, total loss is 21.31 , inference loss is 14.24, weight deacy loss is 7.07, training accuracy is 0.250000, time 223.570 samples/sec
+epoch 8, total_step 137040, total loss is 17.26 , inference loss is 10.20, weight deacy loss is 7.07, training accuracy is 0.281250, time 210.029 samples/sec
+epoch 8, total_step 137060, total loss is 17.96 , inference loss is 10.89, weight deacy loss is 7.06, training accuracy is 0.343750, time 217.526 samples/sec
+epoch 8, total_step 137080, total loss is 18.47 , inference loss is 11.41, weight deacy loss is 7.06, training accuracy is 0.312500, time 218.075 samples/sec
+epoch 8, total_step 137100, total loss is 18.90 , inference loss is 11.84, weight deacy loss is 7.06, training accuracy is 0.312500, time 12.843 samples/sec
+epoch 8, total_step 137120, total loss is 16.85 , inference loss is 9.79, weight deacy loss is 7.06, training accuracy is 0.218750, time 215.233 samples/sec
+epoch 8, total_step 137140, total loss is 19.26 , inference loss is 12.20, weight deacy loss is 7.06, training accuracy is 0.312500, time 220.383 samples/sec
+epoch 8, total_step 137160, total loss is 18.19 , inference loss is 11.13, weight deacy loss is 7.06, training accuracy is 0.406250, time 209.576 samples/sec
+epoch 8, total_step 137180, total loss is 19.13 , inference loss is 12.07, weight deacy loss is 7.06, training accuracy is 0.156250, time 216.374 samples/sec
+epoch 8, total_step 137200, total loss is 13.95 , inference loss is 6.89, weight deacy loss is 7.06, training accuracy is 0.406250, time 209.458 samples/sec
+epoch 8, total_step 137220, total loss is 16.81 , inference loss is 9.75, weight deacy loss is 7.06, training accuracy is 0.250000, time 220.161 samples/sec
+epoch 8, total_step 137240, total loss is 19.86 , inference loss is 12.79, weight deacy loss is 7.06, training accuracy is 0.250000, time 214.920 samples/sec
+epoch 8, total_step 137260, total loss is 17.59 , inference loss is 10.53, weight deacy loss is 7.06, training accuracy is 0.281250, time 232.405 samples/sec
+epoch 8, total_step 137280, total loss is 17.14 , inference loss is 10.08, weight deacy loss is 7.06, training accuracy is 0.312500, time 223.835 samples/sec
+epoch 8, total_step 137300, total loss is 19.54 , inference loss is 12.48, weight deacy loss is 7.06, training accuracy is 0.281250, time 214.629 samples/sec
+epoch 8, total_step 137320, total loss is 17.37 , inference loss is 10.31, weight deacy loss is 7.06, training accuracy is 0.406250, time 214.192 samples/sec
+epoch 8, total_step 137340, total loss is 21.77 , inference loss is 14.71, weight deacy loss is 7.06, training accuracy is 0.250000, time 209.654 samples/sec
+epoch 8, total_step 137360, total loss is 15.47 , inference loss is 8.41, weight deacy loss is 7.06, training accuracy is 0.312500, time 216.568 samples/sec
+epoch 8, total_step 137380, total loss is 20.56 , inference loss is 13.50, weight deacy loss is 7.06, training accuracy is 0.187500, time 299.811 samples/sec
+epoch 8, total_step 137400, total loss is 19.66 , inference loss is 12.60, weight deacy loss is 7.06, training accuracy is 0.250000, time 4.502 samples/sec
+epoch 8, total_step 137420, total loss is 16.48 , inference loss is 9.43, weight deacy loss is 7.06, training accuracy is 0.343750, time 224.783 samples/sec
+epoch 8, total_step 137440, total loss is 15.83 , inference loss is 8.77, weight deacy loss is 7.06, training accuracy is 0.218750, time 204.320 samples/sec
+epoch 8, total_step 137460, total loss is 20.09 , inference loss is 13.04, weight deacy loss is 7.06, training accuracy is 0.375000, time 216.503 samples/sec
+epoch 8, total_step 137480, total loss is 16.52 , inference loss is 9.46, weight deacy loss is 7.06, training accuracy is 0.312500, time 208.758 samples/sec
+epoch 8, total_step 137500, total loss is 21.62 , inference loss is 14.57, weight deacy loss is 7.06, training accuracy is 0.125000, time 195.986 samples/sec
+epoch 8, total_step 137520, total loss is 13.55 , inference loss is 6.49, weight deacy loss is 7.05, training accuracy is 0.375000, time 219.642 samples/sec
+epoch 8, total_step 137540, total loss is 20.11 , inference loss is 13.06, weight deacy loss is 7.05, training accuracy is 0.250000, time 299.170 samples/sec
+epoch 8, total_step 137560, total loss is 17.50 , inference loss is 10.45, weight deacy loss is 7.05, training accuracy is 0.250000, time 300.037 samples/sec
+epoch 8, total_step 137580, total loss is 18.48 , inference loss is 11.43, weight deacy loss is 7.05, training accuracy is 0.187500, time 298.515 samples/sec
+epoch 8, total_step 137600, total loss is 18.65 , inference loss is 11.59, weight deacy loss is 7.05, training accuracy is 0.250000, time 299.566 samples/sec
+epoch 8, total_step 137620, total loss is 15.58 , inference loss is 8.53, weight deacy loss is 7.05, training accuracy is 0.281250, time 298.819 samples/sec
+epoch 8, total_step 137640, total loss is 19.63 , inference loss is 12.58, weight deacy loss is 7.05, training accuracy is 0.343750, time 298.739 samples/sec
+epoch 8, total_step 137660, total loss is 13.80 , inference loss is 6.75, weight deacy loss is 7.05, training accuracy is 0.437500, time 222.530 samples/sec
+epoch 8, total_step 137680, total loss is 17.80 , inference loss is 10.75, weight deacy loss is 7.05, training accuracy is 0.218750, time 175.909 samples/sec
+epoch 8, total_step 137700, total loss is 17.78 , inference loss is 10.73, weight deacy loss is 7.05, training accuracy is 0.281250, time 3.570 samples/sec
+epoch 8, total_step 137720, total loss is 14.94 , inference loss is 7.89, weight deacy loss is 7.05, training accuracy is 0.375000, time 299.916 samples/sec
+epoch 8, total_step 137740, total loss is 21.17 , inference loss is 14.12, weight deacy loss is 7.05, training accuracy is 0.156250, time 298.662 samples/sec
+epoch 8, total_step 137760, total loss is 17.74 , inference loss is 10.69, weight deacy loss is 7.05, training accuracy is 0.250000, time 299.473 samples/sec
+epoch 8, total_step 137780, total loss is 16.27 , inference loss is 9.22, weight deacy loss is 7.05, training accuracy is 0.156250, time 298.746 samples/sec
+epoch 8, total_step 137800, total loss is 21.73 , inference loss is 14.68, weight deacy loss is 7.05, training accuracy is 0.156250, time 298.505 samples/sec
+epoch 8, total_step 137820, total loss is 23.26 , inference loss is 16.21, weight deacy loss is 7.05, training accuracy is 0.406250, time 298.353 samples/sec2022-12-01 09:20:03.005269: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7416 of 10000
+2022-12-01 09:20:06.544714: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 8, total_step 137840, total loss is 16.16 , inference loss is 9.11, weight deacy loss is 7.05, training accuracy is 0.437500, time 299.352 samples/sec
+epoch 8, total_step 137860, total loss is 13.77 , inference loss is 6.72, weight deacy loss is 7.05, training accuracy is 0.468750, time 298.754 samples/sec
+epoch 8, total_step 137880, total loss is 18.65 , inference loss is 11.60, weight deacy loss is 7.05, training accuracy is 0.406250, time 202.256 samples/sec
+epoch 8, total_step 137900, total loss is 16.23 , inference loss is 9.18, weight deacy loss is 7.05, training accuracy is 0.250000, time 171.017 samples/sec
+epoch 8, total_step 137920, total loss is 14.74 , inference loss is 7.69, weight deacy loss is 7.05, training accuracy is 0.437500, time 175.351 samples/sec
+epoch 8, total_step 137940, total loss is 18.72 , inference loss is 11.67, weight deacy loss is 7.05, training accuracy is 0.250000, time 175.878 samples/sec
+epoch 8, total_step 137960, total loss is 17.51 , inference loss is 10.47, weight deacy loss is 7.05, training accuracy is 0.187500, time 177.047 samples/sec
+End of epoch 8
+epoch 9, total_step 137980, total loss is 29.03 , inference loss is 21.98, weight deacy loss is 7.04, training accuracy is 0.000000, time 203.906 samples/sec
+testing verification..
+(12000, 512)
+infer time 34.38512400000001
+best_threshold_index 110 0.9842592592592593
+best_threshold_index 110 0.9831481481481481
+best_threshold_index 110 0.9837037037037037
+best_threshold_index 110 0.985
+best_threshold_index 110 0.9848148148148148
+best_threshold_index 110 0.9824074074074074
+best_threshold_index 110 0.9848148148148148
+best_threshold_index 110 0.9838888888888889
+best_threshold_index 110 0.982037037037037
+best_threshold_index 110 0.9825925925925926
+[lfw][138000]XNorm: 18.740737
+[lfw][138000]Accuracy-Flip: 0.98367+-0.00921
+test accuracy is: 0.9836666666666668
+epoch 9, total_step 138000, total loss is 14.06 , inference loss is 7.01, weight deacy loss is 7.04, training accuracy is 0.406250, time 3.788 samples/sec
+epoch 9, total_step 138020, total loss is 14.46 , inference loss is 7.41, weight deacy loss is 7.04, training accuracy is 0.437500, time 298.894 samples/sec
+epoch 9, total_step 138040, total loss is 14.82 , inference loss is 7.77, weight deacy loss is 7.04, training accuracy is 0.562500, time 298.451 samples/sec
+epoch 9, total_step 138060, total loss is 15.79 , inference loss is 8.75, weight deacy loss is 7.04, training accuracy is 0.437500, time 298.587 samples/sec
+epoch 9, total_step 138080, total loss is 8.78 , inference loss is 1.74, weight deacy loss is 7.04, training accuracy is 0.718750, time 299.016 samples/sec
+epoch 9, total_step 138100, total loss is 13.19 , inference loss is 6.15, weight deacy loss is 7.04, training accuracy is 0.500000, time 299.561 samples/sec
+epoch 9, total_step 138120, total loss is 12.59 , inference loss is 5.55, weight deacy loss is 7.04, training accuracy is 0.687500, time 299.469 samples/sec
+epoch 9, total_step 138140, total loss is 16.75 , inference loss is 9.71, weight deacy loss is 7.04, training accuracy is 0.468750, time 296.026 samples/sec
+epoch 9, total_step 138160, total loss is 10.62 , inference loss is 3.58, weight deacy loss is 7.04, training accuracy is 0.781250, time 249.892 samples/sec
+epoch 9, total_step 138180, total loss is 10.59 , inference loss is 3.55, weight deacy loss is 7.04, training accuracy is 0.687500, time 248.614 samples/sec
+epoch 9, total_step 138200, total loss is 12.21 , inference loss is 5.17, weight deacy loss is 7.04, training accuracy is 0.718750, time 244.888 samples/sec
+epoch 9, total_step 138220, total loss is 11.51 , inference loss is 4.47, weight deacy loss is 7.04, training accuracy is 0.656250, time 257.241 samples/sec
+epoch 9, total_step 138240, total loss is 10.46 , inference loss is 3.42, weight deacy loss is 7.04, training accuracy is 0.687500, time 245.535 samples/sec
+epoch 9, total_step 138260, total loss is 16.55 , inference loss is 9.51, weight deacy loss is 7.04, training accuracy is 0.562500, time 255.343 samples/sec
+epoch 9, total_step 138280, total loss is 14.05 , inference loss is 7.01, weight deacy loss is 7.04, training accuracy is 0.718750, time 247.485 samples/sec
+epoch 9, total_step 138300, total loss is 10.72 , inference loss is 3.68, weight deacy loss is 7.04, training accuracy is 0.625000, time 3.743 samples/sec
+epoch 9, total_step 138320, total loss is 17.37 , inference loss is 10.33, weight deacy loss is 7.04, training accuracy is 0.500000, time 299.556 samples/sec
+epoch 9, total_step 138340, total loss is 13.00 , inference loss is 5.96, weight deacy loss is 7.04, training accuracy is 0.562500, time 299.083 samples/sec
+epoch 9, total_step 138360, total loss is 14.07 , inference loss is 7.04, weight deacy loss is 7.04, training accuracy is 0.437500, time 300.149 samples/sec
+epoch 9, total_step 138380, total loss is 14.65 , inference loss is 7.61, weight deacy loss is 7.04, training accuracy is 0.562500, time 299.833 samples/sec
+epoch 9, total_step 138400, total loss is 9.23 , inference loss is 2.20, weight deacy loss is 7.04, training accuracy is 0.656250, time 298.550 samples/sec
+epoch 9, total_step 138420, total loss is 14.50 , inference loss is 7.46, weight deacy loss is 7.04, training accuracy is 0.406250, time 299.183 samples/sec
+epoch 9, total_step 138440, total loss is 13.51 , inference loss is 6.48, weight deacy loss is 7.04, training accuracy is 0.500000, time 299.060 samples/sec
+epoch 9, total_step 138460, total loss is 10.27 , inference loss is 3.23, weight deacy loss is 7.03, training accuracy is 0.750000, time 299.594 samples/sec
+epoch 9, total_step 138480, total loss is 11.15 , inference loss is 4.11, weight deacy loss is 7.03, training accuracy is 0.593750, time 299.642 samples/sec
+epoch 9, total_step 138500, total loss is 8.98 , inference loss is 1.94, weight deacy loss is 7.03, training accuracy is 0.718750, time 299.765 samples/sec
+epoch 9, total_step 138520, total loss is 13.73 , inference loss is 6.69, weight deacy loss is 7.03, training accuracy is 0.437500, time 299.880 samples/sec
+epoch 9, total_step 138540, total loss is 13.50 , inference loss is 6.47, weight deacy loss is 7.03, training accuracy is 0.531250, time 299.158 samples/sec
+epoch 9, total_step 138560, total loss is 15.14 , inference loss is 8.10, weight deacy loss is 7.03, training accuracy is 0.562500, time 299.564 samples/sec
+epoch 9, total_step 138580, total loss is 15.06 , inference loss is 8.03, weight deacy loss is 7.03, training accuracy is 0.500000, time 299.534 samples/sec
+epoch 9, total_step 138600, total loss is 12.31 , inference loss is 5.27, weight deacy loss is 7.03, training accuracy is 0.500000, time 2.788 samples/sec
+epoch 9, total_step 138620, total loss is 14.81 , inference loss is 7.78, weight deacy loss is 7.03, training accuracy is 0.437500, time 216.526 samples/sec
+epoch 9, total_step 138640, total loss is 11.33 , inference loss is 4.29, weight deacy loss is 7.03, training accuracy is 0.531250, time 223.669 samples/sec
+epoch 9, total_step 138660, total loss is 15.52 , inference loss is 8.49, weight deacy loss is 7.03, training accuracy is 0.593750, time 210.999 samples/sec
+epoch 9, total_step 138680, total loss is 16.21 , inference loss is 9.18, weight deacy loss is 7.03, training accuracy is 0.562500, time 216.928 samples/sec
+epoch 9, total_step 138700, total loss is 16.41 , inference loss is 9.38, weight deacy loss is 7.03, training accuracy is 0.500000, time 280.233 samples/sec
+epoch 9, total_step 138720, total loss is 15.13 , inference loss is 8.10, weight deacy loss is 7.03, training accuracy is 0.437500, time 299.514 samples/sec
+epoch 9, total_step 138740, total loss is 13.55 , inference loss is 6.52, weight deacy loss is 7.03, training accuracy is 0.593750, time 298.506 samples/sec
+epoch 9, total_step 138760, total loss is 13.32 , inference loss is 6.29, weight deacy loss is 7.03, training accuracy is 0.500000, time 299.269 samples/sec
+epoch 9, total_step 138780, total loss is 19.00 , inference loss is 11.97, weight deacy loss is 7.03, training accuracy is 0.343750, time 294.344 samples/sec
+epoch 9, total_step 138800, total loss is 15.14 , inference loss is 8.11, weight deacy loss is 7.03, training accuracy is 0.500000, time 298.599 samples/sec
+epoch 9, total_step 138820, total loss is 14.01 , inference loss is 6.98, weight deacy loss is 7.03, training accuracy is 0.562500, time 299.122 samples/sec
+epoch 9, total_step 138840, total loss is 12.72 , inference loss is 5.69, weight deacy loss is 7.03, training accuracy is 0.625000, time 219.757 samples/sec
+epoch 9, total_step 138860, total loss is 14.13 , inference loss is 7.10, weight deacy loss is 7.03, training accuracy is 0.500000, time 235.966 samples/sec
+epoch 9, total_step 138880, total loss is 13.58 , inference loss is 6.55, weight deacy loss is 7.03, training accuracy is 0.531250, time 229.119 samples/sec
+epoch 9, total_step 138900, total loss is 14.55 , inference loss is 7.52, weight deacy loss is 7.03, training accuracy is 0.406250, time 6.294 samples/sec
+epoch 9, total_step 138920, total loss is 16.50 , inference loss is 9.47, weight deacy loss is 7.03, training accuracy is 0.250000, time 224.860 samples/sec
+epoch 9, total_step 138940, total loss is 11.27 , inference loss is 4.24, weight deacy loss is 7.03, training accuracy is 0.500000, time 217.915 samples/sec
+epoch 9, total_step 138960, total loss is 15.95 , inference loss is 8.92, weight deacy loss is 7.03, training accuracy is 0.500000, time 296.253 samples/sec
+epoch 9, total_step 138980, total loss is 15.09 , inference loss is 8.06, weight deacy loss is 7.03, training accuracy is 0.437500, time 299.204 samples/sec
+epoch 9, total_step 139000, total loss is 15.09 , inference loss is 8.06, weight deacy loss is 7.03, training accuracy is 0.531250, time 298.225 samples/sec
+epoch 9, total_step 139020, total loss is 15.42 , inference loss is 8.39, weight deacy loss is 7.03, training accuracy is 0.437500, time 298.861 samples/sec
+epoch 9, total_step 139040, total loss is 13.13 , inference loss is 6.10, weight deacy loss is 7.03, training accuracy is 0.531250, time 215.141 samples/sec
+epoch 9, total_step 139060, total loss is 13.52 , inference loss is 6.49, weight deacy loss is 7.03, training accuracy is 0.562500, time 236.585 samples/sec
+epoch 9, total_step 139080, total loss is 14.48 , inference loss is 7.45, weight deacy loss is 7.03, training accuracy is 0.562500, time 220.871 samples/sec
+epoch 9, total_step 139100, total loss is 10.23 , inference loss is 3.20, weight deacy loss is 7.03, training accuracy is 0.656250, time 236.552 samples/sec
+epoch 9, total_step 139120, total loss is 18.12 , inference loss is 11.09, weight deacy loss is 7.03, training accuracy is 0.437500, time 225.711 samples/sec
+epoch 9, total_step 139140, total loss is 12.97 , inference loss is 5.94, weight deacy loss is 7.03, training accuracy is 0.531250, time 234.865 samples/sec
+epoch 9, total_step 139160, total loss is 17.49 , inference loss is 10.46, weight deacy loss is 7.03, training accuracy is 0.406250, time 219.035 samples/sec
+epoch 9, total_step 139180, total loss is 13.00 , inference loss is 5.97, weight deacy loss is 7.03, training accuracy is 0.687500, time 225.225 samples/sec
+epoch 9, total_step 139200, total loss is 12.73 , inference loss is 5.70, weight deacy loss is 7.03, training accuracy is 0.562500, time 1.883 samples/sec
+epoch 9, total_step 139220, total loss is 20.37 , inference loss is 13.34, weight deacy loss is 7.03, training accuracy is 0.312500, time 226.997 samples/sec
+epoch 9, total_step 139240, total loss is 13.54 , inference loss is 6.51, weight deacy loss is 7.03, training accuracy is 0.562500, time 213.979 samples/sec
+epoch 9, total_step 139260, total loss is 12.48 , inference loss is 5.45, weight deacy loss is 7.03, training accuracy is 0.500000, time 227.375 samples/sec
+epoch 9, total_step 139280, total loss is 15.40 , inference loss is 8.37, weight deacy loss is 7.03, training accuracy is 0.531250, time 206.984 samples/sec
+epoch 9, total_step 139300, total loss is 13.96 , inference loss is 6.93, weight deacy loss is 7.03, training accuracy is 0.531250, time 216.622 samples/sec
+epoch 9, total_step 139320, total loss is 15.16 , inference loss is 8.13, weight deacy loss is 7.03, training accuracy is 0.406250, time 218.054 samples/sec
+epoch 9, total_step 139340, total loss is 16.78 , inference loss is 9.75, weight deacy loss is 7.03, training accuracy is 0.312500, time 201.036 samples/sec
+epoch 9, total_step 139360, total loss is 16.91 , inference loss is 9.88, weight deacy loss is 7.03, training accuracy is 0.375000, time 241.920 samples/sec
+epoch 9, total_step 139380, total loss is 13.73 , inference loss is 6.70, weight deacy loss is 7.03, training accuracy is 0.531250, time 299.285 samples/sec
+epoch 9, total_step 139400, total loss is 17.37 , inference loss is 10.34, weight deacy loss is 7.03, training accuracy is 0.343750, time 299.617 samples/sec
+epoch 9, total_step 139420, total loss is 11.50 , inference loss is 4.47, weight deacy loss is 7.03, training accuracy is 0.531250, time 299.394 samples/sec
+epoch 9, total_step 139440, total loss is 14.23 , inference loss is 7.20, weight deacy loss is 7.03, training accuracy is 0.343750, time 299.906 samples/sec
+epoch 9, total_step 139460, total loss is 15.92 , inference loss is 8.89, weight deacy loss is 7.03, training accuracy is 0.406250, time 299.060 samples/sec
+epoch 9, total_step 139480, total loss is 10.91 , inference loss is 3.89, weight deacy loss is 7.03, training accuracy is 0.625000, time 296.595 samples/sec
+epoch 9, total_step 139500, total loss is 11.89 , inference loss is 4.87, weight deacy loss is 7.03, training accuracy is 0.593750, time 2.866 samples/sec
+epoch 9, total_step 139520, total loss is 14.25 , inference loss is 7.23, weight deacy loss is 7.03, training accuracy is 0.406250, time 290.605 samples/sec
+epoch 9, total_step 139540, total loss is 11.59 , inference loss is 4.57, weight deacy loss is 7.03, training accuracy is 0.562500, time 243.289 samples/sec
+epoch 9, total_step 139560, total loss is 14.13 , inference loss is 7.11, weight deacy loss is 7.03, training accuracy is 0.531250, time 298.973 samples/sec
+epoch 9, total_step 139580, total loss is 19.67 , inference loss is 12.65, weight deacy loss is 7.03, training accuracy is 0.468750, time 228.078 samples/sec
+epoch 9, total_step 139600, total loss is 12.39 , inference loss is 5.36, weight deacy loss is 7.02, training accuracy is 0.625000, time 215.517 samples/sec
+epoch 9, total_step 139620, total loss is 13.92 , inference loss is 6.89, weight deacy loss is 7.02, training accuracy is 0.500000, time 202.950 samples/sec
+epoch 9, total_step 139640, total loss is 11.16 , inference loss is 4.14, weight deacy loss is 7.02, training accuracy is 0.593750, time 213.302 samples/sec
+epoch 9, total_step 139660, total loss is 13.88 , inference loss is 6.86, weight deacy loss is 7.02, training accuracy is 0.437500, time 217.428 samples/sec
+epoch 9, total_step 139680, total loss is 16.25 , inference loss is 9.23, weight deacy loss is 7.02, training accuracy is 0.281250, time 212.163 samples/sec
+epoch 9, total_step 139700, total loss is 14.52 , inference loss is 7.49, weight deacy loss is 7.02, training accuracy is 0.531250, time 229.538 samples/sec
+epoch 9, total_step 139720, total loss is 14.34 , inference loss is 7.32, weight deacy loss is 7.02, training accuracy is 0.375000, time 224.661 samples/sec
+epoch 9, total_step 139740, total loss is 16.93 , inference loss is 9.91, weight deacy loss is 7.02, training accuracy is 0.437500, time 214.047 samples/sec
+epoch 9, total_step 139760, total loss is 16.43 , inference loss is 9.40, weight deacy loss is 7.02, training accuracy is 0.562500, time 227.434 samples/sec
+epoch 9, total_step 139780, total loss is 14.67 , inference loss is 7.64, weight deacy loss is 7.02, training accuracy is 0.562500, time 215.101 samples/sec
+epoch 9, total_step 139800, total loss is 13.57 , inference loss is 6.55, weight deacy loss is 7.02, training accuracy is 0.562500, time 3.456 samples/sec
+epoch 9, total_step 139820, total loss is 14.75 , inference loss is 7.73, weight deacy loss is 7.02, training accuracy is 0.406250, time 210.028 samples/sec
+epoch 9, total_step 139840, total loss is 16.13 , inference loss is 9.11, weight deacy loss is 7.02, training accuracy is 0.406250, time 219.094 samples/sec
+epoch 9, total_step 139860, total loss is 15.52 , inference loss is 8.49, weight deacy loss is 7.02, training accuracy is 0.468750, time 207.869 samples/sec
+epoch 9, total_step 139880, total loss is 11.27 , inference loss is 4.24, weight deacy loss is 7.02, training accuracy is 0.625000, time 207.927 samples/sec
+epoch 9, total_step 139900, total loss is 16.68 , inference loss is 9.66, weight deacy loss is 7.02, training accuracy is 0.375000, time 219.729 samples/sec
+epoch 9, total_step 139920, total loss is 11.18 , inference loss is 4.16, weight deacy loss is 7.02, training accuracy is 0.593750, time 216.864 samples/sec
+epoch 9, total_step 139940, total loss is 16.37 , inference loss is 9.35, weight deacy loss is 7.02, training accuracy is 0.406250, time 214.994 samples/sec
+epoch 9, total_step 139960, total loss is 15.05 , inference loss is 8.03, weight deacy loss is 7.02, training accuracy is 0.468750, time 224.804 samples/sec
+epoch 9, total_step 139980, total loss is 12.55 , inference loss is 5.53, weight deacy loss is 7.02, training accuracy is 0.562500, time 226.075 samples/sec
+testing verification..
+(12000, 512)
+infer time 26.396229000000012
+best_threshold_index 120 0.987037037037037
+best_threshold_index 119 0.9868518518518519
+best_threshold_index 112 0.9859259259259259
+best_threshold_index 116 0.9874074074074074
+best_threshold_index 120 0.9864814814814815
+best_threshold_index 120 0.9855555555555555
+best_threshold_index 120 0.9868518518518519
+best_threshold_index 120 0.9855555555555555
+best_threshold_index 120 0.9861111111111112
+best_threshold_index 120 0.9861111111111112
+[lfw][140000]XNorm: 22.054561
+[lfw][140000]Accuracy-Flip: 0.98483+-0.00598
+test accuracy is: 0.9848333333333332
+epoch 9, total_step 140000, total loss is 11.16 , inference loss is 4.14, weight deacy loss is 7.02, training accuracy is 0.562500, time 11.377 samples/sec
+epoch 9, total_step 140020, total loss is 14.90 , inference loss is 7.88, weight deacy loss is 7.02, training accuracy is 0.531250, time 287.287 samples/sec
+epoch 9, total_step 140040, total loss is 12.40 , inference loss is 5.38, weight deacy loss is 7.02, training accuracy is 0.562500, time 208.859 samples/sec
+epoch 9, total_step 140060, total loss is 14.33 , inference loss is 7.31, weight deacy loss is 7.02, training accuracy is 0.375000, time 213.249 samples/sec
+epoch 9, total_step 140080, total loss is 13.82 , inference loss is 6.81, weight deacy loss is 7.02, training accuracy is 0.500000, time 227.163 samples/sec
+epoch 9, total_step 140100, total loss is 14.90 , inference loss is 7.88, weight deacy loss is 7.02, training accuracy is 0.406250, time 4.719 samples/sec
+epoch 9, total_step 140120, total loss is 14.43 , inference loss is 7.41, weight deacy loss is 7.02, training accuracy is 0.406250, time 203.389 samples/sec
+epoch 9, total_step 140140, total loss is 13.67 , inference loss is 6.65, weight deacy loss is 7.02, training accuracy is 0.625000, time 298.117 samples/sec
+epoch 9, total_step 140160, total loss is 11.28 , inference loss is 4.27, weight deacy loss is 7.02, training accuracy is 0.562500, time 298.326 samples/sec
+epoch 9, total_step 140180, total loss is 15.29 , inference loss is 8.27, weight deacy loss is 7.02, training accuracy is 0.500000, time 296.311 samples/sec
+epoch 9, total_step 140200, total loss is 13.74 , inference loss is 6.72, weight deacy loss is 7.02, training accuracy is 0.593750, time 296.354 samples/sec
+epoch 9, total_step 140220, total loss is 8.77 , inference loss is 1.76, weight deacy loss is 7.02, training accuracy is 0.843750, time 299.712 samples/sec
+epoch 9, total_step 140240, total loss is 12.56 , inference loss is 5.54, weight deacy loss is 7.02, training accuracy is 0.687500, time 299.532 samples/sec
+epoch 9, total_step 140260, total loss is 13.17 , inference loss is 6.16, weight deacy loss is 7.02, training accuracy is 0.562500, time 298.980 samples/sec
+epoch 9, total_step 140280, total loss is 10.73 , inference loss is 3.72, weight deacy loss is 7.01, training accuracy is 0.593750, time 215.895 samples/sec
+epoch 9, total_step 140300, total loss is 9.96 , inference loss is 2.95, weight deacy loss is 7.01, training accuracy is 0.531250, time 223.651 samples/sec
+epoch 9, total_step 140320, total loss is 11.95 , inference loss is 4.94, weight deacy loss is 7.01, training accuracy is 0.625000, time 213.981 samples/sec
+epoch 9, total_step 140340, total loss is 15.88 , inference loss is 8.87, weight deacy loss is 7.01, training accuracy is 0.437500, time 226.586 samples/sec
+epoch 9, total_step 140360, total loss is 18.11 , inference loss is 11.09, weight deacy loss is 7.01, training accuracy is 0.468750, time 236.402 samples/sec
+epoch 9, total_step 140380, total loss is 10.72 , inference loss is 3.71, weight deacy loss is 7.01, training accuracy is 0.625000, time 210.102 samples/sec
+epoch 9, total_step 140400, total loss is 12.46 , inference loss is 5.45, weight deacy loss is 7.01, training accuracy is 0.593750, time 4.567 samples/sec
+epoch 9, total_step 140420, total loss is 10.34 , inference loss is 3.33, weight deacy loss is 7.01, training accuracy is 0.593750, time 298.319 samples/sec
+epoch 9, total_step 140440, total loss is 10.86 , inference loss is 3.84, weight deacy loss is 7.01, training accuracy is 0.593750, time 207.937 samples/sec
+epoch 9, total_step 140460, total loss is 12.20 , inference loss is 5.19, weight deacy loss is 7.01, training accuracy is 0.531250, time 210.939 samples/sec
+epoch 9, total_step 140480, total loss is 13.70 , inference loss is 6.69, weight deacy loss is 7.01, training accuracy is 0.437500, time 216.096 samples/sec
+epoch 9, total_step 140500, total loss is 12.08 , inference loss is 5.06, weight deacy loss is 7.01, training accuracy is 0.562500, time 216.191 samples/sec
+epoch 9, total_step 140520, total loss is 13.74 , inference loss is 6.73, weight deacy loss is 7.01, training accuracy is 0.437500, time 226.865 samples/sec
+epoch 9, total_step 140540, total loss is 11.17 , inference loss is 4.16, weight deacy loss is 7.01, training accuracy is 0.593750, time 212.726 samples/sec
+epoch 9, total_step 140560, total loss is 14.79 , inference loss is 7.78, weight deacy loss is 7.01, training accuracy is 0.562500, time 209.374 samples/sec
+epoch 9, total_step 140580, total loss is 15.48 , inference loss is 8.47, weight deacy loss is 7.01, training accuracy is 0.406250, time 213.028 samples/sec
+epoch 9, total_step 140600, total loss is 10.03 , inference loss is 3.02, weight deacy loss is 7.01, training accuracy is 0.625000, time 215.098 samples/sec
+epoch 9, total_step 140620, total loss is 13.81 , inference loss is 6.80, weight deacy loss is 7.01, training accuracy is 0.406250, time 219.761 samples/sec
+epoch 9, total_step 140640, total loss is 12.72 , inference loss is 5.71, weight deacy loss is 7.01, training accuracy is 0.625000, time 211.859 samples/sec
+epoch 9, total_step 140660, total loss is 15.12 , inference loss is 8.11, weight deacy loss is 7.01, training accuracy is 0.562500, time 211.668 samples/sec
+epoch 9, total_step 140680, total loss is 13.03 , inference loss is 6.03, weight deacy loss is 7.01, training accuracy is 0.375000, time 210.481 samples/sec
+epoch 9, total_step 140700, total loss is 10.18 , inference loss is 3.17, weight deacy loss is 7.01, training accuracy is 0.687500, time 3.408 samples/sec
+epoch 9, total_step 140720, total loss is 11.87 , inference loss is 4.86, weight deacy loss is 7.01, training accuracy is 0.656250, time 227.483 samples/sec
+epoch 9, total_step 140740, total loss is 11.77 , inference loss is 4.77, weight deacy loss is 7.01, training accuracy is 0.531250, time 219.309 samples/sec
+epoch 9, total_step 140760, total loss is 12.47 , inference loss is 5.46, weight deacy loss is 7.01, training accuracy is 0.562500, time 230.234 samples/sec
+epoch 9, total_step 140780, total loss is 15.00 , inference loss is 7.99, weight deacy loss is 7.01, training accuracy is 0.437500, time 209.414 samples/sec
+epoch 9, total_step 140800, total loss is 14.37 , inference loss is 7.37, weight deacy loss is 7.00, training accuracy is 0.437500, time 224.666 samples/sec
+epoch 9, total_step 140820, total loss is 12.23 , inference loss is 5.22, weight deacy loss is 7.00, training accuracy is 0.625000, time 219.180 samples/sec
+epoch 9, total_step 140840, total loss is 14.80 , inference loss is 7.80, weight deacy loss is 7.00, training accuracy is 0.500000, time 214.848 samples/sec
+epoch 9, total_step 140860, total loss is 15.99 , inference loss is 8.99, weight deacy loss is 7.00, training accuracy is 0.437500, time 211.854 samples/sec
+epoch 9, total_step 140880, total loss is 12.23 , inference loss is 5.22, weight deacy loss is 7.00, training accuracy is 0.625000, time 218.922 samples/sec
+epoch 9, total_step 140900, total loss is 13.31 , inference loss is 6.31, weight deacy loss is 7.00, training accuracy is 0.718750, time 291.341 samples/sec
+epoch 9, total_step 140920, total loss is 13.02 , inference loss is 6.02, weight deacy loss is 7.00, training accuracy is 0.531250, time 299.594 samples/sec
+epoch 9, total_step 140940, total loss is 10.84 , inference loss is 3.84, weight deacy loss is 7.00, training accuracy is 0.656250, time 299.150 samples/sec
+epoch 9, total_step 140960, total loss is 13.31 , inference loss is 6.31, weight deacy loss is 7.00, training accuracy is 0.562500, time 299.441 samples/sec
+epoch 9, total_step 140980, total loss is 11.69 , inference loss is 4.68, weight deacy loss is 7.00, training accuracy is 0.562500, time 199.610 samples/sec
+epoch 9, total_step 141000, total loss is 11.66 , inference loss is 4.66, weight deacy loss is 7.00, training accuracy is 0.531250, time 6.945 samples/sec
+epoch 9, total_step 141020, total loss is 11.28 , inference loss is 4.28, weight deacy loss is 7.00, training accuracy is 0.718750, time 208.926 samples/sec
+epoch 9, total_step 141040, total loss is 12.96 , inference loss is 5.96, weight deacy loss is 7.00, training accuracy is 0.437500, time 207.731 samples/sec
+epoch 9, total_step 141060, total loss is 12.69 , inference loss is 5.69, weight deacy loss is 7.00, training accuracy is 0.687500, time 240.938 samples/sec
+epoch 9, total_step 141080, total loss is 13.19 , inference loss is 6.19, weight deacy loss is 7.00, training accuracy is 0.718750, time 211.444 samples/sec
+epoch 9, total_step 141100, total loss is 15.53 , inference loss is 8.54, weight deacy loss is 7.00, training accuracy is 0.562500, time 213.725 samples/sec
+epoch 9, total_step 141120, total loss is 12.43 , inference loss is 5.43, weight deacy loss is 7.00, training accuracy is 0.531250, time 219.473 samples/sec
+epoch 9, total_step 141140, total loss is 12.60 , inference loss is 5.60, weight deacy loss is 7.00, training accuracy is 0.562500, time 225.721 samples/sec
+epoch 9, total_step 141160, total loss is 12.37 , inference loss is 5.37, weight deacy loss is 7.00, training accuracy is 0.593750, time 215.320 samples/sec
+epoch 9, total_step 141180, total loss is 12.58 , inference loss is 5.58, weight deacy loss is 7.00, training accuracy is 0.500000, time 214.928 samples/sec
+epoch 9, total_step 141200, total loss is 14.72 , inference loss is 7.72, weight deacy loss is 7.00, training accuracy is 0.593750, time 299.374 samples/sec
+epoch 9, total_step 141220, total loss is 13.59 , inference loss is 6.59, weight deacy loss is 7.00, training accuracy is 0.593750, time 300.041 samples/sec
+epoch 9, total_step 141240, total loss is 19.19 , inference loss is 12.19, weight deacy loss is 7.00, training accuracy is 0.468750, time 298.785 samples/sec
+epoch 9, total_step 141260, total loss is 14.55 , inference loss is 7.55, weight deacy loss is 7.00, training accuracy is 0.500000, time 299.202 samples/sec
+epoch 9, total_step 141280, total loss is 13.96 , inference loss is 6.96, weight deacy loss is 7.00, training accuracy is 0.531250, time 215.611 samples/sec
+epoch 9, total_step 141300, total loss is 16.62 , inference loss is 9.63, weight deacy loss is 7.00, training accuracy is 0.437500, time 5.446 samples/sec
+epoch 9, total_step 141320, total loss is 8.14 , inference loss is 1.15, weight deacy loss is 7.00, training accuracy is 0.781250, time 199.885 samples/sec
+epoch 9, total_step 141340, total loss is 14.44 , inference loss is 7.45, weight deacy loss is 7.00, training accuracy is 0.343750, time 211.602 samples/sec
+epoch 9, total_step 141360, total loss is 13.71 , inference loss is 6.72, weight deacy loss is 7.00, training accuracy is 0.406250, time 232.075 samples/sec
+epoch 9, total_step 141380, total loss is 15.35 , inference loss is 8.35, weight deacy loss is 7.00, training accuracy is 0.437500, time 207.159 samples/sec
+epoch 9, total_step 141400, total loss is 13.35 , inference loss is 6.35, weight deacy loss is 7.00, training accuracy is 0.531250, time 205.870 samples/sec
+epoch 9, total_step 141420, total loss is 12.73 , inference loss is 5.74, weight deacy loss is 7.00, training accuracy is 0.500000, time 298.195 samples/sec
+epoch 9, total_step 141440, total loss is 11.49 , inference loss is 4.50, weight deacy loss is 7.00, training accuracy is 0.562500, time 299.931 samples/sec
+epoch 9, total_step 141460, total loss is 13.27 , inference loss is 6.27, weight deacy loss is 6.99, training accuracy is 0.500000, time 299.309 samples/sec
+epoch 9, total_step 141480, total loss is 15.88 , inference loss is 8.89, weight deacy loss is 6.99, training accuracy is 0.406250, time 299.602 samples/sec
+epoch 9, total_step 141500, total loss is 11.91 , inference loss is 4.92, weight deacy loss is 6.99, training accuracy is 0.593750, time 299.077 samples/sec
+epoch 9, total_step 141520, total loss is 12.62 , inference loss is 5.63, weight deacy loss is 6.99, training accuracy is 0.531250, time 214.982 samples/sec
+epoch 9, total_step 141540, total loss is 18.45 , inference loss is 11.46, weight deacy loss is 6.99, training accuracy is 0.343750, time 209.738 samples/sec
+epoch 9, total_step 141560, total loss is 12.91 , inference loss is 5.92, weight deacy loss is 6.99, training accuracy is 0.531250, time 218.646 samples/sec
+epoch 9, total_step 141580, total loss is 12.07 , inference loss is 5.07, weight deacy loss is 6.99, training accuracy is 0.531250, time 216.535 samples/sec
+epoch 9, total_step 141600, total loss is 15.47 , inference loss is 8.48, weight deacy loss is 6.99, training accuracy is 0.343750, time 8.112 samples/sec
+epoch 9, total_step 141620, total loss is 20.61 , inference loss is 13.62, weight deacy loss is 6.99, training accuracy is 0.406250, time 214.429 samples/sec
+epoch 9, total_step 141640, total loss is 12.90 , inference loss is 5.91, weight deacy loss is 6.99, training accuracy is 0.625000, time 213.807 samples/sec
+epoch 9, total_step 141660, total loss is 13.37 , inference loss is 6.38, weight deacy loss is 6.99, training accuracy is 0.375000, time 216.667 samples/sec
+epoch 9, total_step 141680, total loss is 19.08 , inference loss is 12.09, weight deacy loss is 6.99, training accuracy is 0.312500, time 213.334 samples/sec
+epoch 9, total_step 141700, total loss is 16.71 , inference loss is 9.72, weight deacy loss is 6.99, training accuracy is 0.468750, time 215.118 samples/sec
+epoch 9, total_step 141720, total loss is 12.38 , inference loss is 5.39, weight deacy loss is 6.99, training accuracy is 0.437500, time 218.011 samples/sec
+epoch 9, total_step 141740, total loss is 14.11 , inference loss is 7.12, weight deacy loss is 6.99, training accuracy is 0.468750, time 226.806 samples/sec
+epoch 9, total_step 141760, total loss is 19.15 , inference loss is 12.16, weight deacy loss is 6.99, training accuracy is 0.406250, time 299.615 samples/sec
+epoch 9, total_step 141780, total loss is 14.97 , inference loss is 7.98, weight deacy loss is 6.99, training accuracy is 0.406250, time 299.537 samples/sec
+epoch 9, total_step 141800, total loss is 19.80 , inference loss is 12.81, weight deacy loss is 6.99, training accuracy is 0.343750, time 203.365 samples/sec
+epoch 9, total_step 141820, total loss is 15.91 , inference loss is 8.92, weight deacy loss is 6.99, training accuracy is 0.406250, time 217.459 samples/sec
+epoch 9, total_step 141840, total loss is 18.35 , inference loss is 11.36, weight deacy loss is 6.99, training accuracy is 0.531250, time 214.227 samples/sec
+epoch 9, total_step 141860, total loss is 11.54 , inference loss is 4.55, weight deacy loss is 6.99, training accuracy is 0.468750, time 207.575 samples/sec
+epoch 9, total_step 141880, total loss is 14.17 , inference loss is 7.18, weight deacy loss is 6.99, training accuracy is 0.593750, time 212.660 samples/sec
+epoch 9, total_step 141900, total loss is 14.18 , inference loss is 7.19, weight deacy loss is 6.99, training accuracy is 0.562500, time 13.328 samples/sec
+epoch 9, total_step 141920, total loss is 18.05 , inference loss is 11.06, weight deacy loss is 6.99, training accuracy is 0.468750, time 208.509 samples/sec
+epoch 9, total_step 141940, total loss is 14.93 , inference loss is 7.94, weight deacy loss is 6.99, training accuracy is 0.437500, time 209.860 samples/sec
+epoch 9, total_step 141960, total loss is 13.12 , inference loss is 6.13, weight deacy loss is 6.99, training accuracy is 0.468750, time 222.040 samples/sec
+epoch 9, total_step 141980, total loss is 20.78 , inference loss is 13.80, weight deacy loss is 6.99, training accuracy is 0.343750, time 218.022 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.718047999999996
+best_threshold_index 123 0.9877777777777778
+best_threshold_index 114 0.987037037037037
+best_threshold_index 123 0.9874074074074074
+best_threshold_index 114 0.9879629629629629
+best_threshold_index 123 0.9874074074074074
+best_threshold_index 114 0.9864814814814815
+best_threshold_index 114 0.987037037037037
+best_threshold_index 123 0.9862962962962963
+best_threshold_index 114 0.9859259259259259
+best_threshold_index 123 0.9866666666666667
+[lfw][142000]XNorm: 21.868196
+[lfw][142000]Accuracy-Flip: 0.98533+-0.00562
+test accuracy is: 0.9853333333333334
+epoch 9, total_step 142000, total loss is 17.41 , inference loss is 10.42, weight deacy loss is 6.99, training accuracy is 0.312500, time 17.808 samples/sec
+epoch 9, total_step 142020, total loss is 16.27 , inference loss is 9.28, weight deacy loss is 6.99, training accuracy is 0.437500, time 228.149 samples/sec
+epoch 9, total_step 142040, total loss is 13.95 , inference loss is 6.97, weight deacy loss is 6.99, training accuracy is 0.468750, time 216.259 samples/sec
+epoch 9, total_step 142060, total loss is 13.01 , inference loss is 6.03, weight deacy loss is 6.99, training accuracy is 0.500000, time 216.801 samples/sec
+epoch 9, total_step 142080, total loss is 16.79 , inference loss is 9.81, weight deacy loss is 6.99, training accuracy is 0.500000, time 214.531 samples/sec
+epoch 9, total_step 142100, total loss is 13.01 , inference loss is 6.02, weight deacy loss is 6.99, training accuracy is 0.531250, time 219.072 samples/sec
+epoch 9, total_step 142120, total loss is 15.14 , inference loss is 8.16, weight deacy loss is 6.98, training accuracy is 0.625000, time 226.185 samples/sec
+epoch 9, total_step 142140, total loss is 11.88 , inference loss is 4.90, weight deacy loss is 6.98, training accuracy is 0.718750, time 207.987 samples/sec
+epoch 9, total_step 142160, total loss is 19.03 , inference loss is 12.04, weight deacy loss is 6.98, training accuracy is 0.375000, time 214.700 samples/sec
+epoch 9, total_step 142180, total loss is 14.15 , inference loss is 7.17, weight deacy loss is 6.98, training accuracy is 0.343750, time 218.069 samples/sec
+epoch 9, total_step 142200, total loss is 15.32 , inference loss is 8.34, weight deacy loss is 6.98, training accuracy is 0.312500, time 13.438 samples/sec
+epoch 9, total_step 142220, total loss is 16.21 , inference loss is 9.22, weight deacy loss is 6.98, training accuracy is 0.343750, time 212.629 samples/sec
+epoch 9, total_step 142240, total loss is 14.58 , inference loss is 7.60, weight deacy loss is 6.98, training accuracy is 0.500000, time 217.304 samples/sec
+epoch 9, total_step 142260, total loss is 13.64 , inference loss is 6.66, weight deacy loss is 6.98, training accuracy is 0.468750, time 210.153 samples/sec
+epoch 9, total_step 142280, total loss is 13.87 , inference loss is 6.89, weight deacy loss is 6.98, training accuracy is 0.468750, time 209.612 samples/sec
+epoch 9, total_step 142300, total loss is 12.28 , inference loss is 5.29, weight deacy loss is 6.98, training accuracy is 0.468750, time 216.122 samples/sec
+epoch 9, total_step 142320, total loss is 16.64 , inference loss is 9.65, weight deacy loss is 6.98, training accuracy is 0.343750, time 213.653 samples/sec
+epoch 9, total_step 142340, total loss is 24.89 , inference loss is 17.91, weight deacy loss is 6.98, training accuracy is 0.468750, time 225.102 samples/sec
+epoch 9, total_step 142360, total loss is 17.08 , inference loss is 10.10, weight deacy loss is 6.98, training accuracy is 0.406250, time 209.099 samples/sec
+epoch 9, total_step 142380, total loss is 15.67 , inference loss is 8.69, weight deacy loss is 6.98, training accuracy is 0.437500, time 221.649 samples/sec
+epoch 9, total_step 142400, total loss is 18.10 , inference loss is 11.12, weight deacy loss is 6.98, training accuracy is 0.437500, time 215.530 samples/sec
+epoch 9, total_step 142420, total loss is 15.98 , inference loss is 9.00, weight deacy loss is 6.98, training accuracy is 0.468750, time 212.985 samples/sec
+epoch 9, total_step 142440, total loss is 18.32 , inference loss is 11.34, weight deacy loss is 6.98, training accuracy is 0.343750, time 208.303 samples/sec
+epoch 9, total_step 142460, total loss is 16.11 , inference loss is 9.13, weight deacy loss is 6.98, training accuracy is 0.437500, time 217.519 samples/sec
+epoch 9, total_step 142480, total loss is 15.38 , inference loss is 8.40, weight deacy loss is 6.98, training accuracy is 0.312500, time 213.848 samples/sec
+epoch 9, total_step 142500, total loss is 13.31 , inference loss is 6.33, weight deacy loss is 6.98, training accuracy is 0.468750, time 13.943 samples/sec
+epoch 9, total_step 142520, total loss is 17.49 , inference loss is 10.51, weight deacy loss is 6.98, training accuracy is 0.437500, time 216.302 samples/sec
+epoch 9, total_step 142540, total loss is 11.99 , inference loss is 5.02, weight deacy loss is 6.98, training accuracy is 0.593750, time 209.363 samples/sec
+epoch 9, total_step 142560, total loss is 15.10 , inference loss is 8.12, weight deacy loss is 6.98, training accuracy is 0.500000, time 220.935 samples/sec
+epoch 9, total_step 142580, total loss is 17.95 , inference loss is 10.97, weight deacy loss is 6.98, training accuracy is 0.375000, time 214.912 samples/sec
+epoch 9, total_step 142600, total loss is 15.89 , inference loss is 8.91, weight deacy loss is 6.98, training accuracy is 0.437500, time 222.962 samples/sec
+epoch 9, total_step 142620, total loss is 15.96 , inference loss is 8.98, weight deacy loss is 6.98, training accuracy is 0.406250, time 224.307 samples/sec
+epoch 9, total_step 142640, total loss is 12.39 , inference loss is 5.41, weight deacy loss is 6.98, training accuracy is 0.500000, time 222.265 samples/sec
+epoch 9, total_step 142660, total loss is 14.22 , inference loss is 7.25, weight deacy loss is 6.98, training accuracy is 0.468750, time 212.005 samples/sec
+epoch 9, total_step 142680, total loss is 13.54 , inference loss is 6.56, weight deacy loss is 6.98, training accuracy is 0.500000, time 221.271 samples/sec
+epoch 9, total_step 142700, total loss is 14.04 , inference loss is 7.07, weight deacy loss is 6.98, training accuracy is 0.468750, time 217.291 samples/sec
+epoch 9, total_step 142720, total loss is 16.20 , inference loss is 9.22, weight deacy loss is 6.97, training accuracy is 0.562500, time 237.915 samples/sec
+epoch 9, total_step 142740, total loss is 13.69 , inference loss is 6.71, weight deacy loss is 6.97, training accuracy is 0.406250, time 215.173 samples/sec
+epoch 9, total_step 142760, total loss is 14.85 , inference loss is 7.87, weight deacy loss is 6.97, training accuracy is 0.375000, time 219.074 samples/sec
+epoch 9, total_step 142780, total loss is 19.15 , inference loss is 12.18, weight deacy loss is 6.97, training accuracy is 0.375000, time 207.897 samples/sec
+epoch 9, total_step 142800, total loss is 16.87 , inference loss is 9.90, weight deacy loss is 6.97, training accuracy is 0.437500, time 12.925 samples/sec
+epoch 9, total_step 142820, total loss is 12.34 , inference loss is 5.36, weight deacy loss is 6.97, training accuracy is 0.500000, time 212.002 samples/sec
+epoch 9, total_step 142840, total loss is 13.21 , inference loss is 6.24, weight deacy loss is 6.97, training accuracy is 0.562500, time 216.213 samples/sec
+epoch 9, total_step 142860, total loss is 12.27 , inference loss is 5.30, weight deacy loss is 6.97, training accuracy is 0.437500, time 213.424 samples/sec
+epoch 9, total_step 142880, total loss is 15.91 , inference loss is 8.93, weight deacy loss is 6.97, training accuracy is 0.531250, time 210.216 samples/sec
+epoch 9, total_step 142900, total loss is 14.49 , inference loss is 7.52, weight deacy loss is 6.97, training accuracy is 0.468750, time 213.292 samples/sec
+epoch 9, total_step 142920, total loss is 16.55 , inference loss is 9.57, weight deacy loss is 6.97, training accuracy is 0.468750, time 214.350 samples/sec
+epoch 9, total_step 142940, total loss is 16.07 , inference loss is 9.10, weight deacy loss is 6.97, training accuracy is 0.437500, time 214.452 samples/sec
+epoch 9, total_step 142960, total loss is 15.56 , inference loss is 8.59, weight deacy loss is 6.97, training accuracy is 0.437500, time 210.508 samples/sec
+epoch 9, total_step 142980, total loss is 17.17 , inference loss is 10.20, weight deacy loss is 6.97, training accuracy is 0.312500, time 214.228 samples/sec
+epoch 9, total_step 143000, total loss is 14.77 , inference loss is 7.79, weight deacy loss is 6.97, training accuracy is 0.406250, time 225.179 samples/sec
+epoch 9, total_step 143020, total loss is 15.45 , inference loss is 8.48, weight deacy loss is 6.97, training accuracy is 0.531250, time 219.587 samples/sec
+epoch 9, total_step 143040, total loss is 13.06 , inference loss is 6.09, weight deacy loss is 6.97, training accuracy is 0.562500, time 215.954 samples/sec
+epoch 9, total_step 143060, total loss is 16.47 , inference loss is 9.50, weight deacy loss is 6.97, training accuracy is 0.437500, time 218.622 samples/sec
+epoch 9, total_step 143080, total loss is 15.11 , inference loss is 8.14, weight deacy loss is 6.97, training accuracy is 0.437500, time 298.801 samples/sec
+epoch 9, total_step 143100, total loss is 12.99 , inference loss is 6.02, weight deacy loss is 6.97, training accuracy is 0.500000, time 13.241 samples/sec
+epoch 9, total_step 143120, total loss is 13.60 , inference loss is 6.63, weight deacy loss is 6.97, training accuracy is 0.468750, time 213.057 samples/sec
+epoch 9, total_step 143140, total loss is 15.47 , inference loss is 8.50, weight deacy loss is 6.97, training accuracy is 0.500000, time 229.080 samples/sec
+epoch 9, total_step 143160, total loss is 14.40 , inference loss is 7.44, weight deacy loss is 6.97, training accuracy is 0.437500, time 224.374 samples/sec
+epoch 9, total_step 143180, total loss is 15.75 , inference loss is 8.78, weight deacy loss is 6.97, training accuracy is 0.281250, time 213.343 samples/sec
+epoch 9, total_step 143200, total loss is 14.36 , inference loss is 7.40, weight deacy loss is 6.97, training accuracy is 0.531250, time 221.056 samples/sec
+epoch 9, total_step 143220, total loss is 17.73 , inference loss is 10.76, weight deacy loss is 6.97, training accuracy is 0.437500, time 215.918 samples/sec
+epoch 9, total_step 143240, total loss is 19.52 , inference loss is 12.56, weight deacy loss is 6.97, training accuracy is 0.343750, time 211.574 samples/sec
+epoch 9, total_step 143260, total loss is 12.64 , inference loss is 5.67, weight deacy loss is 6.97, training accuracy is 0.562500, time 209.862 samples/sec
+epoch 9, total_step 143280, total loss is 14.91 , inference loss is 7.94, weight deacy loss is 6.97, training accuracy is 0.312500, time 214.386 samples/sec
+epoch 9, total_step 143300, total loss is 14.27 , inference loss is 7.31, weight deacy loss is 6.97, training accuracy is 0.531250, time 210.374 samples/sec
+epoch 9, total_step 143320, total loss is 12.36 , inference loss is 5.40, weight deacy loss is 6.97, training accuracy is 0.500000, time 216.782 samples/sec
+epoch 9, total_step 143340, total loss is 15.46 , inference loss is 8.49, weight deacy loss is 6.97, training accuracy is 0.468750, time 297.857 samples/sec
+epoch 9, total_step 143360, total loss is 18.30 , inference loss is 11.33, weight deacy loss is 6.96, training accuracy is 0.343750, time 213.582 samples/sec
+epoch 9, total_step 143380, total loss is 10.60 , inference loss is 3.63, weight deacy loss is 6.96, training accuracy is 0.656250, time 213.967 samples/sec
+epoch 9, total_step 143400, total loss is 12.07 , inference loss is 5.10, weight deacy loss is 6.96, training accuracy is 0.625000, time 12.856 samples/sec
+epoch 9, total_step 143420, total loss is 14.00 , inference loss is 7.03, weight deacy loss is 6.96, training accuracy is 0.468750, time 218.475 samples/sec
+epoch 9, total_step 143440, total loss is 12.43 , inference loss is 5.47, weight deacy loss is 6.96, training accuracy is 0.562500, time 215.823 samples/sec
+epoch 9, total_step 143460, total loss is 13.99 , inference loss is 7.03, weight deacy loss is 6.96, training accuracy is 0.531250, time 219.830 samples/sec
+epoch 9, total_step 143480, total loss is 11.31 , inference loss is 4.35, weight deacy loss is 6.96, training accuracy is 0.500000, time 215.842 samples/sec
+epoch 9, total_step 143500, total loss is 15.49 , inference loss is 8.53, weight deacy loss is 6.96, training accuracy is 0.343750, time 213.177 samples/sec
+epoch 9, total_step 143520, total loss is 20.42 , inference loss is 13.46, weight deacy loss is 6.96, training accuracy is 0.312500, time 218.928 samples/sec
+epoch 9, total_step 143540, total loss is 17.36 , inference loss is 10.40, weight deacy loss is 6.96, training accuracy is 0.468750, time 216.619 samples/sec
+epoch 9, total_step 143560, total loss is 15.83 , inference loss is 8.86, weight deacy loss is 6.96, training accuracy is 0.468750, time 300.265 samples/sec
+epoch 9, total_step 143580, total loss is 13.95 , inference loss is 6.99, weight deacy loss is 6.96, training accuracy is 0.468750, time 298.250 samples/sec
+epoch 9, total_step 143600, total loss is 11.57 , inference loss is 4.61, weight deacy loss is 6.96, training accuracy is 0.437500, time 207.493 samples/sec
+epoch 9, total_step 143620, total loss is 14.21 , inference loss is 7.25, weight deacy loss is 6.96, training accuracy is 0.562500, time 213.758 samples/sec
+epoch 9, total_step 143640, total loss is 14.92 , inference loss is 7.96, weight deacy loss is 6.96, training accuracy is 0.468750, time 214.854 samples/sec
+epoch 9, total_step 143660, total loss is 12.63 , inference loss is 5.67, weight deacy loss is 6.96, training accuracy is 0.468750, time 213.781 samples/sec
+epoch 9, total_step 143680, total loss is 12.32 , inference loss is 5.36, weight deacy loss is 6.96, training accuracy is 0.593750, time 212.716 samples/sec
+epoch 9, total_step 143700, total loss is 11.76 , inference loss is 4.80, weight deacy loss is 6.96, training accuracy is 0.593750, time 13.143 samples/sec
+epoch 9, total_step 143720, total loss is 19.31 , inference loss is 12.35, weight deacy loss is 6.96, training accuracy is 0.312500, time 207.958 samples/sec
+epoch 9, total_step 143740, total loss is 15.97 , inference loss is 9.01, weight deacy loss is 6.96, training accuracy is 0.437500, time 296.710 samples/sec
+epoch 9, total_step 143760, total loss is 12.97 , inference loss is 6.02, weight deacy loss is 6.96, training accuracy is 0.437500, time 216.243 samples/sec
+epoch 9, total_step 143780, total loss is 16.65 , inference loss is 9.69, weight deacy loss is 6.96, training accuracy is 0.343750, time 213.607 samples/sec
+epoch 9, total_step 143800, total loss is 15.01 , inference loss is 8.05, weight deacy loss is 6.96, training accuracy is 0.437500, time 218.864 samples/sec
+epoch 9, total_step 143820, total loss is 18.29 , inference loss is 11.33, weight deacy loss is 6.96, training accuracy is 0.281250, time 214.475 samples/sec
+epoch 9, total_step 143840, total loss is 15.76 , inference loss is 8.80, weight deacy loss is 6.96, training accuracy is 0.500000, time 214.340 samples/sec
+epoch 9, total_step 143860, total loss is 13.48 , inference loss is 6.52, weight deacy loss is 6.96, training accuracy is 0.500000, time 211.034 samples/sec
+epoch 9, total_step 143880, total loss is 13.19 , inference loss is 6.23, weight deacy loss is 6.96, training accuracy is 0.562500, time 214.711 samples/sec
+epoch 9, total_step 143900, total loss is 13.17 , inference loss is 6.22, weight deacy loss is 6.96, training accuracy is 0.500000, time 207.465 samples/sec
+epoch 9, total_step 143920, total loss is 18.18 , inference loss is 11.22, weight deacy loss is 6.95, training accuracy is 0.343750, time 224.141 samples/sec
+epoch 9, total_step 143940, total loss is 12.79 , inference loss is 5.83, weight deacy loss is 6.95, training accuracy is 0.437500, time 216.173 samples/sec
+epoch 9, total_step 143960, total loss is 19.07 , inference loss is 12.12, weight deacy loss is 6.95, training accuracy is 0.500000, time 216.086 samples/sec
+epoch 9, total_step 143980, total loss is 16.73 , inference loss is 9.77, weight deacy loss is 6.95, training accuracy is 0.281250, time 216.537 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.473917999999973
+best_threshold_index 114 0.9888888888888889
+best_threshold_index 114 0.9875925925925926
+best_threshold_index 117 0.9881481481481481
+best_threshold_index 114 0.9887037037037038
+best_threshold_index 117 0.9888888888888889
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 114 0.9885185185185185
+best_threshold_index 113 0.9883333333333333
+best_threshold_index 114 0.9866666666666667
+best_threshold_index 117 0.9874074074074074
+[lfw][144000]XNorm: 22.618802
+[lfw][144000]Accuracy-Flip: 0.98717+-0.00637
+test accuracy is: 0.9871666666666667
+epoch 9, total_step 144000, total loss is 13.94 , inference loss is 6.98, weight deacy loss is 6.95, training accuracy is 0.406250, time 17.685 samples/sec
+epoch 9, total_step 144020, total loss is 15.37 , inference loss is 8.41, weight deacy loss is 6.95, training accuracy is 0.500000, time 220.546 samples/sec
+epoch 9, total_step 144040, total loss is 15.56 , inference loss is 8.61, weight deacy loss is 6.95, training accuracy is 0.375000, time 213.657 samples/sec
+epoch 9, total_step 144060, total loss is 14.50 , inference loss is 7.55, weight deacy loss is 6.95, training accuracy is 0.437500, time 224.629 samples/sec
+epoch 9, total_step 144080, total loss is 13.28 , inference loss is 6.32, weight deacy loss is 6.95, training accuracy is 0.375000, time 211.398 samples/sec
+epoch 9, total_step 144100, total loss is 13.35 , inference loss is 6.40, weight deacy loss is 6.95, training accuracy is 0.406250, time 215.665 samples/sec
+epoch 9, total_step 144120, total loss is 13.77 , inference loss is 6.82, weight deacy loss is 6.95, training accuracy is 0.406250, time 297.014 samples/sec
+epoch 9, total_step 144140, total loss is 17.00 , inference loss is 10.05, weight deacy loss is 6.95, training accuracy is 0.406250, time 221.413 samples/sec
+epoch 9, total_step 144160, total loss is 15.02 , inference loss is 8.07, weight deacy loss is 6.95, training accuracy is 0.406250, time 225.943 samples/sec
+epoch 9, total_step 144180, total loss is 12.81 , inference loss is 5.86, weight deacy loss is 6.95, training accuracy is 0.437500, time 209.391 samples/sec
+epoch 9, total_step 144200, total loss is 17.95 , inference loss is 11.00, weight deacy loss is 6.95, training accuracy is 0.375000, time 221.540 samples/sec
+epoch 9, total_step 144220, total loss is 20.03 , inference loss is 13.08, weight deacy loss is 6.95, training accuracy is 0.187500, time 218.401 samples/sec
+epoch 9, total_step 144240, total loss is 12.08 , inference loss is 5.13, weight deacy loss is 6.95, training accuracy is 0.531250, time 222.382 samples/sec
+epoch 9, total_step 144260, total loss is 17.43 , inference loss is 10.48, weight deacy loss is 6.95, training accuracy is 0.437500, time 216.324 samples/sec
+epoch 9, total_step 144280, total loss is 16.36 , inference loss is 9.41, weight deacy loss is 6.95, training accuracy is 0.500000, time 231.334 samples/sec
+epoch 9, total_step 144300, total loss is 12.68 , inference loss is 5.73, weight deacy loss is 6.95, training accuracy is 0.562500, time 13.321 samples/sec
+epoch 9, total_step 144320, total loss is 16.64 , inference loss is 9.70, weight deacy loss is 6.95, training accuracy is 0.218750, time 222.427 samples/sec
+epoch 9, total_step 144340, total loss is 13.08 , inference loss is 6.14, weight deacy loss is 6.95, training accuracy is 0.406250, time 214.251 samples/sec
+epoch 9, total_step 144360, total loss is 17.87 , inference loss is 10.92, weight deacy loss is 6.95, training accuracy is 0.250000, time 228.228 samples/sec
+epoch 9, total_step 144380, total loss is 18.25 , inference loss is 11.31, weight deacy loss is 6.95, training accuracy is 0.468750, time 298.431 samples/sec
+epoch 9, total_step 144400, total loss is 17.13 , inference loss is 10.18, weight deacy loss is 6.95, training accuracy is 0.375000, time 216.453 samples/sec
+epoch 9, total_step 144420, total loss is 19.20 , inference loss is 12.25, weight deacy loss is 6.95, training accuracy is 0.218750, time 208.380 samples/sec
+epoch 9, total_step 144440, total loss is 21.11 , inference loss is 14.16, weight deacy loss is 6.95, training accuracy is 0.218750, time 215.177 samples/sec
+epoch 9, total_step 144460, total loss is 14.28 , inference loss is 7.34, weight deacy loss is 6.95, training accuracy is 0.437500, time 217.499 samples/sec
+epoch 9, total_step 144480, total loss is 17.66 , inference loss is 10.72, weight deacy loss is 6.95, training accuracy is 0.343750, time 234.980 samples/sec
+epoch 9, total_step 144500, total loss is 13.23 , inference loss is 6.29, weight deacy loss is 6.95, training accuracy is 0.406250, time 216.970 samples/sec
+epoch 9, total_step 144520, total loss is 18.14 , inference loss is 11.20, weight deacy loss is 6.95, training accuracy is 0.312500, time 218.002 samples/sec
+epoch 9, total_step 144540, total loss is 12.88 , inference loss is 5.94, weight deacy loss is 6.94, training accuracy is 0.406250, time 227.333 samples/sec
+epoch 9, total_step 144560, total loss is 17.18 , inference loss is 10.24, weight deacy loss is 6.94, training accuracy is 0.406250, time 224.852 samples/sec
+epoch 9, total_step 144580, total loss is 12.85 , inference loss is 5.91, weight deacy loss is 6.94, training accuracy is 0.406250, time 212.083 samples/sec
+epoch 9, total_step 144600, total loss is 12.55 , inference loss is 5.61, weight deacy loss is 6.94, training accuracy is 0.406250, time 12.763 samples/sec
+epoch 9, total_step 144620, total loss is 16.40 , inference loss is 9.45, weight deacy loss is 6.94, training accuracy is 0.437500, time 205.889 samples/sec
+epoch 9, total_step 144640, total loss is 18.70 , inference loss is 11.76, weight deacy loss is 6.94, training accuracy is 0.343750, time 298.931 samples/sec
+epoch 9, total_step 144660, total loss is 13.84 , inference loss is 6.90, weight deacy loss is 6.94, training accuracy is 0.531250, time 214.495 samples/sec
+epoch 9, total_step 144680, total loss is 16.13 , inference loss is 9.18, weight deacy loss is 6.94, training accuracy is 0.375000, time 224.633 samples/sec
+epoch 9, total_step 144700, total loss is 16.07 , inference loss is 9.12, weight deacy loss is 6.94, training accuracy is 0.406250, time 216.353 samples/sec
+epoch 9, total_step 144720, total loss is 12.26 , inference loss is 5.32, weight deacy loss is 6.94, training accuracy is 0.468750, time 216.038 samples/sec
+epoch 9, total_step 144740, total loss is 11.93 , inference loss is 4.99, weight deacy loss is 6.94, training accuracy is 0.437500, time 221.936 samples/sec
+epoch 9, total_step 144760, total loss is 16.56 , inference loss is 9.62, weight deacy loss is 6.94, training accuracy is 0.312500, time 219.668 samples/sec
+epoch 9, total_step 144780, total loss is 19.36 , inference loss is 12.42, weight deacy loss is 6.94, training accuracy is 0.218750, time 218.738 samples/sec
+epoch 9, total_step 144800, total loss is 15.25 , inference loss is 8.31, weight deacy loss is 6.94, training accuracy is 0.468750, time 210.032 samples/sec
+epoch 9, total_step 144820, total loss is 19.51 , inference loss is 12.57, weight deacy loss is 6.94, training accuracy is 0.375000, time 209.436 samples/sec
+epoch 9, total_step 144840, total loss is 17.52 , inference loss is 10.58, weight deacy loss is 6.94, training accuracy is 0.312500, time 214.763 samples/sec
+epoch 9, total_step 144860, total loss is 13.34 , inference loss is 6.40, weight deacy loss is 6.94, training accuracy is 0.468750, time 212.193 samples/sec
+epoch 9, total_step 144880, total loss is 19.40 , inference loss is 12.46, weight deacy loss is 6.94, training accuracy is 0.437500, time 216.314 samples/sec
+epoch 9, total_step 144900, total loss is 17.60 , inference loss is 10.66, weight deacy loss is 6.94, training accuracy is 0.250000, time 12.684 samples/sec
+epoch 9, total_step 144920, total loss is 12.38 , inference loss is 5.44, weight deacy loss is 6.94, training accuracy is 0.500000, time 210.829 samples/sec
+epoch 9, total_step 144940, total loss is 11.34 , inference loss is 4.41, weight deacy loss is 6.94, training accuracy is 0.468750, time 210.114 samples/sec
+epoch 9, total_step 144960, total loss is 19.78 , inference loss is 12.84, weight deacy loss is 6.94, training accuracy is 0.343750, time 221.525 samples/sec
+epoch 9, total_step 144980, total loss is 19.11 , inference loss is 12.17, weight deacy loss is 6.94, training accuracy is 0.375000, time 243.524 samples/sec
+epoch 9, total_step 145000, total loss is 20.78 , inference loss is 13.84, weight deacy loss is 6.94, training accuracy is 0.218750, time 214.625 samples/sec
+epoch 9, total_step 145020, total loss is 13.40 , inference loss is 6.47, weight deacy loss is 6.94, training accuracy is 0.468750, time 219.591 samples/sec
+epoch 9, total_step 145040, total loss is 14.48 , inference loss is 7.55, weight deacy loss is 6.94, training accuracy is 0.468750, time 222.841 samples/sec
+epoch 9, total_step 145060, total loss is 12.00 , inference loss is 5.06, weight deacy loss is 6.94, training accuracy is 0.562500, time 213.681 samples/sec
+epoch 9, total_step 145080, total loss is 13.23 , inference loss is 6.30, weight deacy loss is 6.94, training accuracy is 0.500000, time 214.519 samples/sec
+epoch 9, total_step 145100, total loss is 14.65 , inference loss is 7.72, weight deacy loss is 6.94, training accuracy is 0.468750, time 216.245 samples/sec
+epoch 9, total_step 145120, total loss is 15.95 , inference loss is 9.01, weight deacy loss is 6.93, training accuracy is 0.281250, time 217.071 samples/sec
+epoch 9, total_step 145140, total loss is 14.45 , inference loss is 7.52, weight deacy loss is 6.93, training accuracy is 0.500000, time 218.279 samples/sec
+epoch 9, total_step 145160, total loss is 14.08 , inference loss is 7.15, weight deacy loss is 6.93, training accuracy is 0.468750, time 223.963 samples/sec
+epoch 9, total_step 145180, total loss is 12.64 , inference loss is 5.71, weight deacy loss is 6.93, training accuracy is 0.531250, time 211.341 samples/sec
+epoch 9, total_step 145200, total loss is 14.72 , inference loss is 7.78, weight deacy loss is 6.93, training accuracy is 0.406250, time 13.854 samples/sec
+epoch 9, total_step 145220, total loss is 15.89 , inference loss is 8.95, weight deacy loss is 6.93, training accuracy is 0.437500, time 218.324 samples/sec
+epoch 9, total_step 145240, total loss is 16.07 , inference loss is 9.13, weight deacy loss is 6.93, training accuracy is 0.375000, time 212.643 samples/sec
+epoch 9, total_step 145260, total loss is 16.99 , inference loss is 10.06, weight deacy loss is 6.93, training accuracy is 0.312500, time 221.004 samples/sec
+epoch 9, total_step 145280, total loss is 14.48 , inference loss is 7.55, weight deacy loss is 6.93, training accuracy is 0.375000, time 216.940 samples/sec
+epoch 9, total_step 145300, total loss is 15.61 , inference loss is 8.68, weight deacy loss is 6.93, training accuracy is 0.312500, time 216.372 samples/sec
+epoch 9, total_step 145320, total loss is 19.13 , inference loss is 12.20, weight deacy loss is 6.93, training accuracy is 0.343750, time 248.988 samples/sec
+epoch 9, total_step 145340, total loss is 16.35 , inference loss is 9.42, weight deacy loss is 6.93, training accuracy is 0.406250, time 298.766 samples/sec
+epoch 9, total_step 145360, total loss is 16.99 , inference loss is 10.06, weight deacy loss is 6.93, training accuracy is 0.375000, time 216.987 samples/sec
+epoch 9, total_step 145380, total loss is 13.28 , inference loss is 6.35, weight deacy loss is 6.93, training accuracy is 0.500000, time 207.289 samples/sec
+epoch 9, total_step 145400, total loss is 12.60 , inference loss is 5.67, weight deacy loss is 6.93, training accuracy is 0.531250, time 217.999 samples/sec
+epoch 9, total_step 145420, total loss is 16.96 , inference loss is 10.03, weight deacy loss is 6.93, training accuracy is 0.375000, time 212.261 samples/sec
+epoch 9, total_step 145440, total loss is 17.01 , inference loss is 10.09, weight deacy loss is 6.93, training accuracy is 0.500000, time 210.119 samples/sec
+epoch 9, total_step 145460, total loss is 16.06 , inference loss is 9.13, weight deacy loss is 6.93, training accuracy is 0.406250, time 212.304 samples/sec
+epoch 9, total_step 145480, total loss is 15.60 , inference loss is 8.68, weight deacy loss is 6.93, training accuracy is 0.375000, time 225.309 samples/sec
+epoch 9, total_step 145500, total loss is 14.57 , inference loss is 7.64, weight deacy loss is 6.93, training accuracy is 0.531250, time 13.257 samples/sec
+epoch 9, total_step 145520, total loss is 13.76 , inference loss is 6.83, weight deacy loss is 6.93, training accuracy is 0.500000, time 211.893 samples/sec
+epoch 9, total_step 145540, total loss is 11.13 , inference loss is 4.20, weight deacy loss is 6.93, training accuracy is 0.593750, time 233.139 samples/sec
+epoch 9, total_step 145560, total loss is 15.68 , inference loss is 8.76, weight deacy loss is 6.93, training accuracy is 0.468750, time 215.487 samples/sec
+epoch 9, total_step 145580, total loss is 12.23 , inference loss is 5.31, weight deacy loss is 6.93, training accuracy is 0.437500, time 215.703 samples/sec
+epoch 9, total_step 145600, total loss is 13.33 , inference loss is 6.40, weight deacy loss is 6.93, training accuracy is 0.406250, time 223.987 samples/sec
+epoch 9, total_step 145620, total loss is 16.01 , inference loss is 9.08, weight deacy loss is 6.93, training accuracy is 0.375000, time 192.297 samples/sec
+epoch 9, total_step 145640, total loss is 13.88 , inference loss is 6.96, weight deacy loss is 6.93, training accuracy is 0.468750, time 219.572 samples/sec
+epoch 9, total_step 145660, total loss is 13.97 , inference loss is 7.04, weight deacy loss is 6.92, training accuracy is 0.406250, time 213.320 samples/sec
+epoch 9, total_step 145680, total loss is 16.11 , inference loss is 9.19, weight deacy loss is 6.92, training accuracy is 0.437500, time 213.406 samples/sec
+epoch 9, total_step 145700, total loss is 16.07 , inference loss is 9.14, weight deacy loss is 6.92, training accuracy is 0.531250, time 211.018 samples/sec
+epoch 9, total_step 145720, total loss is 17.20 , inference loss is 10.28, weight deacy loss is 6.92, training accuracy is 0.312500, time 208.663 samples/sec
+epoch 9, total_step 145740, total loss is 12.61 , inference loss is 5.69, weight deacy loss is 6.92, training accuracy is 0.500000, time 217.731 samples/sec
+epoch 9, total_step 145760, total loss is 14.73 , inference loss is 7.81, weight deacy loss is 6.92, training accuracy is 0.343750, time 219.818 samples/sec
+epoch 9, total_step 145780, total loss is 12.13 , inference loss is 5.21, weight deacy loss is 6.92, training accuracy is 0.437500, time 213.850 samples/sec
+epoch 9, total_step 145800, total loss is 18.89 , inference loss is 11.97, weight deacy loss is 6.92, training accuracy is 0.375000, time 13.005 samples/sec
+epoch 9, total_step 145820, total loss is 14.96 , inference loss is 8.03, weight deacy loss is 6.92, training accuracy is 0.468750, time 216.039 samples/sec
+epoch 9, total_step 145840, total loss is 12.80 , inference loss is 5.88, weight deacy loss is 6.92, training accuracy is 0.468750, time 211.540 samples/sec
+epoch 9, total_step 145860, total loss is 12.25 , inference loss is 5.33, weight deacy loss is 6.92, training accuracy is 0.406250, time 211.207 samples/sec
+epoch 9, total_step 145880, total loss is 14.78 , inference loss is 7.86, weight deacy loss is 6.92, training accuracy is 0.468750, time 297.884 samples/sec
+epoch 9, total_step 145900, total loss is 14.78 , inference loss is 7.86, weight deacy loss is 6.92, training accuracy is 0.312500, time 234.588 samples/sec
+epoch 9, total_step 145920, total loss is 13.43 , inference loss is 6.51, weight deacy loss is 6.92, training accuracy is 0.593750, time 213.782 samples/sec
+epoch 9, total_step 145940, total loss is 14.12 , inference loss is 7.20, weight deacy loss is 6.92, training accuracy is 0.531250, time 211.786 samples/sec
+epoch 9, total_step 145960, total loss is 14.77 , inference loss is 7.85, weight deacy loss is 6.92, training accuracy is 0.500000, time 212.533 samples/sec
+epoch 9, total_step 145980, total loss is 17.70 , inference loss is 10.78, weight deacy loss is 6.92, training accuracy is 0.312500, time 206.880 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.14554600000001
+best_threshold_index 114 0.9898148148148148
+best_threshold_index 114 0.9877777777777778
+best_threshold_index 114 0.9887037037037038
+best_threshold_index 114 0.99
+best_threshold_index 114 0.9892592592592593
+best_threshold_index 114 0.9875925925925926
+best_threshold_index 114 0.9887037037037038
+best_threshold_index 114 0.9877777777777778
+best_threshold_index 114 0.9877777777777778
+best_threshold_index 114 0.9875925925925926
+[lfw][146000]XNorm: 21.526729
+[lfw][146000]Accuracy-Flip: 0.98850+-0.00797
+test accuracy is: 0.9884999999999999
+epoch 9, total_step 146000, total loss is 13.36 , inference loss is 6.45, weight deacy loss is 6.92, training accuracy is 0.406250, time 17.477 samples/sec
+epoch 9, total_step 146020, total loss is 16.22 , inference loss is 9.31, weight deacy loss is 6.92, training accuracy is 0.281250, time 212.897 samples/sec
+epoch 9, total_step 146040, total loss is 13.62 , inference loss is 6.71, weight deacy loss is 6.92, training accuracy is 0.406250, time 218.137 samples/sec
+epoch 9, total_step 146060, total loss is 15.21 , inference loss is 8.30, weight deacy loss is 6.92, training accuracy is 0.468750, time 221.203 samples/sec
+epoch 9, total_step 146080, total loss is 18.42 , inference loss is 11.50, weight deacy loss is 6.92, training accuracy is 0.343750, time 215.192 samples/sec
+epoch 9, total_step 146100, total loss is 13.73 , inference loss is 6.82, weight deacy loss is 6.92, training accuracy is 0.468750, time 12.614 samples/sec
+epoch 9, total_step 146120, total loss is 14.44 , inference loss is 7.52, weight deacy loss is 6.92, training accuracy is 0.375000, time 218.389 samples/sec
+epoch 9, total_step 146140, total loss is 17.14 , inference loss is 10.22, weight deacy loss is 6.92, training accuracy is 0.406250, time 218.959 samples/sec
+epoch 9, total_step 146160, total loss is 16.05 , inference loss is 9.13, weight deacy loss is 6.92, training accuracy is 0.312500, time 213.736 samples/sec
+epoch 9, total_step 146180, total loss is 11.45 , inference loss is 4.54, weight deacy loss is 6.92, training accuracy is 0.562500, time 219.149 samples/sec
+epoch 9, total_step 146200, total loss is 14.45 , inference loss is 7.54, weight deacy loss is 6.92, training accuracy is 0.500000, time 217.529 samples/sec
+epoch 9, total_step 146220, total loss is 16.59 , inference loss is 9.67, weight deacy loss is 6.92, training accuracy is 0.343750, time 213.916 samples/sec
+epoch 9, total_step 146240, total loss is 14.06 , inference loss is 7.14, weight deacy loss is 6.91, training accuracy is 0.500000, time 297.714 samples/sec
+epoch 9, total_step 146260, total loss is 17.12 , inference loss is 10.21, weight deacy loss is 6.91, training accuracy is 0.343750, time 218.399 samples/sec
+epoch 9, total_step 146280, total loss is 13.15 , inference loss is 6.24, weight deacy loss is 6.91, training accuracy is 0.531250, time 214.655 samples/sec
+epoch 9, total_step 146300, total loss is 18.72 , inference loss is 11.81, weight deacy loss is 6.91, training accuracy is 0.281250, time 212.641 samples/sec
+epoch 9, total_step 146320, total loss is 13.74 , inference loss is 6.83, weight deacy loss is 6.91, training accuracy is 0.375000, time 213.698 samples/sec
+epoch 9, total_step 146340, total loss is 15.39 , inference loss is 8.47, weight deacy loss is 6.91, training accuracy is 0.281250, time 212.883 samples/sec
+epoch 9, total_step 146360, total loss is 16.42 , inference loss is 9.50, weight deacy loss is 6.91, training accuracy is 0.437500, time 205.845 samples/sec
+epoch 9, total_step 146380, total loss is 16.42 , inference loss is 9.51, weight deacy loss is 6.91, training accuracy is 0.406250, time 214.470 samples/sec
+epoch 9, total_step 146400, total loss is 16.41 , inference loss is 9.50, weight deacy loss is 6.91, training accuracy is 0.218750, time 12.826 samples/sec
+epoch 9, total_step 146420, total loss is 17.03 , inference loss is 10.12, weight deacy loss is 6.91, training accuracy is 0.468750, time 211.355 samples/sec
+epoch 9, total_step 146440, total loss is 13.19 , inference loss is 6.28, weight deacy loss is 6.91, training accuracy is 0.437500, time 212.198 samples/sec
+epoch 9, total_step 146460, total loss is 16.17 , inference loss is 9.26, weight deacy loss is 6.91, training accuracy is 0.500000, time 214.565 samples/sec
+epoch 9, total_step 146480, total loss is 11.78 , inference loss is 4.87, weight deacy loss is 6.91, training accuracy is 0.562500, time 220.574 samples/sec
+epoch 9, total_step 146500, total loss is 12.45 , inference loss is 5.54, weight deacy loss is 6.91, training accuracy is 0.625000, time 216.132 samples/sec
+epoch 9, total_step 146520, total loss is 16.60 , inference loss is 9.69, weight deacy loss is 6.91, training accuracy is 0.375000, time 225.317 samples/sec
+epoch 9, total_step 146540, total loss is 19.81 , inference loss is 12.90, weight deacy loss is 6.91, training accuracy is 0.468750, time 212.563 samples/sec
+epoch 9, total_step 146560, total loss is 13.70 , inference loss is 6.79, weight deacy loss is 6.91, training accuracy is 0.437500, time 214.874 samples/sec
+epoch 9, total_step 146580, total loss is 16.97 , inference loss is 10.06, weight deacy loss is 6.91, training accuracy is 0.375000, time 208.167 samples/sec
+epoch 9, total_step 146600, total loss is 17.24 , inference loss is 10.33, weight deacy loss is 6.91, training accuracy is 0.375000, time 225.357 samples/sec
+epoch 9, total_step 146620, total loss is 15.75 , inference loss is 8.84, weight deacy loss is 6.91, training accuracy is 0.562500, time 223.440 samples/sec
+epoch 9, total_step 146640, total loss is 13.98 , inference loss is 7.07, weight deacy loss is 6.91, training accuracy is 0.406250, time 214.011 samples/sec
+epoch 9, total_step 146660, total loss is 16.13 , inference loss is 9.22, weight deacy loss is 6.91, training accuracy is 0.312500, time 208.807 samples/sec
+epoch 9, total_step 146680, total loss is 20.13 , inference loss is 13.22, weight deacy loss is 6.91, training accuracy is 0.312500, time 217.264 samples/sec
+epoch 9, total_step 146700, total loss is 15.82 , inference loss is 8.91, weight deacy loss is 6.91, training accuracy is 0.468750, time 13.060 samples/sec
+epoch 9, total_step 146720, total loss is 17.50 , inference loss is 10.59, weight deacy loss is 6.91, training accuracy is 0.281250, time 218.295 samples/sec
+epoch 9, total_step 146740, total loss is 14.70 , inference loss is 7.79, weight deacy loss is 6.91, training accuracy is 0.375000, time 212.243 samples/sec
+epoch 9, total_step 146760, total loss is 14.74 , inference loss is 7.83, weight deacy loss is 6.91, training accuracy is 0.437500, time 211.530 samples/sec
+epoch 9, total_step 146780, total loss is 13.63 , inference loss is 6.72, weight deacy loss is 6.91, training accuracy is 0.312500, time 299.415 samples/sec
+epoch 9, total_step 146800, total loss is 14.98 , inference loss is 8.07, weight deacy loss is 6.91, training accuracy is 0.375000, time 213.401 samples/sec
+epoch 9, total_step 146820, total loss is 15.59 , inference loss is 8.69, weight deacy loss is 6.91, training accuracy is 0.343750, time 229.954 samples/sec
+epoch 9, total_step 146840, total loss is 17.29 , inference loss is 10.38, weight deacy loss is 6.91, training accuracy is 0.406250, time 221.814 samples/sec
+epoch 9, total_step 146860, total loss is 15.10 , inference loss is 8.20, weight deacy loss is 6.90, training accuracy is 0.406250, time 218.032 samples/sec
+epoch 9, total_step 146880, total loss is 20.20 , inference loss is 13.30, weight deacy loss is 6.90, training accuracy is 0.312500, time 205.666 samples/sec
+epoch 9, total_step 146900, total loss is 22.98 , inference loss is 16.08, weight deacy loss is 6.90, training accuracy is 0.343750, time 212.991 samples/sec
+epoch 9, total_step 146920, total loss is 15.16 , inference loss is 8.25, weight deacy loss is 6.90, training accuracy is 0.437500, time 214.261 samples/sec
+epoch 9, total_step 146940, total loss is 13.37 , inference loss is 6.47, weight deacy loss is 6.90, training accuracy is 0.531250, time 220.772 samples/sec
+epoch 9, total_step 146960, total loss is 17.41 , inference loss is 10.51, weight deacy loss is 6.90, training accuracy is 0.281250, time 222.368 samples/sec
+epoch 9, total_step 146980, total loss is 14.30 , inference loss is 7.40, weight deacy loss is 6.90, training accuracy is 0.343750, time 206.387 samples/sec
+epoch 9, total_step 147000, total loss is 10.83 , inference loss is 3.93, weight deacy loss is 6.90, training accuracy is 0.593750, time 13.497 samples/sec
+epoch 9, total_step 147020, total loss is 13.98 , inference loss is 7.08, weight deacy loss is 6.90, training accuracy is 0.375000, time 230.545 samples/sec
+epoch 9, total_step 147040, total loss is 17.91 , inference loss is 11.00, weight deacy loss is 6.90, training accuracy is 0.281250, time 217.019 samples/sec
+epoch 9, total_step 147060, total loss is 16.15 , inference loss is 9.25, weight deacy loss is 6.90, training accuracy is 0.406250, time 219.049 samples/sec
+epoch 9, total_step 147080, total loss is 15.75 , inference loss is 8.85, weight deacy loss is 6.90, training accuracy is 0.531250, time 219.533 samples/sec
+epoch 9, total_step 147100, total loss is 18.29 , inference loss is 11.39, weight deacy loss is 6.90, training accuracy is 0.312500, time 215.732 samples/sec
+epoch 9, total_step 147120, total loss is 16.63 , inference loss is 9.73, weight deacy loss is 6.90, training accuracy is 0.406250, time 298.568 samples/sec
+epoch 9, total_step 147140, total loss is 20.55 , inference loss is 13.65, weight deacy loss is 6.90, training accuracy is 0.312500, time 222.066 samples/sec
+epoch 9, total_step 147160, total loss is 15.10 , inference loss is 8.20, weight deacy loss is 6.90, training accuracy is 0.312500, time 212.487 samples/sec
+epoch 9, total_step 147180, total loss is 15.33 , inference loss is 8.43, weight deacy loss is 6.90, training accuracy is 0.500000, time 210.256 samples/sec
+epoch 9, total_step 147200, total loss is 18.07 , inference loss is 11.17, weight deacy loss is 6.90, training accuracy is 0.343750, time 209.499 samples/sec
+epoch 9, total_step 147220, total loss is 18.02 , inference loss is 11.12, weight deacy loss is 6.90, training accuracy is 0.312500, time 216.560 samples/sec
+epoch 9, total_step 147240, total loss is 14.17 , inference loss is 7.27, weight deacy loss is 6.90, training accuracy is 0.406250, time 298.459 samples/sec
+epoch 9, total_step 147260, total loss is 11.83 , inference loss is 4.93, weight deacy loss is 6.90, training accuracy is 0.625000, time 214.329 samples/sec
+epoch 9, total_step 147280, total loss is 18.51 , inference loss is 11.61, weight deacy loss is 6.90, training accuracy is 0.375000, time 216.765 samples/sec
+epoch 9, total_step 147300, total loss is 15.33 , inference loss is 8.43, weight deacy loss is 6.90, training accuracy is 0.343750, time 13.218 samples/sec
+epoch 9, total_step 147320, total loss is 16.53 , inference loss is 9.63, weight deacy loss is 6.90, training accuracy is 0.281250, time 217.423 samples/sec
+epoch 9, total_step 147340, total loss is 18.79 , inference loss is 11.90, weight deacy loss is 6.90, training accuracy is 0.312500, time 209.461 samples/sec
+epoch 9, total_step 147360, total loss is 12.21 , inference loss is 5.31, weight deacy loss is 6.90, training accuracy is 0.468750, time 218.567 samples/sec
+epoch 9, total_step 147380, total loss is 11.62 , inference loss is 4.72, weight deacy loss is 6.90, training accuracy is 0.406250, time 216.517 samples/sec
+epoch 9, total_step 147400, total loss is 13.97 , inference loss is 7.07, weight deacy loss is 6.90, training accuracy is 0.406250, time 221.687 samples/sec
+epoch 9, total_step 147420, total loss is 20.35 , inference loss is 13.45, weight deacy loss is 6.90, training accuracy is 0.437500, time 218.081 samples/sec
+epoch 9, total_step 147440, total loss is 12.59 , inference loss is 5.69, weight deacy loss is 6.90, training accuracy is 0.531250, time 214.531 samples/sec
+epoch 9, total_step 147460, total loss is 16.69 , inference loss is 9.79, weight deacy loss is 6.90, training accuracy is 0.312500, time 231.534 samples/sec
+epoch 9, total_step 147480, total loss is 13.36 , inference loss is 6.47, weight deacy loss is 6.90, training accuracy is 0.500000, time 217.028 samples/sec
+epoch 9, total_step 147500, total loss is 15.01 , inference loss is 8.12, weight deacy loss is 6.89, training accuracy is 0.468750, time 299.623 samples/sec
+epoch 9, total_step 147520, total loss is 16.11 , inference loss is 9.22, weight deacy loss is 6.89, training accuracy is 0.406250, time 221.205 samples/sec
+epoch 9, total_step 147540, total loss is 14.98 , inference loss is 8.08, weight deacy loss is 6.89, training accuracy is 0.343750, time 222.913 samples/sec
+epoch 9, total_step 147560, total loss is 16.07 , inference loss is 9.18, weight deacy loss is 6.89, training accuracy is 0.343750, time 213.975 samples/sec
+epoch 9, total_step 147580, total loss is 18.47 , inference loss is 11.57, weight deacy loss is 6.89, training accuracy is 0.250000, time 213.423 samples/sec
+epoch 9, total_step 147600, total loss is 17.72 , inference loss is 10.82, weight deacy loss is 6.89, training accuracy is 0.500000, time 12.694 samples/sec
+epoch 9, total_step 147620, total loss is 13.32 , inference loss is 6.43, weight deacy loss is 6.89, training accuracy is 0.437500, time 210.845 samples/sec
+epoch 9, total_step 147640, total loss is 11.78 , inference loss is 4.89, weight deacy loss is 6.89, training accuracy is 0.437500, time 223.857 samples/sec
+epoch 9, total_step 147660, total loss is 18.00 , inference loss is 11.11, weight deacy loss is 6.89, training accuracy is 0.281250, time 214.726 samples/sec
+epoch 9, total_step 147680, total loss is 17.74 , inference loss is 10.85, weight deacy loss is 6.89, training accuracy is 0.406250, time 213.169 samples/sec
+epoch 9, total_step 147700, total loss is 13.80 , inference loss is 6.91, weight deacy loss is 6.89, training accuracy is 0.437500, time 211.708 samples/sec
+epoch 9, total_step 147720, total loss is 15.82 , inference loss is 8.93, weight deacy loss is 6.89, training accuracy is 0.406250, time 218.923 samples/sec
+epoch 9, total_step 147740, total loss is 14.21 , inference loss is 7.32, weight deacy loss is 6.89, training accuracy is 0.437500, time 213.222 samples/sec
+epoch 9, total_step 147760, total loss is 12.64 , inference loss is 5.75, weight deacy loss is 6.89, training accuracy is 0.500000, time 297.348 samples/sec
+epoch 9, total_step 147780, total loss is 13.41 , inference loss is 6.52, weight deacy loss is 6.89, training accuracy is 0.437500, time 207.754 samples/sec
+epoch 9, total_step 147800, total loss is 13.55 , inference loss is 6.66, weight deacy loss is 6.89, training accuracy is 0.250000, time 215.293 samples/sec
+epoch 9, total_step 147820, total loss is 12.25 , inference loss is 5.36, weight deacy loss is 6.89, training accuracy is 0.562500, time 210.844 samples/sec
+epoch 9, total_step 147840, total loss is 13.93 , inference loss is 7.05, weight deacy loss is 6.89, training accuracy is 0.500000, time 213.229 samples/sec
+epoch 9, total_step 147860, total loss is 14.07 , inference loss is 7.19, weight deacy loss is 6.89, training accuracy is 0.375000, time 215.574 samples/sec
+epoch 9, total_step 147880, total loss is 16.04 , inference loss is 9.15, weight deacy loss is 6.89, training accuracy is 0.500000, time 217.260 samples/sec
+epoch 9, total_step 147900, total loss is 12.51 , inference loss is 5.62, weight deacy loss is 6.89, training accuracy is 0.562500, time 13.146 samples/sec
+epoch 9, total_step 147920, total loss is 17.05 , inference loss is 10.16, weight deacy loss is 6.89, training accuracy is 0.312500, time 218.403 samples/sec
+epoch 9, total_step 147940, total loss is 14.93 , inference loss is 8.05, weight deacy loss is 6.89, training accuracy is 0.187500, time 210.123 samples/sec
+epoch 9, total_step 147960, total loss is 12.73 , inference loss is 5.84, weight deacy loss is 6.89, training accuracy is 0.343750, time 217.732 samples/sec
+epoch 9, total_step 147980, total loss is 19.46 , inference loss is 12.57, weight deacy loss is 6.89, training accuracy is 0.437500, time 207.850 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.863634000000022
+best_threshold_index 107 0.9879629629629629
+best_threshold_index 107 0.9864814814814815
+best_threshold_index 107 0.9862962962962963
+best_threshold_index 107 0.9881481481481481
+best_threshold_index 107 0.9877777777777778
+best_threshold_index 110 0.9861111111111112
+best_threshold_index 110 0.987037037037037
+best_threshold_index 107 0.9861111111111112
+best_threshold_index 107 0.9853703703703703
+best_threshold_index 107 0.9857407407407407
+[lfw][148000]XNorm: 22.830338
+[lfw][148000]Accuracy-Flip: 0.98600+-0.00831
+test accuracy is: 0.986
+epoch 9, total_step 148000, total loss is 15.50 , inference loss is 8.62, weight deacy loss is 6.89, training accuracy is 0.406250, time 16.809 samples/sec
+epoch 9, total_step 148020, total loss is 13.73 , inference loss is 6.85, weight deacy loss is 6.89, training accuracy is 0.468750, time 224.312 samples/sec
+epoch 9, total_step 148040, total loss is 13.59 , inference loss is 6.70, weight deacy loss is 6.89, training accuracy is 0.656250, time 217.584 samples/sec
+epoch 9, total_step 148060, total loss is 14.44 , inference loss is 7.55, weight deacy loss is 6.89, training accuracy is 0.531250, time 225.759 samples/sec
+epoch 9, total_step 148080, total loss is 14.12 , inference loss is 7.23, weight deacy loss is 6.88, training accuracy is 0.468750, time 208.351 samples/sec
+epoch 9, total_step 148100, total loss is 14.31 , inference loss is 7.42, weight deacy loss is 6.88, training accuracy is 0.468750, time 217.842 samples/sec
+epoch 9, total_step 148120, total loss is 12.16 , inference loss is 5.28, weight deacy loss is 6.88, training accuracy is 0.562500, time 215.838 samples/sec
+epoch 9, total_step 148140, total loss is 13.46 , inference loss is 6.58, weight deacy loss is 6.88, training accuracy is 0.281250, time 299.552 samples/sec
+epoch 9, total_step 148160, total loss is 16.60 , inference loss is 9.72, weight deacy loss is 6.88, training accuracy is 0.312500, time 212.410 samples/sec
+epoch 9, total_step 148180, total loss is 16.90 , inference loss is 10.02, weight deacy loss is 6.88, training accuracy is 0.312500, time 215.180 samples/sec
+epoch 9, total_step 148200, total loss is 16.60 , inference loss is 9.72, weight deacy loss is 6.88, training accuracy is 0.562500, time 12.812 samples/sec
+epoch 9, total_step 148220, total loss is 18.85 , inference loss is 11.96, weight deacy loss is 6.88, training accuracy is 0.312500, time 227.548 samples/sec
+epoch 9, total_step 148240, total loss is 14.60 , inference loss is 7.72, weight deacy loss is 6.88, training accuracy is 0.500000, time 210.942 samples/sec
+epoch 9, total_step 148260, total loss is 14.08 , inference loss is 7.20, weight deacy loss is 6.88, training accuracy is 0.406250, time 207.017 samples/sec
+epoch 9, total_step 148280, total loss is 14.15 , inference loss is 7.27, weight deacy loss is 6.88, training accuracy is 0.531250, time 224.155 samples/sec
+epoch 9, total_step 148300, total loss is 13.58 , inference loss is 6.70, weight deacy loss is 6.88, training accuracy is 0.406250, time 215.523 samples/sec
+epoch 9, total_step 148320, total loss is 15.09 , inference loss is 8.21, weight deacy loss is 6.88, training accuracy is 0.375000, time 210.538 samples/sec
+epoch 9, total_step 148340, total loss is 11.56 , inference loss is 4.68, weight deacy loss is 6.88, training accuracy is 0.437500, time 212.257 samples/sec
+epoch 9, total_step 148360, total loss is 15.98 , inference loss is 9.10, weight deacy loss is 6.88, training accuracy is 0.343750, time 215.995 samples/sec
+epoch 9, total_step 148380, total loss is 15.22 , inference loss is 8.34, weight deacy loss is 6.88, training accuracy is 0.312500, time 214.772 samples/sec
+epoch 9, total_step 148400, total loss is 18.01 , inference loss is 11.13, weight deacy loss is 6.88, training accuracy is 0.437500, time 297.668 samples/sec
+epoch 9, total_step 148420, total loss is 11.94 , inference loss is 5.06, weight deacy loss is 6.88, training accuracy is 0.500000, time 212.999 samples/sec
+epoch 9, total_step 148440, total loss is 14.83 , inference loss is 7.95, weight deacy loss is 6.88, training accuracy is 0.468750, time 215.890 samples/sec
+epoch 9, total_step 148460, total loss is 15.93 , inference loss is 9.06, weight deacy loss is 6.88, training accuracy is 0.343750, time 210.715 samples/sec
+epoch 9, total_step 148480, total loss is 16.66 , inference loss is 9.78, weight deacy loss is 6.88, training accuracy is 0.468750, time 221.916 samples/sec
+epoch 9, total_step 148500, total loss is 19.48 , inference loss is 12.60, weight deacy loss is 6.88, training accuracy is 0.343750, time 13.493 samples/sec
+epoch 9, total_step 148520, total loss is 13.46 , inference loss is 6.58, weight deacy loss is 6.88, training accuracy is 0.437500, time 206.655 samples/sec
+epoch 9, total_step 148540, total loss is 15.06 , inference loss is 8.18, weight deacy loss is 6.88, training accuracy is 0.375000, time 223.260 samples/sec
+epoch 9, total_step 148560, total loss is 10.61 , inference loss is 3.74, weight deacy loss is 6.88, training accuracy is 0.625000, time 206.845 samples/sec
+epoch 9, total_step 148580, total loss is 19.87 , inference loss is 12.99, weight deacy loss is 6.88, training accuracy is 0.156250, time 222.298 samples/sec
+epoch 9, total_step 148600, total loss is 15.45 , inference loss is 8.57, weight deacy loss is 6.88, training accuracy is 0.281250, time 207.247 samples/sec
+epoch 9, total_step 148620, total loss is 15.14 , inference loss is 8.26, weight deacy loss is 6.88, training accuracy is 0.500000, time 215.149 samples/sec
+epoch 9, total_step 148640, total loss is 14.68 , inference loss is 7.81, weight deacy loss is 6.88, training accuracy is 0.437500, time 214.577 samples/sec
+epoch 9, total_step 148660, total loss is 16.91 , inference loss is 10.04, weight deacy loss is 6.87, training accuracy is 0.375000, time 297.350 samples/sec
+epoch 9, total_step 148680, total loss is 15.11 , inference loss is 8.24, weight deacy loss is 6.87, training accuracy is 0.468750, time 215.746 samples/sec
+epoch 9, total_step 148700, total loss is 16.83 , inference loss is 9.95, weight deacy loss is 6.87, training accuracy is 0.312500, time 214.398 samples/sec
+epoch 9, total_step 148720, total loss is 14.53 , inference loss is 7.66, weight deacy loss is 6.87, training accuracy is 0.375000, time 207.603 samples/sec
+epoch 9, total_step 148740, total loss is 14.60 , inference loss is 7.73, weight deacy loss is 6.87, training accuracy is 0.468750, time 215.790 samples/sec
+epoch 9, total_step 148760, total loss is 12.92 , inference loss is 6.05, weight deacy loss is 6.87, training accuracy is 0.531250, time 216.148 samples/sec
+epoch 9, total_step 148780, total loss is 18.35 , inference loss is 11.48, weight deacy loss is 6.87, training accuracy is 0.250000, time 217.377 samples/sec
+epoch 9, total_step 148800, total loss is 14.21 , inference loss is 7.34, weight deacy loss is 6.87, training accuracy is 0.375000, time 12.967 samples/sec
+epoch 9, total_step 148820, total loss is 14.96 , inference loss is 8.08, weight deacy loss is 6.87, training accuracy is 0.500000, time 216.211 samples/sec
+epoch 9, total_step 148840, total loss is 15.57 , inference loss is 8.69, weight deacy loss is 6.87, training accuracy is 0.343750, time 225.463 samples/sec
+epoch 9, total_step 148860, total loss is 14.67 , inference loss is 7.80, weight deacy loss is 6.87, training accuracy is 0.406250, time 297.298 samples/sec
+epoch 9, total_step 148880, total loss is 17.51 , inference loss is 10.64, weight deacy loss is 6.87, training accuracy is 0.375000, time 297.679 samples/sec
+epoch 9, total_step 148900, total loss is 15.05 , inference loss is 8.18, weight deacy loss is 6.87, training accuracy is 0.437500, time 224.826 samples/sec
+epoch 9, total_step 148920, total loss is 18.01 , inference loss is 11.14, weight deacy loss is 6.87, training accuracy is 0.281250, time 207.834 samples/sec
+epoch 9, total_step 148940, total loss is 14.51 , inference loss is 7.64, weight deacy loss is 6.87, training accuracy is 0.500000, time 218.485 samples/sec
+epoch 9, total_step 148960, total loss is 14.20 , inference loss is 7.33, weight deacy loss is 6.87, training accuracy is 0.406250, time 215.243 samples/sec
+epoch 9, total_step 148980, total loss is 18.09 , inference loss is 11.22, weight deacy loss is 6.87, training accuracy is 0.437500, time 208.741 samples/sec
+epoch 9, total_step 149000, total loss is 15.61 , inference loss is 8.74, weight deacy loss is 6.87, training accuracy is 0.468750, time 220.036 samples/sec
+epoch 9, total_step 149020, total loss is 16.68 , inference loss is 9.81, weight deacy loss is 6.87, training accuracy is 0.468750, time 212.648 samples/sec
+epoch 9, total_step 149040, total loss is 15.37 , inference loss is 8.50, weight deacy loss is 6.87, training accuracy is 0.406250, time 214.396 samples/sec
+epoch 9, total_step 149060, total loss is 12.39 , inference loss is 5.52, weight deacy loss is 6.87, training accuracy is 0.468750, time 210.544 samples/sec
+epoch 9, total_step 149080, total loss is 14.00 , inference loss is 7.13, weight deacy loss is 6.87, training accuracy is 0.343750, time 211.087 samples/sec
+epoch 9, total_step 149100, total loss is 13.37 , inference loss is 6.50, weight deacy loss is 6.87, training accuracy is 0.375000, time 13.172 samples/sec
+epoch 9, total_step 149120, total loss is 20.82 , inference loss is 13.96, weight deacy loss is 6.87, training accuracy is 0.343750, time 213.782 samples/sec
+epoch 9, total_step 149140, total loss is 11.77 , inference loss is 4.90, weight deacy loss is 6.87, training accuracy is 0.468750, time 217.526 samples/sec
+epoch 9, total_step 149160, total loss is 18.32 , inference loss is 11.45, weight deacy loss is 6.87, training accuracy is 0.375000, time 212.557 samples/sec
+epoch 9, total_step 149180, total loss is 13.99 , inference loss is 7.12, weight deacy loss is 6.87, training accuracy is 0.531250, time 218.263 samples/sec
+epoch 9, total_step 149200, total loss is 12.62 , inference loss is 5.75, weight deacy loss is 6.87, training accuracy is 0.437500, time 210.246 samples/sec
+epoch 9, total_step 149220, total loss is 16.61 , inference loss is 9.75, weight deacy loss is 6.87, training accuracy is 0.343750, time 214.043 samples/sec
+epoch 9, total_step 149240, total loss is 15.35 , inference loss is 8.48, weight deacy loss is 6.87, training accuracy is 0.468750, time 209.486 samples/sec
+epoch 9, total_step 149260, total loss is 15.09 , inference loss is 8.23, weight deacy loss is 6.87, training accuracy is 0.468750, time 208.243 samples/sec
+epoch 9, total_step 149280, total loss is 12.89 , inference loss is 6.02, weight deacy loss is 6.87, training accuracy is 0.406250, time 221.211 samples/sec
+epoch 9, total_step 149300, total loss is 11.90 , inference loss is 5.04, weight deacy loss is 6.86, training accuracy is 0.468750, time 209.793 samples/sec
+epoch 9, total_step 149320, total loss is 14.85 , inference loss is 7.98, weight deacy loss is 6.86, training accuracy is 0.375000, time 212.419 samples/sec
+epoch 9, total_step 149340, total loss is 14.46 , inference loss is 7.60, weight deacy loss is 6.86, training accuracy is 0.437500, time 211.706 samples/sec
+epoch 9, total_step 149360, total loss is 15.48 , inference loss is 8.61, weight deacy loss is 6.86, training accuracy is 0.375000, time 296.043 samples/sec
+epoch 9, total_step 149380, total loss is 13.64 , inference loss is 6.78, weight deacy loss is 6.86, training accuracy is 0.500000, time 210.862 samples/sec
+epoch 9, total_step 149400, total loss is 14.20 , inference loss is 7.34, weight deacy loss is 6.86, training accuracy is 0.406250, time 5.647 samples/sec
+epoch 9, total_step 149420, total loss is 13.62 , inference loss is 6.76, weight deacy loss is 6.86, training accuracy is 0.562500, time 211.510 samples/sec
+epoch 9, total_step 149440, total loss is 13.08 , inference loss is 6.22, weight deacy loss is 6.86, training accuracy is 0.343750, time 216.256 samples/sec
+epoch 9, total_step 149460, total loss is 15.46 , inference loss is 8.60, weight deacy loss is 6.86, training accuracy is 0.375000, time 213.319 samples/sec
+epoch 9, total_step 149480, total loss is 17.99 , inference loss is 11.12, weight deacy loss is 6.86, training accuracy is 0.406250, time 223.109 samples/sec
+epoch 9, total_step 149500, total loss is 15.62 , inference loss is 8.76, weight deacy loss is 6.86, training accuracy is 0.375000, time 232.986 samples/sec
+epoch 9, total_step 149520, total loss is 17.00 , inference loss is 10.14, weight deacy loss is 6.86, training accuracy is 0.343750, time 215.404 samples/sec
+epoch 9, total_step 149540, total loss is 14.89 , inference loss is 8.03, weight deacy loss is 6.86, training accuracy is 0.375000, time 215.192 samples/sec
+epoch 9, total_step 149560, total loss is 13.61 , inference loss is 6.75, weight deacy loss is 6.86, training accuracy is 0.343750, time 210.517 samples/sec
+epoch 9, total_step 149580, total loss is 17.16 , inference loss is 10.30, weight deacy loss is 6.86, training accuracy is 0.312500, time 211.765 samples/sec
+epoch 9, total_step 149600, total loss is 12.64 , inference loss is 5.78, weight deacy loss is 6.86, training accuracy is 0.468750, time 299.838 samples/sec
+epoch 9, total_step 149620, total loss is 14.86 , inference loss is 8.00, weight deacy loss is 6.86, training accuracy is 0.281250, time 217.374 samples/sec
+epoch 9, total_step 149640, total loss is 12.70 , inference loss is 5.84, weight deacy loss is 6.86, training accuracy is 0.312500, time 216.367 samples/sec
+epoch 9, total_step 149660, total loss is 17.35 , inference loss is 10.49, weight deacy loss is 6.86, training accuracy is 0.312500, time 212.195 samples/sec
+epoch 9, total_step 149680, total loss is 12.37 , inference loss is 5.51, weight deacy loss is 6.86, training accuracy is 0.531250, time 219.906 samples/sec
+epoch 9, total_step 149700, total loss is 18.01 , inference loss is 11.15, weight deacy loss is 6.86, training accuracy is 0.312500, time 11.569 samples/sec
+epoch 9, total_step 149720, total loss is 15.85 , inference loss is 8.99, weight deacy loss is 6.86, training accuracy is 0.437500, time 218.965 samples/sec
+epoch 9, total_step 149740, total loss is 14.59 , inference loss is 7.73, weight deacy loss is 6.86, training accuracy is 0.312500, time 213.904 samples/sec
+epoch 9, total_step 149760, total loss is 14.65 , inference loss is 7.79, weight deacy loss is 6.86, training accuracy is 0.312500, time 214.073 samples/sec
+epoch 9, total_step 149780, total loss is 17.23 , inference loss is 10.38, weight deacy loss is 6.86, training accuracy is 0.343750, time 213.096 samples/sec
+epoch 9, total_step 149800, total loss is 17.41 , inference loss is 10.56, weight deacy loss is 6.86, training accuracy is 0.281250, time 209.960 samples/sec
+epoch 9, total_step 149820, total loss is 16.16 , inference loss is 9.30, weight deacy loss is 6.86, training accuracy is 0.312500, time 210.233 samples/sec
+epoch 9, total_step 149840, total loss is 18.22 , inference loss is 11.37, weight deacy loss is 6.86, training accuracy is 0.312500, time 214.480 samples/sec
+epoch 9, total_step 149860, total loss is 13.32 , inference loss is 6.46, weight deacy loss is 6.86, training accuracy is 0.375000, time 299.323 samples/sec
+epoch 9, total_step 149880, total loss is 16.78 , inference loss is 9.92, weight deacy loss is 6.86, training accuracy is 0.281250, time 209.680 samples/sec
+epoch 9, total_step 149900, total loss is 15.23 , inference loss is 8.38, weight deacy loss is 6.85, training accuracy is 0.218750, time 215.279 samples/sec
+epoch 9, total_step 149920, total loss is 12.08 , inference loss is 5.23, weight deacy loss is 6.85, training accuracy is 0.500000, time 212.144 samples/sec
+epoch 9, total_step 149940, total loss is 18.21 , inference loss is 11.36, weight deacy loss is 6.85, training accuracy is 0.281250, time 213.060 samples/sec
+epoch 9, total_step 149960, total loss is 15.38 , inference loss is 8.52, weight deacy loss is 6.85, training accuracy is 0.375000, time 215.114 samples/sec
+epoch 9, total_step 149980, total loss is 15.54 , inference loss is 8.69, weight deacy loss is 6.85, training accuracy is 0.406250, time 214.867 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.16458600000002
+best_threshold_index 113 0.9877777777777778
+best_threshold_index 113 0.987037037037037
+best_threshold_index 113 0.9874074074074074
+best_threshold_index 112 0.9875925925925926
+best_threshold_index 113 0.9879629629629629
+best_threshold_index 113 0.987037037037037
+best_threshold_index 113 0.9872222222222222
+best_threshold_index 113 0.9868518518518519
+best_threshold_index 113 0.9861111111111112
+best_threshold_index 113 0.9866666666666667
+[lfw][150000]XNorm: 21.880269
+[lfw][150000]Accuracy-Flip: 0.98700+-0.00488
+test accuracy is: 0.9869999999999999
+epoch 9, total_step 150000, total loss is 16.05 , inference loss is 9.19, weight deacy loss is 6.85, training accuracy is 0.312500, time 15.452 samples/sec
+epoch 9, total_step 150020, total loss is 16.42 , inference loss is 9.57, weight deacy loss is 6.85, training accuracy is 0.500000, time 213.441 samples/sec
+epoch 9, total_step 150040, total loss is 13.99 , inference loss is 7.13, weight deacy loss is 6.85, training accuracy is 0.562500, time 209.147 samples/sec
+epoch 9, total_step 150060, total loss is 14.13 , inference loss is 7.28, weight deacy loss is 6.85, training accuracy is 0.437500, time 212.614 samples/sec
+epoch 9, total_step 150080, total loss is 15.47 , inference loss is 8.61, weight deacy loss is 6.85, training accuracy is 0.406250, time 217.650 samples/sec
+epoch 9, total_step 150100, total loss is 17.84 , inference loss is 10.99, weight deacy loss is 6.85, training accuracy is 0.312500, time 211.604 samples/sec
+epoch 9, total_step 150120, total loss is 15.82 , inference loss is 8.97, weight deacy loss is 6.85, training accuracy is 0.406250, time 222.292 samples/sec
+epoch 9, total_step 150140, total loss is 17.03 , inference loss is 10.18, weight deacy loss is 6.85, training accuracy is 0.281250, time 221.199 samples/sec
+epoch 9, total_step 150160, total loss is 13.16 , inference loss is 6.31, weight deacy loss is 6.85, training accuracy is 0.531250, time 300.018 samples/sec
+epoch 9, total_step 150180, total loss is 14.01 , inference loss is 7.16, weight deacy loss is 6.85, training accuracy is 0.531250, time 219.440 samples/sec
+epoch 9, total_step 150200, total loss is 19.42 , inference loss is 12.57, weight deacy loss is 6.85, training accuracy is 0.281250, time 217.532 samples/sec
+epoch 9, total_step 150220, total loss is 17.03 , inference loss is 10.18, weight deacy loss is 6.85, training accuracy is 0.343750, time 215.515 samples/sec
+epoch 9, total_step 150240, total loss is 14.58 , inference loss is 7.73, weight deacy loss is 6.85, training accuracy is 0.468750, time 217.792 samples/sec
+epoch 9, total_step 150260, total loss is 13.72 , inference loss is 6.88, weight deacy loss is 6.85, training accuracy is 0.312500, time 222.106 samples/sec
+epoch 9, total_step 150280, total loss is 15.06 , inference loss is 8.22, weight deacy loss is 6.85, training accuracy is 0.312500, time 219.085 samples/sec
+epoch 9, total_step 150300, total loss is 15.87 , inference loss is 9.02, weight deacy loss is 6.85, training accuracy is 0.500000, time 14.320 samples/sec
+epoch 9, total_step 150320, total loss is 13.01 , inference loss is 6.17, weight deacy loss is 6.85, training accuracy is 0.500000, time 221.358 samples/sec
+epoch 9, total_step 150340, total loss is 16.86 , inference loss is 10.01, weight deacy loss is 6.85, training accuracy is 0.250000, time 222.347 samples/sec
+epoch 9, total_step 150360, total loss is 15.48 , inference loss is 8.63, weight deacy loss is 6.85, training accuracy is 0.312500, time 215.187 samples/sec
+epoch 9, total_step 150380, total loss is 15.04 , inference loss is 8.19, weight deacy loss is 6.85, training accuracy is 0.281250, time 225.420 samples/sec
+epoch 9, total_step 150400, total loss is 15.14 , inference loss is 8.30, weight deacy loss is 6.85, training accuracy is 0.312500, time 210.271 samples/sec
+epoch 9, total_step 150420, total loss is 15.43 , inference loss is 8.58, weight deacy loss is 6.85, training accuracy is 0.437500, time 299.088 samples/sec
+epoch 9, total_step 150440, total loss is 16.29 , inference loss is 9.44, weight deacy loss is 6.85, training accuracy is 0.375000, time 299.349 samples/sec
+epoch 9, total_step 150460, total loss is 18.59 , inference loss is 11.74, weight deacy loss is 6.85, training accuracy is 0.312500, time 215.984 samples/sec
+epoch 9, total_step 150480, total loss is 11.61 , inference loss is 4.76, weight deacy loss is 6.85, training accuracy is 0.500000, time 214.865 samples/sec
+epoch 9, total_step 150500, total loss is 14.50 , inference loss is 7.66, weight deacy loss is 6.84, training accuracy is 0.375000, time 213.962 samples/sec
+epoch 9, total_step 150520, total loss is 15.72 , inference loss is 8.88, weight deacy loss is 6.84, training accuracy is 0.406250, time 219.269 samples/sec
+epoch 9, total_step 150540, total loss is 13.65 , inference loss is 6.80, weight deacy loss is 6.84, training accuracy is 0.375000, time 215.193 samples/sec
+epoch 9, total_step 150560, total loss is 17.07 , inference loss is 10.23, weight deacy loss is 6.84, training accuracy is 0.406250, time 215.255 samples/sec
+epoch 9, total_step 150580, total loss is 12.50 , inference loss is 5.66, weight deacy loss is 6.84, training accuracy is 0.531250, time 228.213 samples/sec
+epoch 9, total_step 150600, total loss is 15.06 , inference loss is 8.22, weight deacy loss is 6.84, training accuracy is 0.281250, time 13.089 samples/sec
+epoch 9, total_step 150620, total loss is 14.34 , inference loss is 7.49, weight deacy loss is 6.84, training accuracy is 0.281250, time 216.424 samples/sec
+epoch 9, total_step 150640, total loss is 17.08 , inference loss is 10.24, weight deacy loss is 6.84, training accuracy is 0.250000, time 207.920 samples/sec
+epoch 9, total_step 150660, total loss is 15.26 , inference loss is 8.41, weight deacy loss is 6.84, training accuracy is 0.312500, time 211.528 samples/sec
+epoch 9, total_step 150680, total loss is 16.22 , inference loss is 9.38, weight deacy loss is 6.84, training accuracy is 0.281250, time 213.576 samples/sec
+epoch 9, total_step 150700, total loss is 15.70 , inference loss is 8.86, weight deacy loss is 6.84, training accuracy is 0.250000, time 208.846 samples/sec
+epoch 9, total_step 150720, total loss is 19.56 , inference loss is 12.71, weight deacy loss is 6.84, training accuracy is 0.343750, time 209.312 samples/sec
+epoch 9, total_step 150740, total loss is 13.29 , inference loss is 6.44, weight deacy loss is 6.84, training accuracy is 0.531250, time 217.822 samples/sec
+epoch 9, total_step 150760, total loss is 14.40 , inference loss is 7.56, weight deacy loss is 6.84, training accuracy is 0.437500, time 209.720 samples/sec
+epoch 9, total_step 150780, total loss is 13.24 , inference loss is 6.40, weight deacy loss is 6.84, training accuracy is 0.343750, time 212.251 samples/sec
+epoch 9, total_step 150800, total loss is 20.19 , inference loss is 13.35, weight deacy loss is 6.84, training accuracy is 0.312500, time 216.283 samples/sec
+epoch 9, total_step 150820, total loss is 13.98 , inference loss is 7.14, weight deacy loss is 6.84, training accuracy is 0.468750, time 211.421 samples/sec
+epoch 9, total_step 150840, total loss is 16.41 , inference loss is 9.57, weight deacy loss is 6.84, training accuracy is 0.218750, time 216.698 samples/sec
+epoch 9, total_step 150860, total loss is 15.57 , inference loss is 8.73, weight deacy loss is 6.84, training accuracy is 0.281250, time 203.742 samples/sec
+epoch 9, total_step 150880, total loss is 18.98 , inference loss is 12.14, weight deacy loss is 6.84, training accuracy is 0.218750, time 297.633 samples/sec
+epoch 9, total_step 150900, total loss is 15.35 , inference loss is 8.51, weight deacy loss is 6.84, training accuracy is 0.406250, time 13.203 samples/sec
+epoch 9, total_step 150920, total loss is 16.16 , inference loss is 9.33, weight deacy loss is 6.84, training accuracy is 0.281250, time 211.828 samples/sec
+epoch 9, total_step 150940, total loss is 13.08 , inference loss is 6.24, weight deacy loss is 6.84, training accuracy is 0.343750, time 226.285 samples/sec
+epoch 9, total_step 150960, total loss is 21.18 , inference loss is 14.35, weight deacy loss is 6.84, training accuracy is 0.125000, time 212.777 samples/sec
+epoch 9, total_step 150980, total loss is 18.45 , inference loss is 11.61, weight deacy loss is 6.84, training accuracy is 0.343750, time 213.654 samples/sec
+epoch 9, total_step 151000, total loss is 17.21 , inference loss is 10.37, weight deacy loss is 6.84, training accuracy is 0.281250, time 218.580 samples/sec
+epoch 9, total_step 151020, total loss is 17.45 , inference loss is 10.61, weight deacy loss is 6.84, training accuracy is 0.312500, time 208.333 samples/sec
+epoch 9, total_step 151040, total loss is 14.50 , inference loss is 7.66, weight deacy loss is 6.84, training accuracy is 0.343750, time 211.939 samples/sec
+epoch 9, total_step 151060, total loss is 14.96 , inference loss is 8.12, weight deacy loss is 6.84, training accuracy is 0.281250, time 211.537 samples/sec
+epoch 9, total_step 151080, total loss is 13.57 , inference loss is 6.73, weight deacy loss is 6.83, training accuracy is 0.281250, time 221.375 samples/sec
+epoch 9, total_step 151100, total loss is 20.22 , inference loss is 13.38, weight deacy loss is 6.83, training accuracy is 0.281250, time 212.584 samples/sec
+epoch 9, total_step 151120, total loss is 16.98 , inference loss is 10.14, weight deacy loss is 6.83, training accuracy is 0.375000, time 213.364 samples/sec
+epoch 9, total_step 151140, total loss is 18.38 , inference loss is 11.55, weight deacy loss is 6.83, training accuracy is 0.250000, time 180.475 samples/sec
+epoch 9, total_step 151160, total loss is 12.71 , inference loss is 5.87, weight deacy loss is 6.83, training accuracy is 0.375000, time 298.479 samples/sec
+epoch 9, total_step 151180, total loss is 21.30 , inference loss is 14.47, weight deacy loss is 6.83, training accuracy is 0.312500, time 214.501 samples/sec
+epoch 9, total_step 151200, total loss is 16.80 , inference loss is 9.97, weight deacy loss is 6.83, training accuracy is 0.312500, time 12.857 samples/sec
+epoch 9, total_step 151220, total loss is 19.34 , inference loss is 12.51, weight deacy loss is 6.83, training accuracy is 0.343750, time 213.493 samples/sec
+epoch 9, total_step 151240, total loss is 13.82 , inference loss is 6.98, weight deacy loss is 6.83, training accuracy is 0.437500, time 217.181 samples/sec
+epoch 9, total_step 151260, total loss is 12.06 , inference loss is 5.22, weight deacy loss is 6.83, training accuracy is 0.437500, time 216.557 samples/sec
+epoch 9, total_step 151280, total loss is 15.60 , inference loss is 8.77, weight deacy loss is 6.83, training accuracy is 0.437500, time 214.571 samples/sec
+epoch 9, total_step 151300, total loss is 16.05 , inference loss is 9.22, weight deacy loss is 6.83, training accuracy is 0.343750, time 220.069 samples/sec
+epoch 9, total_step 151320, total loss is 18.38 , inference loss is 11.55, weight deacy loss is 6.83, training accuracy is 0.312500, time 212.874 samples/sec
+epoch 9, total_step 151340, total loss is 11.91 , inference loss is 5.08, weight deacy loss is 6.83, training accuracy is 0.531250, time 217.851 samples/sec
+epoch 9, total_step 151360, total loss is 18.70 , inference loss is 11.87, weight deacy loss is 6.83, training accuracy is 0.437500, time 217.547 samples/sec
+epoch 9, total_step 151380, total loss is 15.35 , inference loss is 8.52, weight deacy loss is 6.83, training accuracy is 0.406250, time 227.776 samples/sec
+epoch 9, total_step 151400, total loss is 17.54 , inference loss is 10.72, weight deacy loss is 6.83, training accuracy is 0.218750, time 217.854 samples/sec
+epoch 9, total_step 151420, total loss is 16.08 , inference loss is 9.25, weight deacy loss is 6.83, training accuracy is 0.343750, time 298.560 samples/sec
+epoch 9, total_step 151440, total loss is 13.36 , inference loss is 6.53, weight deacy loss is 6.83, training accuracy is 0.343750, time 235.222 samples/sec
+epoch 9, total_step 151460, total loss is 16.83 , inference loss is 10.00, weight deacy loss is 6.83, training accuracy is 0.218750, time 213.539 samples/sec
+epoch 9, total_step 151480, total loss is 15.27 , inference loss is 8.45, weight deacy loss is 6.83, training accuracy is 0.312500, time 216.163 samples/sec
+epoch 9, total_step 151500, total loss is 15.81 , inference loss is 8.99, weight deacy loss is 6.83, training accuracy is 0.375000, time 12.568 samples/sec
+epoch 9, total_step 151520, total loss is 18.48 , inference loss is 11.65, weight deacy loss is 6.83, training accuracy is 0.218750, time 216.124 samples/sec
+epoch 9, total_step 151540, total loss is 14.93 , inference loss is 8.10, weight deacy loss is 6.83, training accuracy is 0.250000, time 207.489 samples/sec
+epoch 9, total_step 151560, total loss is 19.52 , inference loss is 12.69, weight deacy loss is 6.83, training accuracy is 0.218750, time 212.587 samples/sec
+epoch 9, total_step 151580, total loss is 16.59 , inference loss is 9.77, weight deacy loss is 6.83, training accuracy is 0.312500, time 213.554 samples/sec
+epoch 9, total_step 151600, total loss is 13.79 , inference loss is 6.96, weight deacy loss is 6.83, training accuracy is 0.375000, time 213.931 samples/sec
+epoch 9, total_step 151620, total loss is 18.15 , inference loss is 11.33, weight deacy loss is 6.83, training accuracy is 0.250000, time 217.290 samples/sec
+epoch 9, total_step 151640, total loss is 18.20 , inference loss is 11.38, weight deacy loss is 6.82, training accuracy is 0.375000, time 215.043 samples/sec
+epoch 9, total_step 151660, total loss is 11.99 , inference loss is 5.16, weight deacy loss is 6.82, training accuracy is 0.531250, time 212.934 samples/sec
+epoch 9, total_step 151680, total loss is 16.53 , inference loss is 9.71, weight deacy loss is 6.82, training accuracy is 0.437500, time 298.427 samples/sec
+epoch 9, total_step 151700, total loss is 18.77 , inference loss is 11.95, weight deacy loss is 6.82, training accuracy is 0.250000, time 215.204 samples/sec
+epoch 9, total_step 151720, total loss is 15.95 , inference loss is 9.13, weight deacy loss is 6.82, training accuracy is 0.312500, time 221.439 samples/sec
+epoch 9, total_step 151740, total loss is 12.55 , inference loss is 5.73, weight deacy loss is 6.82, training accuracy is 0.343750, time 209.712 samples/sec
+epoch 9, total_step 151760, total loss is 13.38 , inference loss is 6.56, weight deacy loss is 6.82, training accuracy is 0.406250, time 217.202 samples/sec
+epoch 9, total_step 151780, total loss is 13.98 , inference loss is 7.16, weight deacy loss is 6.82, training accuracy is 0.375000, time 216.286 samples/sec
+epoch 9, total_step 151800, total loss is 13.94 , inference loss is 7.12, weight deacy loss is 6.82, training accuracy is 0.343750, time 12.162 samples/sec
+epoch 9, total_step 151820, total loss is 15.07 , inference loss is 8.25, weight deacy loss is 6.82, training accuracy is 0.343750, time 209.035 samples/sec
+epoch 9, total_step 151840, total loss is 12.92 , inference loss is 6.09, weight deacy loss is 6.82, training accuracy is 0.375000, time 211.953 samples/sec
+epoch 9, total_step 151860, total loss is 17.80 , inference loss is 10.98, weight deacy loss is 6.82, training accuracy is 0.312500, time 212.201 samples/sec
+epoch 9, total_step 151880, total loss is 14.40 , inference loss is 7.58, weight deacy loss is 6.82, training accuracy is 0.312500, time 210.644 samples/sec
+epoch 9, total_step 151900, total loss is 20.31 , inference loss is 13.49, weight deacy loss is 6.82, training accuracy is 0.218750, time 214.162 samples/sec
+epoch 9, total_step 151920, total loss is 16.79 , inference loss is 9.97, weight deacy loss is 6.82, training accuracy is 0.312500, time 216.642 samples/sec
+epoch 9, total_step 151940, total loss is 18.95 , inference loss is 12.13, weight deacy loss is 6.82, training accuracy is 0.375000, time 299.811 samples/sec
+epoch 9, total_step 151960, total loss is 19.61 , inference loss is 12.79, weight deacy loss is 6.82, training accuracy is 0.250000, time 216.184 samples/sec
+epoch 9, total_step 151980, total loss is 14.36 , inference loss is 7.55, weight deacy loss is 6.82, training accuracy is 0.437500, time 218.971 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.629655999999976
+best_threshold_index 109 0.9874074074074074
+best_threshold_index 109 0.9855555555555555
+best_threshold_index 109 0.9861111111111112
+best_threshold_index 109 0.9877777777777778
+best_threshold_index 109 0.9868518518518519
+best_threshold_index 109 0.9853703703703703
+best_threshold_index 109 0.9866666666666667
+best_threshold_index 112 0.9853703703703703
+best_threshold_index 109 0.9853703703703703
+best_threshold_index 109 0.9853703703703703
+[lfw][152000]XNorm: 22.079617
+[lfw][152000]Accuracy-Flip: 0.98583+-0.00779
+test accuracy is: 0.9858333333333335
+epoch 9, total_step 152000, total loss is 17.61 , inference loss is 10.79, weight deacy loss is 6.82, training accuracy is 0.281250, time 16.110 samples/sec
+epoch 9, total_step 152020, total loss is 14.77 , inference loss is 7.95, weight deacy loss is 6.82, training accuracy is 0.218750, time 207.762 samples/sec
+epoch 9, total_step 152040, total loss is 17.20 , inference loss is 10.38, weight deacy loss is 6.82, training accuracy is 0.281250, time 298.388 samples/sec
+epoch 9, total_step 152060, total loss is 15.37 , inference loss is 8.55, weight deacy loss is 6.82, training accuracy is 0.218750, time 210.922 samples/sec
+epoch 9, total_step 152080, total loss is 19.92 , inference loss is 13.10, weight deacy loss is 6.82, training accuracy is 0.218750, time 213.559 samples/sec
+epoch 9, total_step 152100, total loss is 16.81 , inference loss is 9.99, weight deacy loss is 6.82, training accuracy is 0.312500, time 12.014 samples/sec
+epoch 9, total_step 152120, total loss is 17.21 , inference loss is 10.39, weight deacy loss is 6.82, training accuracy is 0.250000, time 220.372 samples/sec
+epoch 9, total_step 152140, total loss is 23.85 , inference loss is 17.03, weight deacy loss is 6.82, training accuracy is 0.218750, time 222.122 samples/sec
+epoch 9, total_step 152160, total loss is 12.77 , inference loss is 5.96, weight deacy loss is 6.82, training accuracy is 0.468750, time 222.843 samples/sec
+epoch 9, total_step 152180, total loss is 13.72 , inference loss is 6.90, weight deacy loss is 6.82, training accuracy is 0.343750, time 216.468 samples/sec
+epoch 9, total_step 152200, total loss is 13.96 , inference loss is 7.14, weight deacy loss is 6.82, training accuracy is 0.343750, time 216.677 samples/sec
+epoch 9, total_step 152220, total loss is 17.57 , inference loss is 10.75, weight deacy loss is 6.81, training accuracy is 0.468750, time 213.317 samples/sec
+epoch 9, total_step 152240, total loss is 18.55 , inference loss is 11.74, weight deacy loss is 6.81, training accuracy is 0.281250, time 225.471 samples/sec
+epoch 9, total_step 152260, total loss is 18.47 , inference loss is 11.66, weight deacy loss is 6.81, training accuracy is 0.250000, time 219.841 samples/sec
+epoch 9, total_step 152280, total loss is 19.43 , inference loss is 12.61, weight deacy loss is 6.81, training accuracy is 0.250000, time 215.249 samples/sec
+epoch 9, total_step 152300, total loss is 16.74 , inference loss is 9.93, weight deacy loss is 6.81, training accuracy is 0.343750, time 215.109 samples/sec
+epoch 9, total_step 152320, total loss is 13.42 , inference loss is 6.60, weight deacy loss is 6.81, training accuracy is 0.406250, time 217.169 samples/sec
+epoch 9, total_step 152340, total loss is 13.87 , inference loss is 7.05, weight deacy loss is 6.81, training accuracy is 0.343750, time 297.819 samples/sec
+epoch 9, total_step 152360, total loss is 16.85 , inference loss is 10.04, weight deacy loss is 6.81, training accuracy is 0.250000, time 298.673 samples/sec
+epoch 9, total_step 152380, total loss is 14.29 , inference loss is 7.48, weight deacy loss is 6.81, training accuracy is 0.406250, time 209.591 samples/sec
+epoch 9, total_step 152400, total loss is 16.37 , inference loss is 9.56, weight deacy loss is 6.81, training accuracy is 0.250000, time 12.005 samples/sec
+epoch 9, total_step 152420, total loss is 20.32 , inference loss is 13.51, weight deacy loss is 6.81, training accuracy is 0.218750, time 299.727 samples/sec
+epoch 9, total_step 152440, total loss is 15.83 , inference loss is 9.02, weight deacy loss is 6.81, training accuracy is 0.343750, time 219.170 samples/sec
+epoch 9, total_step 152460, total loss is 16.93 , inference loss is 10.12, weight deacy loss is 6.81, training accuracy is 0.250000, time 209.527 samples/sec
+epoch 9, total_step 152480, total loss is 16.77 , inference loss is 9.96, weight deacy loss is 6.81, training accuracy is 0.250000, time 217.856 samples/sec
+epoch 9, total_step 152500, total loss is 16.81 , inference loss is 10.00, weight deacy loss is 6.81, training accuracy is 0.343750, time 215.712 samples/sec
+epoch 9, total_step 152520, total loss is 16.87 , inference loss is 10.06, weight deacy loss is 6.81, training accuracy is 0.406250, time 220.119 samples/sec
+epoch 9, total_step 152540, total loss is 18.93 , inference loss is 12.12, weight deacy loss is 6.81, training accuracy is 0.343750, time 221.848 samples/sec
+epoch 9, total_step 152560, total loss is 20.77 , inference loss is 13.96, weight deacy loss is 6.81, training accuracy is 0.218750, time 219.730 samples/sec
+epoch 9, total_step 152580, total loss is 12.91 , inference loss is 6.10, weight deacy loss is 6.81, training accuracy is 0.375000, time 221.385 samples/sec
+epoch 9, total_step 152600, total loss is 12.60 , inference loss is 5.80, weight deacy loss is 6.81, training accuracy is 0.406250, time 214.451 samples/sec
+epoch 9, total_step 152620, total loss is 15.69 , inference loss is 8.88, weight deacy loss is 6.81, training accuracy is 0.250000, time 216.527 samples/sec
+epoch 9, total_step 152640, total loss is 16.84 , inference loss is 10.04, weight deacy loss is 6.81, training accuracy is 0.375000, time 214.834 samples/sec
+epoch 9, total_step 152660, total loss is 17.13 , inference loss is 10.32, weight deacy loss is 6.81, training accuracy is 0.187500, time 214.106 samples/sec
+epoch 9, total_step 152680, total loss is 19.43 , inference loss is 12.62, weight deacy loss is 6.81, training accuracy is 0.375000, time 221.585 samples/sec
+epoch 9, total_step 152700, total loss is 17.74 , inference loss is 10.93, weight deacy loss is 6.81, training accuracy is 0.125000, time 12.795 samples/sec
+epoch 9, total_step 152720, total loss is 17.73 , inference loss is 10.92, weight deacy loss is 6.81, training accuracy is 0.218750, time 218.445 samples/sec
+epoch 9, total_step 152740, total loss is 14.13 , inference loss is 7.32, weight deacy loss is 6.81, training accuracy is 0.343750, time 217.081 samples/sec
+epoch 9, total_step 152760, total loss is 19.77 , inference loss is 12.97, weight deacy loss is 6.81, training accuracy is 0.187500, time 221.963 samples/sec
+epoch 9, total_step 152780, total loss is 16.22 , inference loss is 9.42, weight deacy loss is 6.81, training accuracy is 0.375000, time 216.051 samples/sec
+epoch 9, total_step 152800, total loss is 13.74 , inference loss is 6.94, weight deacy loss is 6.80, training accuracy is 0.281250, time 221.891 samples/sec
+epoch 9, total_step 152820, total loss is 17.42 , inference loss is 10.61, weight deacy loss is 6.80, training accuracy is 0.218750, time 211.427 samples/sec
+epoch 9, total_step 152840, total loss is 14.96 , inference loss is 8.15, weight deacy loss is 6.80, training accuracy is 0.343750, time 217.329 samples/sec
+epoch 9, total_step 152860, total loss is 16.02 , inference loss is 9.21, weight deacy loss is 6.80, training accuracy is 0.093750, time 215.766 samples/sec
+epoch 9, total_step 152880, total loss is 18.48 , inference loss is 11.67, weight deacy loss is 6.80, training accuracy is 0.187500, time 217.727 samples/sec
+epoch 9, total_step 152900, total loss is 16.90 , inference loss is 10.10, weight deacy loss is 6.80, training accuracy is 0.187500, time 218.700 samples/sec
+epoch 9, total_step 152920, total loss is 15.16 , inference loss is 8.36, weight deacy loss is 6.80, training accuracy is 0.250000, time 215.263 samples/sec
+epoch 9, total_step 152940, total loss is 16.08 , inference loss is 9.28, weight deacy loss is 6.80, training accuracy is 0.281250, time 212.045 samples/sec
+epoch 9, total_step 152960, total loss is 16.92 , inference loss is 10.12, weight deacy loss is 6.80, training accuracy is 0.218750, time 219.920 samples/sec
+epoch 9, total_step 152980, total loss is 18.20 , inference loss is 11.39, weight deacy loss is 6.80, training accuracy is 0.312500, time 215.998 samples/sec
+epoch 9, total_step 153000, total loss is 19.81 , inference loss is 13.01, weight deacy loss is 6.80, training accuracy is 0.250000, time 11.584 samples/sec
+epoch 9, total_step 153020, total loss is 14.56 , inference loss is 7.76, weight deacy loss is 6.80, training accuracy is 0.250000, time 227.316 samples/sec2022-12-01 10:27:12.292107: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7784 of 10000
+2022-12-01 10:27:15.149566: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 9, total_step 153040, total loss is 15.17 , inference loss is 8.37, weight deacy loss is 6.80, training accuracy is 0.281250, time 226.813 samples/sec
+epoch 9, total_step 153060, total loss is 14.99 , inference loss is 8.19, weight deacy loss is 6.80, training accuracy is 0.281250, time 243.604 samples/sec
+epoch 9, total_step 153080, total loss is 15.88 , inference loss is 9.08, weight deacy loss is 6.80, training accuracy is 0.281250, time 222.416 samples/sec
+epoch 9, total_step 153100, total loss is 17.96 , inference loss is 11.16, weight deacy loss is 6.80, training accuracy is 0.187500, time 227.114 samples/sec
+epoch 9, total_step 153120, total loss is 15.19 , inference loss is 8.39, weight deacy loss is 6.80, training accuracy is 0.343750, time 225.964 samples/sec
+epoch 9, total_step 153140, total loss is 19.26 , inference loss is 12.46, weight deacy loss is 6.80, training accuracy is 0.250000, time 237.191 samples/sec
+epoch 9, total_step 153160, total loss is 21.59 , inference loss is 14.80, weight deacy loss is 6.80, training accuracy is 0.281250, time 226.419 samples/sec
+epoch 9, total_step 153180, total loss is 16.30 , inference loss is 9.50, weight deacy loss is 6.80, training accuracy is 0.187500, time 216.135 samples/sec
+epoch 9, total_step 153200, total loss is 15.74 , inference loss is 8.94, weight deacy loss is 6.80, training accuracy is 0.375000, time 200.171 samples/sec
+epoch 9, total_step 153220, total loss is 19.47 , inference loss is 12.68, weight deacy loss is 6.80, training accuracy is 0.375000, time 176.732 samples/sec
+epoch 9, total_step 153240, total loss is 19.74 , inference loss is 12.94, weight deacy loss is 6.80, training accuracy is 0.312500, time 180.248 samples/sec
+epoch 9, total_step 153260, total loss is 16.28 , inference loss is 9.49, weight deacy loss is 6.80, training accuracy is 0.156250, time 177.821 samples/sec
+epoch 9, total_step 153280, total loss is 15.97 , inference loss is 9.17, weight deacy loss is 6.80, training accuracy is 0.250000, time 184.281 samples/sec
+epoch 9, total_step 153300, total loss is 12.10 , inference loss is 5.30, weight deacy loss is 6.80, training accuracy is 0.343750, time 11.376 samples/sec
+End of epoch 9
+epoch 10, total_step 153320, total loss is 19.46 , inference loss is 12.66, weight deacy loss is 6.80, training accuracy is 0.031250, time 212.906 samples/sec
+epoch 10, total_step 153340, total loss is 10.50 , inference loss is 3.71, weight deacy loss is 6.80, training accuracy is 0.625000, time 219.020 samples/sec
+epoch 10, total_step 153360, total loss is 12.35 , inference loss is 5.56, weight deacy loss is 6.80, training accuracy is 0.656250, time 214.291 samples/sec
+epoch 10, total_step 153380, total loss is 12.14 , inference loss is 5.35, weight deacy loss is 6.80, training accuracy is 0.718750, time 215.978 samples/sec
+epoch 10, total_step 153400, total loss is 10.91 , inference loss is 4.12, weight deacy loss is 6.79, training accuracy is 0.656250, time 213.360 samples/sec
+epoch 10, total_step 153420, total loss is 12.58 , inference loss is 5.79, weight deacy loss is 6.79, training accuracy is 0.500000, time 212.346 samples/sec
+epoch 10, total_step 153440, total loss is 10.77 , inference loss is 3.97, weight deacy loss is 6.79, training accuracy is 0.625000, time 224.372 samples/sec
+epoch 10, total_step 153460, total loss is 12.05 , inference loss is 5.25, weight deacy loss is 6.79, training accuracy is 0.656250, time 215.193 samples/sec
+epoch 10, total_step 153480, total loss is 12.24 , inference loss is 5.45, weight deacy loss is 6.79, training accuracy is 0.593750, time 219.273 samples/sec
+epoch 10, total_step 153500, total loss is 11.38 , inference loss is 4.59, weight deacy loss is 6.79, training accuracy is 0.625000, time 221.109 samples/sec
+epoch 10, total_step 153520, total loss is 15.32 , inference loss is 8.53, weight deacy loss is 6.79, training accuracy is 0.562500, time 299.129 samples/sec
+epoch 10, total_step 153540, total loss is 10.04 , inference loss is 3.25, weight deacy loss is 6.79, training accuracy is 0.625000, time 215.592 samples/sec
+epoch 10, total_step 153560, total loss is 10.39 , inference loss is 3.60, weight deacy loss is 6.79, training accuracy is 0.687500, time 215.253 samples/sec
+epoch 10, total_step 153580, total loss is 11.71 , inference loss is 4.92, weight deacy loss is 6.79, training accuracy is 0.468750, time 214.827 samples/sec
+epoch 10, total_step 153600, total loss is 13.09 , inference loss is 6.30, weight deacy loss is 6.79, training accuracy is 0.656250, time 12.205 samples/sec
+epoch 10, total_step 153620, total loss is 12.18 , inference loss is 5.39, weight deacy loss is 6.79, training accuracy is 0.593750, time 215.844 samples/sec
+epoch 10, total_step 153640, total loss is 11.29 , inference loss is 4.50, weight deacy loss is 6.79, training accuracy is 0.750000, time 210.882 samples/sec
+epoch 10, total_step 153660, total loss is 8.47 , inference loss is 1.68, weight deacy loss is 6.79, training accuracy is 0.718750, time 185.206 samples/sec
+epoch 10, total_step 153680, total loss is 18.94 , inference loss is 12.15, weight deacy loss is 6.79, training accuracy is 0.500000, time 181.248 samples/sec
+epoch 10, total_step 153700, total loss is 15.09 , inference loss is 8.30, weight deacy loss is 6.79, training accuracy is 0.562500, time 180.752 samples/sec
+epoch 10, total_step 153720, total loss is 11.87 , inference loss is 5.08, weight deacy loss is 6.79, training accuracy is 0.625000, time 176.359 samples/sec
+epoch 10, total_step 153740, total loss is 15.24 , inference loss is 8.45, weight deacy loss is 6.79, training accuracy is 0.562500, time 298.964 samples/sec
+epoch 10, total_step 153760, total loss is 11.62 , inference loss is 4.83, weight deacy loss is 6.79, training accuracy is 0.625000, time 181.243 samples/sec
+epoch 10, total_step 153780, total loss is 14.49 , inference loss is 7.70, weight deacy loss is 6.79, training accuracy is 0.562500, time 180.174 samples/sec
+epoch 10, total_step 153800, total loss is 12.18 , inference loss is 5.40, weight deacy loss is 6.79, training accuracy is 0.687500, time 180.207 samples/sec
+epoch 10, total_step 153820, total loss is 11.96 , inference loss is 5.17, weight deacy loss is 6.79, training accuracy is 0.593750, time 173.001 samples/sec
+epoch 10, total_step 153840, total loss is 14.52 , inference loss is 7.73, weight deacy loss is 6.79, training accuracy is 0.500000, time 170.070 samples/sec
+epoch 10, total_step 153860, total loss is 9.74 , inference loss is 2.95, weight deacy loss is 6.79, training accuracy is 0.718750, time 167.060 samples/sec
+epoch 10, total_step 153880, total loss is 12.09 , inference loss is 5.31, weight deacy loss is 6.79, training accuracy is 0.562500, time 180.719 samples/sec
+epoch 10, total_step 153900, total loss is 16.07 , inference loss is 9.28, weight deacy loss is 6.79, training accuracy is 0.500000, time 13.084 samples/sec
+epoch 10, total_step 153920, total loss is 11.65 , inference loss is 4.86, weight deacy loss is 6.79, training accuracy is 0.562500, time 214.460 samples/sec
+epoch 10, total_step 153940, total loss is 14.22 , inference loss is 7.43, weight deacy loss is 6.79, training accuracy is 0.593750, time 213.038 samples/sec
+epoch 10, total_step 153960, total loss is 12.04 , inference loss is 5.25, weight deacy loss is 6.79, training accuracy is 0.500000, time 211.351 samples/sec
+epoch 10, total_step 153980, total loss is 13.17 , inference loss is 6.38, weight deacy loss is 6.79, training accuracy is 0.625000, time 218.626 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.73838199999999
+best_threshold_index 113 0.9844444444444445
+best_threshold_index 113 0.9848148148148148
+best_threshold_index 113 0.9829629629629629
+best_threshold_index 113 0.9844444444444445
+best_threshold_index 113 0.9851851851851852
+best_threshold_index 113 0.9837037037037037
+best_threshold_index 113 0.9842592592592593
+best_threshold_index 113 0.985
+best_threshold_index 113 0.9833333333333333
+best_threshold_index 113 0.9835185185185186
+[lfw][154000]XNorm: 17.860944
+[lfw][154000]Accuracy-Flip: 0.98417+-0.00642
+test accuracy is: 0.9841666666666666
+epoch 10, total_step 154000, total loss is 15.33 , inference loss is 8.54, weight deacy loss is 6.79, training accuracy is 0.562500, time 15.731 samples/sec
+epoch 10, total_step 154020, total loss is 14.55 , inference loss is 7.76, weight deacy loss is 6.79, training accuracy is 0.468750, time 216.483 samples/sec
+epoch 10, total_step 154040, total loss is 13.54 , inference loss is 6.75, weight deacy loss is 6.79, training accuracy is 0.687500, time 225.094 samples/sec
+epoch 10, total_step 154060, total loss is 14.22 , inference loss is 7.43, weight deacy loss is 6.79, training accuracy is 0.562500, time 298.719 samples/sec
+epoch 10, total_step 154080, total loss is 18.53 , inference loss is 11.74, weight deacy loss is 6.79, training accuracy is 0.500000, time 203.204 samples/sec
+epoch 10, total_step 154100, total loss is 12.95 , inference loss is 6.16, weight deacy loss is 6.79, training accuracy is 0.437500, time 210.012 samples/sec
+epoch 10, total_step 154120, total loss is 14.67 , inference loss is 7.88, weight deacy loss is 6.79, training accuracy is 0.468750, time 209.098 samples/sec
+epoch 10, total_step 154140, total loss is 18.58 , inference loss is 11.79, weight deacy loss is 6.79, training accuracy is 0.406250, time 214.229 samples/sec
+epoch 10, total_step 154160, total loss is 13.66 , inference loss is 6.88, weight deacy loss is 6.79, training accuracy is 0.687500, time 209.389 samples/sec
+epoch 10, total_step 154180, total loss is 14.42 , inference loss is 7.64, weight deacy loss is 6.79, training accuracy is 0.593750, time 212.464 samples/sec
+epoch 10, total_step 154200, total loss is 17.22 , inference loss is 10.43, weight deacy loss is 6.79, training accuracy is 0.437500, time 12.241 samples/sec
+epoch 10, total_step 154220, total loss is 12.87 , inference loss is 6.08, weight deacy loss is 6.79, training accuracy is 0.593750, time 296.274 samples/sec
+epoch 10, total_step 154240, total loss is 16.30 , inference loss is 9.51, weight deacy loss is 6.79, training accuracy is 0.468750, time 208.980 samples/sec
+epoch 10, total_step 154260, total loss is 14.97 , inference loss is 8.18, weight deacy loss is 6.79, training accuracy is 0.406250, time 210.547 samples/sec
+epoch 10, total_step 154280, total loss is 10.82 , inference loss is 4.03, weight deacy loss is 6.79, training accuracy is 0.562500, time 208.086 samples/sec
+epoch 10, total_step 154300, total loss is 11.81 , inference loss is 5.03, weight deacy loss is 6.79, training accuracy is 0.593750, time 220.267 samples/sec
+epoch 10, total_step 154320, total loss is 11.81 , inference loss is 5.02, weight deacy loss is 6.79, training accuracy is 0.625000, time 213.495 samples/sec
+epoch 10, total_step 154340, total loss is 14.63 , inference loss is 7.84, weight deacy loss is 6.79, training accuracy is 0.500000, time 211.371 samples/sec
+epoch 10, total_step 154360, total loss is 15.02 , inference loss is 8.23, weight deacy loss is 6.79, training accuracy is 0.562500, time 211.065 samples/sec
+epoch 10, total_step 154380, total loss is 14.33 , inference loss is 7.55, weight deacy loss is 6.79, training accuracy is 0.625000, time 205.360 samples/sec
+epoch 10, total_step 154400, total loss is 15.22 , inference loss is 8.43, weight deacy loss is 6.79, training accuracy is 0.468750, time 215.455 samples/sec
+epoch 10, total_step 154420, total loss is 13.67 , inference loss is 6.88, weight deacy loss is 6.79, training accuracy is 0.562500, time 214.802 samples/sec
+epoch 10, total_step 154440, total loss is 16.02 , inference loss is 9.23, weight deacy loss is 6.79, training accuracy is 0.437500, time 204.053 samples/sec
+epoch 10, total_step 154460, total loss is 13.63 , inference loss is 6.84, weight deacy loss is 6.79, training accuracy is 0.468750, time 216.268 samples/sec
+epoch 10, total_step 154480, total loss is 14.99 , inference loss is 8.20, weight deacy loss is 6.79, training accuracy is 0.593750, time 214.385 samples/sec
+epoch 10, total_step 154500, total loss is 13.84 , inference loss is 7.05, weight deacy loss is 6.79, training accuracy is 0.531250, time 12.829 samples/sec
+epoch 10, total_step 154520, total loss is 11.36 , inference loss is 4.58, weight deacy loss is 6.79, training accuracy is 0.593750, time 210.928 samples/sec
+epoch 10, total_step 154540, total loss is 15.37 , inference loss is 8.59, weight deacy loss is 6.79, training accuracy is 0.468750, time 223.697 samples/sec
+epoch 10, total_step 154560, total loss is 11.15 , inference loss is 4.36, weight deacy loss is 6.79, training accuracy is 0.562500, time 216.075 samples/sec
+epoch 10, total_step 154580, total loss is 18.26 , inference loss is 11.47, weight deacy loss is 6.79, training accuracy is 0.437500, time 214.702 samples/sec
+epoch 10, total_step 154600, total loss is 15.67 , inference loss is 8.89, weight deacy loss is 6.79, training accuracy is 0.406250, time 221.371 samples/sec
+epoch 10, total_step 154620, total loss is 13.42 , inference loss is 6.63, weight deacy loss is 6.79, training accuracy is 0.437500, time 211.399 samples/sec
+epoch 10, total_step 154640, total loss is 14.08 , inference loss is 7.29, weight deacy loss is 6.79, training accuracy is 0.531250, time 221.067 samples/sec
+epoch 10, total_step 154660, total loss is 16.50 , inference loss is 9.71, weight deacy loss is 6.79, training accuracy is 0.406250, time 213.739 samples/sec
+epoch 10, total_step 154680, total loss is 15.50 , inference loss is 8.72, weight deacy loss is 6.79, training accuracy is 0.437500, time 214.879 samples/sec
+epoch 10, total_step 154700, total loss is 11.33 , inference loss is 4.54, weight deacy loss is 6.79, training accuracy is 0.625000, time 211.308 samples/sec
+epoch 10, total_step 154720, total loss is 16.79 , inference loss is 10.01, weight deacy loss is 6.79, training accuracy is 0.437500, time 219.757 samples/sec
+epoch 10, total_step 154740, total loss is 11.75 , inference loss is 4.96, weight deacy loss is 6.79, training accuracy is 0.656250, time 212.963 samples/sec
+epoch 10, total_step 154760, total loss is 14.64 , inference loss is 7.85, weight deacy loss is 6.79, training accuracy is 0.468750, time 214.449 samples/sec
+epoch 10, total_step 154780, total loss is 15.26 , inference loss is 8.47, weight deacy loss is 6.79, training accuracy is 0.531250, time 213.601 samples/sec
+epoch 10, total_step 154800, total loss is 13.01 , inference loss is 6.22, weight deacy loss is 6.79, training accuracy is 0.500000, time 12.252 samples/sec
+epoch 10, total_step 154820, total loss is 15.53 , inference loss is 8.74, weight deacy loss is 6.79, training accuracy is 0.406250, time 217.371 samples/sec
+epoch 10, total_step 154840, total loss is 12.04 , inference loss is 5.26, weight deacy loss is 6.79, training accuracy is 0.468750, time 214.582 samples/sec
+epoch 10, total_step 154860, total loss is 18.46 , inference loss is 11.67, weight deacy loss is 6.79, training accuracy is 0.312500, time 209.536 samples/sec
+epoch 10, total_step 154880, total loss is 14.74 , inference loss is 7.96, weight deacy loss is 6.79, training accuracy is 0.562500, time 213.332 samples/sec
+epoch 10, total_step 154900, total loss is 15.09 , inference loss is 8.30, weight deacy loss is 6.79, training accuracy is 0.468750, time 213.720 samples/sec
+epoch 10, total_step 154920, total loss is 13.96 , inference loss is 7.18, weight deacy loss is 6.79, training accuracy is 0.437500, time 215.414 samples/sec
+epoch 10, total_step 154940, total loss is 13.53 , inference loss is 6.74, weight deacy loss is 6.79, training accuracy is 0.531250, time 219.382 samples/sec
+epoch 10, total_step 154960, total loss is 13.51 , inference loss is 6.72, weight deacy loss is 6.79, training accuracy is 0.531250, time 215.491 samples/sec
+epoch 10, total_step 154980, total loss is 16.05 , inference loss is 9.26, weight deacy loss is 6.79, training accuracy is 0.437500, time 218.844 samples/sec
+epoch 10, total_step 155000, total loss is 13.40 , inference loss is 6.62, weight deacy loss is 6.79, training accuracy is 0.406250, time 227.720 samples/sec
+epoch 10, total_step 155020, total loss is 14.11 , inference loss is 7.33, weight deacy loss is 6.79, training accuracy is 0.593750, time 218.329 samples/sec
+epoch 10, total_step 155040, total loss is 14.83 , inference loss is 8.04, weight deacy loss is 6.79, training accuracy is 0.531250, time 213.893 samples/sec
+epoch 10, total_step 155060, total loss is 15.44 , inference loss is 8.65, weight deacy loss is 6.79, training accuracy is 0.468750, time 296.824 samples/sec
+epoch 10, total_step 155080, total loss is 16.98 , inference loss is 10.19, weight deacy loss is 6.79, training accuracy is 0.437500, time 223.900 samples/sec
+epoch 10, total_step 155100, total loss is 14.95 , inference loss is 8.17, weight deacy loss is 6.79, training accuracy is 0.531250, time 12.562 samples/sec
+epoch 10, total_step 155120, total loss is 16.88 , inference loss is 10.10, weight deacy loss is 6.79, training accuracy is 0.500000, time 220.903 samples/sec
+epoch 10, total_step 155140, total loss is 13.85 , inference loss is 7.07, weight deacy loss is 6.79, training accuracy is 0.406250, time 212.303 samples/sec
+epoch 10, total_step 155160, total loss is 13.97 , inference loss is 7.19, weight deacy loss is 6.79, training accuracy is 0.625000, time 217.322 samples/sec
+epoch 10, total_step 155180, total loss is 15.88 , inference loss is 9.10, weight deacy loss is 6.78, training accuracy is 0.437500, time 230.808 samples/sec
+epoch 10, total_step 155200, total loss is 14.93 , inference loss is 8.15, weight deacy loss is 6.78, training accuracy is 0.531250, time 219.896 samples/sec
+epoch 10, total_step 155220, total loss is 11.17 , inference loss is 4.39, weight deacy loss is 6.78, training accuracy is 0.562500, time 213.137 samples/sec
+epoch 10, total_step 155240, total loss is 12.70 , inference loss is 5.91, weight deacy loss is 6.78, training accuracy is 0.625000, time 219.711 samples/sec
+epoch 10, total_step 155260, total loss is 16.74 , inference loss is 9.95, weight deacy loss is 6.78, training accuracy is 0.468750, time 212.149 samples/sec
+epoch 10, total_step 155280, total loss is 12.79 , inference loss is 6.00, weight deacy loss is 6.78, training accuracy is 0.468750, time 217.433 samples/sec
+epoch 10, total_step 155300, total loss is 12.29 , inference loss is 5.51, weight deacy loss is 6.78, training accuracy is 0.593750, time 212.092 samples/sec
+epoch 10, total_step 155320, total loss is 12.50 , inference loss is 5.72, weight deacy loss is 6.78, training accuracy is 0.593750, time 209.985 samples/sec
+epoch 10, total_step 155340, total loss is 14.26 , inference loss is 7.48, weight deacy loss is 6.78, training accuracy is 0.406250, time 258.794 samples/sec
+epoch 10, total_step 155360, total loss is 14.10 , inference loss is 7.32, weight deacy loss is 6.78, training accuracy is 0.656250, time 214.072 samples/sec
+epoch 10, total_step 155380, total loss is 12.20 , inference loss is 5.42, weight deacy loss is 6.78, training accuracy is 0.500000, time 212.283 samples/sec
+epoch 10, total_step 155400, total loss is 18.27 , inference loss is 11.49, weight deacy loss is 6.78, training accuracy is 0.437500, time 11.830 samples/sec
+epoch 10, total_step 155420, total loss is 15.64 , inference loss is 8.85, weight deacy loss is 6.78, training accuracy is 0.437500, time 214.409 samples/sec
+epoch 10, total_step 155440, total loss is 10.39 , inference loss is 3.61, weight deacy loss is 6.78, training accuracy is 0.531250, time 218.191 samples/sec
+epoch 10, total_step 155460, total loss is 13.29 , inference loss is 6.51, weight deacy loss is 6.78, training accuracy is 0.562500, time 211.152 samples/sec
+epoch 10, total_step 155480, total loss is 12.64 , inference loss is 5.86, weight deacy loss is 6.78, training accuracy is 0.468750, time 216.154 samples/sec
+epoch 10, total_step 155500, total loss is 11.89 , inference loss is 5.11, weight deacy loss is 6.78, training accuracy is 0.531250, time 214.837 samples/sec
+epoch 10, total_step 155520, total loss is 14.27 , inference loss is 7.49, weight deacy loss is 6.78, training accuracy is 0.593750, time 218.620 samples/sec
+epoch 10, total_step 155540, total loss is 12.64 , inference loss is 5.86, weight deacy loss is 6.78, training accuracy is 0.500000, time 223.824 samples/sec
+epoch 10, total_step 155560, total loss is 9.44 , inference loss is 2.66, weight deacy loss is 6.78, training accuracy is 0.625000, time 212.064 samples/sec
+epoch 10, total_step 155580, total loss is 12.76 , inference loss is 5.98, weight deacy loss is 6.78, training accuracy is 0.562500, time 209.293 samples/sec
+epoch 10, total_step 155600, total loss is 9.90 , inference loss is 3.12, weight deacy loss is 6.78, training accuracy is 0.687500, time 189.344 samples/sec
+epoch 10, total_step 155620, total loss is 14.33 , inference loss is 7.55, weight deacy loss is 6.78, training accuracy is 0.437500, time 214.618 samples/sec
+epoch 10, total_step 155640, total loss is 11.48 , inference loss is 4.70, weight deacy loss is 6.78, training accuracy is 0.500000, time 216.104 samples/sec
+epoch 10, total_step 155660, total loss is 12.90 , inference loss is 6.12, weight deacy loss is 6.78, training accuracy is 0.593750, time 215.600 samples/sec
+epoch 10, total_step 155680, total loss is 12.31 , inference loss is 5.53, weight deacy loss is 6.78, training accuracy is 0.656250, time 213.202 samples/sec
+epoch 10, total_step 155700, total loss is 12.47 , inference loss is 5.69, weight deacy loss is 6.78, training accuracy is 0.625000, time 12.537 samples/sec
+epoch 10, total_step 155720, total loss is 10.45 , inference loss is 3.67, weight deacy loss is 6.78, training accuracy is 0.593750, time 222.006 samples/sec
+epoch 10, total_step 155740, total loss is 11.07 , inference loss is 4.29, weight deacy loss is 6.78, training accuracy is 0.687500, time 211.751 samples/sec
+epoch 10, total_step 155760, total loss is 12.36 , inference loss is 5.58, weight deacy loss is 6.78, training accuracy is 0.531250, time 214.665 samples/sec
+epoch 10, total_step 155780, total loss is 12.71 , inference loss is 5.93, weight deacy loss is 6.78, training accuracy is 0.437500, time 217.651 samples/sec
+epoch 10, total_step 155800, total loss is 11.31 , inference loss is 4.53, weight deacy loss is 6.78, training accuracy is 0.656250, time 214.061 samples/sec
+epoch 10, total_step 155820, total loss is 14.09 , inference loss is 7.32, weight deacy loss is 6.78, training accuracy is 0.531250, time 212.117 samples/sec
+epoch 10, total_step 155840, total loss is 12.20 , inference loss is 5.42, weight deacy loss is 6.78, training accuracy is 0.656250, time 212.891 samples/sec
+epoch 10, total_step 155860, total loss is 12.83 , inference loss is 6.05, weight deacy loss is 6.78, training accuracy is 0.468750, time 298.930 samples/sec
+epoch 10, total_step 155880, total loss is 12.90 , inference loss is 6.12, weight deacy loss is 6.78, training accuracy is 0.562500, time 298.368 samples/sec
+epoch 10, total_step 155900, total loss is 8.59 , inference loss is 1.81, weight deacy loss is 6.78, training accuracy is 0.687500, time 217.824 samples/sec
+epoch 10, total_step 155920, total loss is 13.78 , inference loss is 7.00, weight deacy loss is 6.77, training accuracy is 0.593750, time 214.004 samples/sec
+epoch 10, total_step 155940, total loss is 13.83 , inference loss is 7.06, weight deacy loss is 6.77, training accuracy is 0.437500, time 217.390 samples/sec
+epoch 10, total_step 155960, total loss is 10.10 , inference loss is 3.33, weight deacy loss is 6.77, training accuracy is 0.687500, time 226.086 samples/sec
+epoch 10, total_step 155980, total loss is 11.79 , inference loss is 5.02, weight deacy loss is 6.77, training accuracy is 0.593750, time 213.416 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.236431999999965
+best_threshold_index 118 0.9883333333333333
+best_threshold_index 118 0.9887037037037038
+best_threshold_index 118 0.9883333333333333
+best_threshold_index 118 0.9890740740740741
+best_threshold_index 118 0.9890740740740741
+best_threshold_index 118 0.9875925925925926
+best_threshold_index 118 0.9888888888888889
+best_threshold_index 120 0.9885185185185185
+best_threshold_index 118 0.9875925925925926
+best_threshold_index 118 0.9874074074074074
+[lfw][156000]XNorm: 21.152376
+[lfw][156000]Accuracy-Flip: 0.98800+-0.00542
+test accuracy is: 0.9879999999999999
+epoch 10, total_step 156000, total loss is 11.53 , inference loss is 4.76, weight deacy loss is 6.77, training accuracy is 0.468750, time 16.354 samples/sec
+epoch 10, total_step 156020, total loss is 15.52 , inference loss is 8.74, weight deacy loss is 6.77, training accuracy is 0.531250, time 215.678 samples/sec
+epoch 10, total_step 156040, total loss is 7.95 , inference loss is 1.18, weight deacy loss is 6.77, training accuracy is 0.718750, time 216.612 samples/sec
+epoch 10, total_step 156060, total loss is 10.15 , inference loss is 3.37, weight deacy loss is 6.77, training accuracy is 0.656250, time 213.123 samples/sec
+epoch 10, total_step 156080, total loss is 9.92 , inference loss is 3.15, weight deacy loss is 6.77, training accuracy is 0.656250, time 222.646 samples/sec
+epoch 10, total_step 156100, total loss is 10.54 , inference loss is 3.77, weight deacy loss is 6.77, training accuracy is 0.656250, time 296.716 samples/sec
+epoch 10, total_step 156120, total loss is 10.33 , inference loss is 3.56, weight deacy loss is 6.77, training accuracy is 0.593750, time 215.111 samples/sec
+epoch 10, total_step 156140, total loss is 14.04 , inference loss is 7.27, weight deacy loss is 6.77, training accuracy is 0.531250, time 217.352 samples/sec
+epoch 10, total_step 156160, total loss is 11.75 , inference loss is 4.98, weight deacy loss is 6.77, training accuracy is 0.625000, time 214.606 samples/sec
+epoch 10, total_step 156180, total loss is 10.39 , inference loss is 3.62, weight deacy loss is 6.77, training accuracy is 0.625000, time 214.194 samples/sec
+epoch 10, total_step 156200, total loss is 10.36 , inference loss is 3.59, weight deacy loss is 6.77, training accuracy is 0.625000, time 210.617 samples/sec
+epoch 10, total_step 156220, total loss is 11.22 , inference loss is 4.45, weight deacy loss is 6.77, training accuracy is 0.718750, time 220.964 samples/sec
+epoch 10, total_step 156240, total loss is 11.00 , inference loss is 4.23, weight deacy loss is 6.77, training accuracy is 0.656250, time 218.418 samples/sec
+epoch 10, total_step 156260, total loss is 10.60 , inference loss is 3.83, weight deacy loss is 6.77, training accuracy is 0.625000, time 212.748 samples/sec
+epoch 10, total_step 156280, total loss is 9.87 , inference loss is 3.10, weight deacy loss is 6.77, training accuracy is 0.687500, time 216.216 samples/sec
+epoch 10, total_step 156300, total loss is 11.85 , inference loss is 5.08, weight deacy loss is 6.77, training accuracy is 0.468750, time 12.189 samples/sec
+epoch 10, total_step 156320, total loss is 14.66 , inference loss is 7.89, weight deacy loss is 6.77, training accuracy is 0.406250, time 212.622 samples/sec
+epoch 10, total_step 156340, total loss is 14.57 , inference loss is 7.80, weight deacy loss is 6.77, training accuracy is 0.437500, time 209.701 samples/sec
+epoch 10, total_step 156360, total loss is 18.25 , inference loss is 11.48, weight deacy loss is 6.77, training accuracy is 0.437500, time 299.642 samples/sec
+epoch 10, total_step 156380, total loss is 15.68 , inference loss is 8.91, weight deacy loss is 6.77, training accuracy is 0.531250, time 217.998 samples/sec
+epoch 10, total_step 156400, total loss is 12.81 , inference loss is 6.04, weight deacy loss is 6.77, training accuracy is 0.562500, time 215.460 samples/sec
+epoch 10, total_step 156420, total loss is 12.51 , inference loss is 5.74, weight deacy loss is 6.77, training accuracy is 0.625000, time 217.271 samples/sec
+epoch 10, total_step 156440, total loss is 12.82 , inference loss is 6.05, weight deacy loss is 6.77, training accuracy is 0.593750, time 216.157 samples/sec
+epoch 10, total_step 156460, total loss is 15.06 , inference loss is 8.30, weight deacy loss is 6.77, training accuracy is 0.437500, time 208.803 samples/sec
+epoch 10, total_step 156480, total loss is 11.05 , inference loss is 4.28, weight deacy loss is 6.77, training accuracy is 0.625000, time 213.602 samples/sec
+epoch 10, total_step 156500, total loss is 16.20 , inference loss is 9.44, weight deacy loss is 6.77, training accuracy is 0.406250, time 205.025 samples/sec
+epoch 10, total_step 156520, total loss is 14.31 , inference loss is 7.54, weight deacy loss is 6.77, training accuracy is 0.500000, time 211.374 samples/sec
+epoch 10, total_step 156540, total loss is 11.69 , inference loss is 4.92, weight deacy loss is 6.77, training accuracy is 0.562500, time 215.666 samples/sec
+epoch 10, total_step 156560, total loss is 10.98 , inference loss is 4.21, weight deacy loss is 6.77, training accuracy is 0.625000, time 210.998 samples/sec
+epoch 10, total_step 156580, total loss is 17.81 , inference loss is 11.04, weight deacy loss is 6.77, training accuracy is 0.406250, time 209.689 samples/sec
+epoch 10, total_step 156600, total loss is 13.54 , inference loss is 6.77, weight deacy loss is 6.77, training accuracy is 0.500000, time 12.342 samples/sec
+epoch 10, total_step 156620, total loss is 18.33 , inference loss is 11.57, weight deacy loss is 6.77, training accuracy is 0.500000, time 298.598 samples/sec
+epoch 10, total_step 156640, total loss is 14.48 , inference loss is 7.72, weight deacy loss is 6.77, training accuracy is 0.468750, time 224.001 samples/sec
+epoch 10, total_step 156660, total loss is 12.59 , inference loss is 5.83, weight deacy loss is 6.77, training accuracy is 0.531250, time 213.771 samples/sec
+epoch 10, total_step 156680, total loss is 9.32 , inference loss is 2.55, weight deacy loss is 6.77, training accuracy is 0.562500, time 212.996 samples/sec
+epoch 10, total_step 156700, total loss is 11.61 , inference loss is 4.85, weight deacy loss is 6.77, training accuracy is 0.593750, time 222.299 samples/sec
+epoch 10, total_step 156720, total loss is 11.76 , inference loss is 5.00, weight deacy loss is 6.77, training accuracy is 0.562500, time 220.295 samples/sec
+epoch 10, total_step 156740, total loss is 16.18 , inference loss is 9.42, weight deacy loss is 6.76, training accuracy is 0.343750, time 209.881 samples/sec
+epoch 10, total_step 156760, total loss is 19.60 , inference loss is 12.84, weight deacy loss is 6.76, training accuracy is 0.312500, time 208.873 samples/sec
+epoch 10, total_step 156780, total loss is 16.26 , inference loss is 9.49, weight deacy loss is 6.76, training accuracy is 0.468750, time 217.953 samples/sec
+epoch 10, total_step 156800, total loss is 15.06 , inference loss is 8.29, weight deacy loss is 6.76, training accuracy is 0.312500, time 212.600 samples/sec
+epoch 10, total_step 156820, total loss is 12.88 , inference loss is 6.11, weight deacy loss is 6.76, training accuracy is 0.437500, time 216.617 samples/sec
+epoch 10, total_step 156840, total loss is 14.11 , inference loss is 7.35, weight deacy loss is 6.76, training accuracy is 0.437500, time 217.836 samples/sec
+epoch 10, total_step 156860, total loss is 14.97 , inference loss is 8.20, weight deacy loss is 6.76, training accuracy is 0.468750, time 211.074 samples/sec
+epoch 10, total_step 156880, total loss is 14.06 , inference loss is 7.30, weight deacy loss is 6.76, training accuracy is 0.375000, time 222.795 samples/sec
+epoch 10, total_step 156900, total loss is 14.19 , inference loss is 7.43, weight deacy loss is 6.76, training accuracy is 0.437500, time 12.391 samples/sec
+epoch 10, total_step 156920, total loss is 11.25 , inference loss is 4.48, weight deacy loss is 6.76, training accuracy is 0.562500, time 213.843 samples/sec
+epoch 10, total_step 156940, total loss is 14.51 , inference loss is 7.74, weight deacy loss is 6.76, training accuracy is 0.531250, time 221.219 samples/sec
+epoch 10, total_step 156960, total loss is 15.90 , inference loss is 9.13, weight deacy loss is 6.76, training accuracy is 0.437500, time 213.993 samples/sec
+epoch 10, total_step 156980, total loss is 14.04 , inference loss is 7.28, weight deacy loss is 6.76, training accuracy is 0.437500, time 218.175 samples/sec
+epoch 10, total_step 157000, total loss is 13.69 , inference loss is 6.93, weight deacy loss is 6.76, training accuracy is 0.375000, time 216.600 samples/sec
+epoch 10, total_step 157020, total loss is 12.90 , inference loss is 6.14, weight deacy loss is 6.76, training accuracy is 0.468750, time 217.862 samples/sec
+epoch 10, total_step 157040, total loss is 16.73 , inference loss is 9.96, weight deacy loss is 6.76, training accuracy is 0.437500, time 214.291 samples/sec
+epoch 10, total_step 157060, total loss is 15.47 , inference loss is 8.71, weight deacy loss is 6.76, training accuracy is 0.562500, time 228.002 samples/sec
+epoch 10, total_step 157080, total loss is 16.87 , inference loss is 10.11, weight deacy loss is 6.76, training accuracy is 0.312500, time 217.617 samples/sec
+epoch 10, total_step 157100, total loss is 15.40 , inference loss is 8.64, weight deacy loss is 6.76, training accuracy is 0.500000, time 214.531 samples/sec
+epoch 10, total_step 157120, total loss is 11.44 , inference loss is 4.68, weight deacy loss is 6.76, training accuracy is 0.593750, time 215.239 samples/sec
+epoch 10, total_step 157140, total loss is 14.07 , inference loss is 7.31, weight deacy loss is 6.76, training accuracy is 0.562500, time 214.605 samples/sec
+epoch 10, total_step 157160, total loss is 18.22 , inference loss is 11.46, weight deacy loss is 6.76, training accuracy is 0.406250, time 220.933 samples/sec
+epoch 10, total_step 157180, total loss is 16.92 , inference loss is 10.16, weight deacy loss is 6.76, training accuracy is 0.343750, time 215.669 samples/sec
+epoch 10, total_step 157200, total loss is 17.06 , inference loss is 10.30, weight deacy loss is 6.76, training accuracy is 0.343750, time 12.186 samples/sec
+epoch 10, total_step 157220, total loss is 13.34 , inference loss is 6.58, weight deacy loss is 6.76, training accuracy is 0.406250, time 213.584 samples/sec
+epoch 10, total_step 157240, total loss is 15.96 , inference loss is 9.20, weight deacy loss is 6.76, training accuracy is 0.500000, time 216.136 samples/sec
+epoch 10, total_step 157260, total loss is 15.12 , inference loss is 8.36, weight deacy loss is 6.76, training accuracy is 0.531250, time 219.290 samples/sec
+epoch 10, total_step 157280, total loss is 16.11 , inference loss is 9.35, weight deacy loss is 6.76, training accuracy is 0.437500, time 227.298 samples/sec
+epoch 10, total_step 157300, total loss is 15.44 , inference loss is 8.68, weight deacy loss is 6.76, training accuracy is 0.343750, time 213.460 samples/sec
+epoch 10, total_step 157320, total loss is 19.72 , inference loss is 12.96, weight deacy loss is 6.76, training accuracy is 0.375000, time 219.193 samples/sec
+epoch 10, total_step 157340, total loss is 17.39 , inference loss is 10.63, weight deacy loss is 6.76, training accuracy is 0.312500, time 215.971 samples/sec
+epoch 10, total_step 157360, total loss is 15.01 , inference loss is 8.25, weight deacy loss is 6.76, training accuracy is 0.312500, time 208.255 samples/sec
+epoch 10, total_step 157380, total loss is 14.13 , inference loss is 7.37, weight deacy loss is 6.76, training accuracy is 0.406250, time 215.690 samples/sec
+epoch 10, total_step 157400, total loss is 17.58 , inference loss is 10.82, weight deacy loss is 6.76, training accuracy is 0.468750, time 215.217 samples/sec
+epoch 10, total_step 157420, total loss is 11.74 , inference loss is 4.98, weight deacy loss is 6.76, training accuracy is 0.468750, time 216.632 samples/sec
+epoch 10, total_step 157440, total loss is 12.49 , inference loss is 5.73, weight deacy loss is 6.76, training accuracy is 0.531250, time 211.309 samples/sec
+epoch 10, total_step 157460, total loss is 12.16 , inference loss is 5.41, weight deacy loss is 6.76, training accuracy is 0.468750, time 296.162 samples/sec
+epoch 10, total_step 157480, total loss is 16.74 , inference loss is 9.98, weight deacy loss is 6.76, training accuracy is 0.437500, time 213.930 samples/sec
+epoch 10, total_step 157500, total loss is 15.01 , inference loss is 8.25, weight deacy loss is 6.76, training accuracy is 0.406250, time 11.632 samples/sec
+epoch 10, total_step 157520, total loss is 15.08 , inference loss is 8.33, weight deacy loss is 6.76, training accuracy is 0.406250, time 216.328 samples/sec
+epoch 10, total_step 157540, total loss is 12.61 , inference loss is 5.86, weight deacy loss is 6.76, training accuracy is 0.531250, time 216.541 samples/sec
+epoch 10, total_step 157560, total loss is 13.34 , inference loss is 6.58, weight deacy loss is 6.76, training accuracy is 0.468750, time 216.916 samples/sec
+epoch 10, total_step 157580, total loss is 16.37 , inference loss is 9.61, weight deacy loss is 6.76, training accuracy is 0.437500, time 218.906 samples/sec
+epoch 10, total_step 157600, total loss is 14.12 , inference loss is 7.36, weight deacy loss is 6.76, training accuracy is 0.375000, time 217.328 samples/sec
+epoch 10, total_step 157620, total loss is 14.07 , inference loss is 7.31, weight deacy loss is 6.76, training accuracy is 0.437500, time 215.643 samples/sec
+epoch 10, total_step 157640, total loss is 14.26 , inference loss is 7.50, weight deacy loss is 6.76, training accuracy is 0.437500, time 215.867 samples/sec
+epoch 10, total_step 157660, total loss is 13.54 , inference loss is 6.79, weight deacy loss is 6.76, training accuracy is 0.593750, time 217.651 samples/sec
+epoch 10, total_step 157680, total loss is 14.31 , inference loss is 7.56, weight deacy loss is 6.76, training accuracy is 0.312500, time 228.558 samples/sec
+epoch 10, total_step 157700, total loss is 14.50 , inference loss is 7.74, weight deacy loss is 6.76, training accuracy is 0.468750, time 213.924 samples/sec
+epoch 10, total_step 157720, total loss is 11.76 , inference loss is 5.01, weight deacy loss is 6.76, training accuracy is 0.531250, time 215.962 samples/sec
+epoch 10, total_step 157740, total loss is 10.20 , inference loss is 3.45, weight deacy loss is 6.76, training accuracy is 0.562500, time 222.074 samples/sec
+epoch 10, total_step 157760, total loss is 13.31 , inference loss is 6.56, weight deacy loss is 6.76, training accuracy is 0.531250, time 299.156 samples/sec
+epoch 10, total_step 157780, total loss is 14.42 , inference loss is 7.67, weight deacy loss is 6.76, training accuracy is 0.562500, time 297.633 samples/sec
+epoch 10, total_step 157800, total loss is 10.86 , inference loss is 4.10, weight deacy loss is 6.75, training accuracy is 0.500000, time 11.939 samples/sec
+epoch 10, total_step 157820, total loss is 10.43 , inference loss is 3.68, weight deacy loss is 6.75, training accuracy is 0.593750, time 215.319 samples/sec
+epoch 10, total_step 157840, total loss is 12.64 , inference loss is 5.88, weight deacy loss is 6.75, training accuracy is 0.562500, time 221.813 samples/sec
+epoch 10, total_step 157860, total loss is 13.36 , inference loss is 6.61, weight deacy loss is 6.75, training accuracy is 0.531250, time 299.330 samples/sec
+epoch 10, total_step 157880, total loss is 17.49 , inference loss is 10.74, weight deacy loss is 6.75, training accuracy is 0.375000, time 222.979 samples/sec
+epoch 10, total_step 157900, total loss is 15.43 , inference loss is 8.68, weight deacy loss is 6.75, training accuracy is 0.406250, time 224.396 samples/sec
+epoch 10, total_step 157920, total loss is 16.06 , inference loss is 9.30, weight deacy loss is 6.75, training accuracy is 0.593750, time 216.392 samples/sec
+epoch 10, total_step 157940, total loss is 14.19 , inference loss is 7.43, weight deacy loss is 6.75, training accuracy is 0.625000, time 215.245 samples/sec
+epoch 10, total_step 157960, total loss is 12.73 , inference loss is 5.98, weight deacy loss is 6.75, training accuracy is 0.562500, time 215.539 samples/sec
+epoch 10, total_step 157980, total loss is 18.06 , inference loss is 11.31, weight deacy loss is 6.75, training accuracy is 0.468750, time 213.442 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.096416000000033
+best_threshold_index 118 0.9883333333333333
+best_threshold_index 120 0.9877777777777778
+best_threshold_index 118 0.9885185185185185
+best_threshold_index 118 0.9888888888888889
+best_threshold_index 120 0.9898148148148148
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 118 0.9872222222222222
+best_threshold_index 118 0.987037037037037
+best_threshold_index 118 0.9877777777777778
+[lfw][158000]XNorm: 21.707070
+[lfw][158000]Accuracy-Flip: 0.98717+-0.00703
+test accuracy is: 0.9871666666666666
+epoch 10, total_step 158000, total loss is 13.37 , inference loss is 6.62, weight deacy loss is 6.75, training accuracy is 0.468750, time 15.750 samples/sec
+epoch 10, total_step 158020, total loss is 15.39 , inference loss is 8.63, weight deacy loss is 6.75, training accuracy is 0.375000, time 211.683 samples/sec
+epoch 10, total_step 158040, total loss is 13.48 , inference loss is 6.73, weight deacy loss is 6.75, training accuracy is 0.468750, time 218.567 samples/sec
+epoch 10, total_step 158060, total loss is 13.76 , inference loss is 7.01, weight deacy loss is 6.75, training accuracy is 0.500000, time 214.326 samples/sec
+epoch 10, total_step 158080, total loss is 14.51 , inference loss is 7.76, weight deacy loss is 6.75, training accuracy is 0.625000, time 212.229 samples/sec
+epoch 10, total_step 158100, total loss is 13.37 , inference loss is 6.61, weight deacy loss is 6.75, training accuracy is 0.500000, time 11.921 samples/sec
+epoch 10, total_step 158120, total loss is 15.25 , inference loss is 8.50, weight deacy loss is 6.75, training accuracy is 0.562500, time 213.988 samples/sec
+epoch 10, total_step 158140, total loss is 15.99 , inference loss is 9.24, weight deacy loss is 6.75, training accuracy is 0.500000, time 217.348 samples/sec
+epoch 10, total_step 158160, total loss is 17.29 , inference loss is 10.54, weight deacy loss is 6.75, training accuracy is 0.437500, time 218.641 samples/sec
+epoch 10, total_step 158180, total loss is 12.54 , inference loss is 5.79, weight deacy loss is 6.75, training accuracy is 0.437500, time 211.136 samples/sec
+epoch 10, total_step 158200, total loss is 11.91 , inference loss is 5.16, weight deacy loss is 6.75, training accuracy is 0.562500, time 214.698 samples/sec
+epoch 10, total_step 158220, total loss is 15.40 , inference loss is 8.65, weight deacy loss is 6.75, training accuracy is 0.406250, time 298.155 samples/sec
+epoch 10, total_step 158240, total loss is 13.55 , inference loss is 6.80, weight deacy loss is 6.75, training accuracy is 0.375000, time 221.630 samples/sec
+epoch 10, total_step 158260, total loss is 16.78 , inference loss is 10.03, weight deacy loss is 6.75, training accuracy is 0.375000, time 213.866 samples/sec
+epoch 10, total_step 158280, total loss is 18.46 , inference loss is 11.71, weight deacy loss is 6.75, training accuracy is 0.250000, time 215.550 samples/sec
+epoch 10, total_step 158300, total loss is 12.55 , inference loss is 5.80, weight deacy loss is 6.75, training accuracy is 0.531250, time 208.244 samples/sec
+epoch 10, total_step 158320, total loss is 12.35 , inference loss is 5.60, weight deacy loss is 6.75, training accuracy is 0.375000, time 214.804 samples/sec
+epoch 10, total_step 158340, total loss is 16.55 , inference loss is 9.80, weight deacy loss is 6.75, training accuracy is 0.500000, time 216.606 samples/sec
+epoch 10, total_step 158360, total loss is 14.59 , inference loss is 7.84, weight deacy loss is 6.75, training accuracy is 0.468750, time 211.027 samples/sec
+epoch 10, total_step 158380, total loss is 13.50 , inference loss is 6.75, weight deacy loss is 6.75, training accuracy is 0.468750, time 215.335 samples/sec
+epoch 10, total_step 158400, total loss is 11.16 , inference loss is 4.41, weight deacy loss is 6.75, training accuracy is 0.468750, time 12.399 samples/sec
+epoch 10, total_step 158420, total loss is 16.84 , inference loss is 10.09, weight deacy loss is 6.75, training accuracy is 0.375000, time 218.647 samples/sec
+epoch 10, total_step 158440, total loss is 13.72 , inference loss is 6.98, weight deacy loss is 6.75, training accuracy is 0.406250, time 216.044 samples/sec
+epoch 10, total_step 158460, total loss is 14.65 , inference loss is 7.90, weight deacy loss is 6.75, training accuracy is 0.500000, time 215.510 samples/sec
+epoch 10, total_step 158480, total loss is 15.81 , inference loss is 9.06, weight deacy loss is 6.75, training accuracy is 0.406250, time 298.207 samples/sec
+epoch 10, total_step 158500, total loss is 16.25 , inference loss is 9.51, weight deacy loss is 6.75, training accuracy is 0.375000, time 225.665 samples/sec
+epoch 10, total_step 158520, total loss is 15.42 , inference loss is 8.67, weight deacy loss is 6.75, training accuracy is 0.375000, time 218.778 samples/sec
+epoch 10, total_step 158540, total loss is 14.31 , inference loss is 7.56, weight deacy loss is 6.75, training accuracy is 0.562500, time 213.928 samples/sec
+epoch 10, total_step 158560, total loss is 15.40 , inference loss is 8.65, weight deacy loss is 6.75, training accuracy is 0.500000, time 210.229 samples/sec
+epoch 10, total_step 158580, total loss is 14.38 , inference loss is 7.63, weight deacy loss is 6.75, training accuracy is 0.437500, time 210.702 samples/sec
+epoch 10, total_step 158600, total loss is 13.27 , inference loss is 6.53, weight deacy loss is 6.75, training accuracy is 0.562500, time 219.391 samples/sec
+epoch 10, total_step 158620, total loss is 11.87 , inference loss is 5.13, weight deacy loss is 6.75, training accuracy is 0.531250, time 216.028 samples/sec
+epoch 10, total_step 158640, total loss is 11.49 , inference loss is 4.74, weight deacy loss is 6.75, training accuracy is 0.562500, time 211.970 samples/sec
+epoch 10, total_step 158660, total loss is 10.76 , inference loss is 4.01, weight deacy loss is 6.75, training accuracy is 0.531250, time 216.319 samples/sec
+epoch 10, total_step 158680, total loss is 12.68 , inference loss is 5.93, weight deacy loss is 6.75, training accuracy is 0.375000, time 213.911 samples/sec
+epoch 10, total_step 158700, total loss is 11.84 , inference loss is 5.09, weight deacy loss is 6.74, training accuracy is 0.625000, time 12.421 samples/sec
+epoch 10, total_step 158720, total loss is 11.61 , inference loss is 4.86, weight deacy loss is 6.74, training accuracy is 0.500000, time 215.688 samples/sec
+epoch 10, total_step 158740, total loss is 13.88 , inference loss is 7.13, weight deacy loss is 6.74, training accuracy is 0.562500, time 297.597 samples/sec
+epoch 10, total_step 158760, total loss is 14.77 , inference loss is 8.02, weight deacy loss is 6.74, training accuracy is 0.375000, time 217.616 samples/sec
+epoch 10, total_step 158780, total loss is 14.73 , inference loss is 7.99, weight deacy loss is 6.74, training accuracy is 0.500000, time 212.589 samples/sec
+epoch 10, total_step 158800, total loss is 19.55 , inference loss is 12.81, weight deacy loss is 6.74, training accuracy is 0.281250, time 213.908 samples/sec
+epoch 10, total_step 158820, total loss is 17.06 , inference loss is 10.32, weight deacy loss is 6.74, training accuracy is 0.375000, time 213.392 samples/sec
+epoch 10, total_step 158840, total loss is 13.13 , inference loss is 6.38, weight deacy loss is 6.74, training accuracy is 0.437500, time 213.115 samples/sec
+epoch 10, total_step 158860, total loss is 14.66 , inference loss is 7.91, weight deacy loss is 6.74, training accuracy is 0.468750, time 210.284 samples/sec
+epoch 10, total_step 158880, total loss is 16.41 , inference loss is 9.67, weight deacy loss is 6.74, training accuracy is 0.531250, time 208.871 samples/sec
+epoch 10, total_step 158900, total loss is 14.99 , inference loss is 8.24, weight deacy loss is 6.74, training accuracy is 0.343750, time 211.259 samples/sec
+epoch 10, total_step 158920, total loss is 13.08 , inference loss is 6.33, weight deacy loss is 6.74, training accuracy is 0.343750, time 212.229 samples/sec
+epoch 10, total_step 158940, total loss is 13.57 , inference loss is 6.82, weight deacy loss is 6.74, training accuracy is 0.562500, time 206.560 samples/sec
+epoch 10, total_step 158960, total loss is 13.21 , inference loss is 6.47, weight deacy loss is 6.74, training accuracy is 0.500000, time 209.825 samples/sec
+epoch 10, total_step 158980, total loss is 13.14 , inference loss is 6.40, weight deacy loss is 6.74, training accuracy is 0.500000, time 208.756 samples/sec
+epoch 10, total_step 159000, total loss is 14.47 , inference loss is 7.73, weight deacy loss is 6.74, training accuracy is 0.531250, time 13.516 samples/sec
+epoch 10, total_step 159020, total loss is 13.69 , inference loss is 6.95, weight deacy loss is 6.74, training accuracy is 0.406250, time 212.066 samples/sec
+epoch 10, total_step 159040, total loss is 16.59 , inference loss is 9.84, weight deacy loss is 6.74, training accuracy is 0.187500, time 212.539 samples/sec
+epoch 10, total_step 159060, total loss is 14.74 , inference loss is 8.00, weight deacy loss is 6.74, training accuracy is 0.593750, time 215.065 samples/sec
+epoch 10, total_step 159080, total loss is 13.21 , inference loss is 6.47, weight deacy loss is 6.74, training accuracy is 0.500000, time 216.172 samples/sec
+epoch 10, total_step 159100, total loss is 19.82 , inference loss is 13.08, weight deacy loss is 6.74, training accuracy is 0.468750, time 212.828 samples/sec
+epoch 10, total_step 159120, total loss is 16.48 , inference loss is 9.74, weight deacy loss is 6.74, training accuracy is 0.406250, time 212.214 samples/sec
+epoch 10, total_step 159140, total loss is 12.98 , inference loss is 6.24, weight deacy loss is 6.74, training accuracy is 0.500000, time 217.597 samples/sec
+epoch 10, total_step 159160, total loss is 13.58 , inference loss is 6.84, weight deacy loss is 6.74, training accuracy is 0.593750, time 213.185 samples/sec
+epoch 10, total_step 159180, total loss is 16.71 , inference loss is 9.97, weight deacy loss is 6.74, training accuracy is 0.375000, time 215.245 samples/sec
+epoch 10, total_step 159200, total loss is 11.79 , inference loss is 5.05, weight deacy loss is 6.74, training accuracy is 0.468750, time 215.396 samples/sec
+epoch 10, total_step 159220, total loss is 16.55 , inference loss is 9.81, weight deacy loss is 6.74, training accuracy is 0.437500, time 216.330 samples/sec
+epoch 10, total_step 159240, total loss is 14.36 , inference loss is 7.63, weight deacy loss is 6.74, training accuracy is 0.468750, time 212.282 samples/sec
+epoch 10, total_step 159260, total loss is 13.91 , inference loss is 7.17, weight deacy loss is 6.74, training accuracy is 0.500000, time 218.088 samples/sec
+epoch 10, total_step 159280, total loss is 12.96 , inference loss is 6.23, weight deacy loss is 6.74, training accuracy is 0.500000, time 213.750 samples/sec
+epoch 10, total_step 159300, total loss is 14.62 , inference loss is 7.88, weight deacy loss is 6.74, training accuracy is 0.375000, time 11.502 samples/sec
+epoch 10, total_step 159320, total loss is 15.84 , inference loss is 9.11, weight deacy loss is 6.74, training accuracy is 0.375000, time 212.860 samples/sec
+epoch 10, total_step 159340, total loss is 15.74 , inference loss is 9.00, weight deacy loss is 6.74, training accuracy is 0.312500, time 217.287 samples/sec
+epoch 10, total_step 159360, total loss is 18.65 , inference loss is 11.91, weight deacy loss is 6.74, training accuracy is 0.437500, time 215.497 samples/sec
+epoch 10, total_step 159380, total loss is 11.25 , inference loss is 4.51, weight deacy loss is 6.74, training accuracy is 0.531250, time 220.498 samples/sec
+epoch 10, total_step 159400, total loss is 14.97 , inference loss is 8.24, weight deacy loss is 6.74, training accuracy is 0.406250, time 217.703 samples/sec
+epoch 10, total_step 159420, total loss is 16.30 , inference loss is 9.57, weight deacy loss is 6.74, training accuracy is 0.375000, time 220.247 samples/sec
+epoch 10, total_step 159440, total loss is 16.32 , inference loss is 9.59, weight deacy loss is 6.74, training accuracy is 0.375000, time 219.107 samples/sec
+epoch 10, total_step 159460, total loss is 15.94 , inference loss is 9.21, weight deacy loss is 6.74, training accuracy is 0.406250, time 214.959 samples/sec
+epoch 10, total_step 159480, total loss is 12.79 , inference loss is 6.05, weight deacy loss is 6.74, training accuracy is 0.562500, time 221.715 samples/sec
+epoch 10, total_step 159500, total loss is 11.26 , inference loss is 4.52, weight deacy loss is 6.73, training accuracy is 0.687500, time 221.170 samples/sec
+epoch 10, total_step 159520, total loss is 14.14 , inference loss is 7.41, weight deacy loss is 6.73, training accuracy is 0.406250, time 298.162 samples/sec
+epoch 10, total_step 159540, total loss is 14.50 , inference loss is 7.77, weight deacy loss is 6.73, training accuracy is 0.562500, time 299.379 samples/sec
+epoch 10, total_step 159560, total loss is 13.05 , inference loss is 6.31, weight deacy loss is 6.73, training accuracy is 0.562500, time 215.899 samples/sec
+epoch 10, total_step 159580, total loss is 12.75 , inference loss is 6.01, weight deacy loss is 6.73, training accuracy is 0.500000, time 212.922 samples/sec
+epoch 10, total_step 159600, total loss is 12.72 , inference loss is 5.98, weight deacy loss is 6.73, training accuracy is 0.500000, time 13.004 samples/sec
+epoch 10, total_step 159620, total loss is 17.79 , inference loss is 11.05, weight deacy loss is 6.73, training accuracy is 0.406250, time 222.854 samples/sec
+epoch 10, total_step 159640, total loss is 13.47 , inference loss is 6.73, weight deacy loss is 6.73, training accuracy is 0.531250, time 212.774 samples/sec
+epoch 10, total_step 159660, total loss is 14.96 , inference loss is 8.22, weight deacy loss is 6.73, training accuracy is 0.531250, time 214.379 samples/sec
+epoch 10, total_step 159680, total loss is 15.90 , inference loss is 9.17, weight deacy loss is 6.73, training accuracy is 0.500000, time 209.518 samples/sec
+epoch 10, total_step 159700, total loss is 16.72 , inference loss is 9.99, weight deacy loss is 6.73, training accuracy is 0.312500, time 298.963 samples/sec
+epoch 10, total_step 159720, total loss is 13.19 , inference loss is 6.46, weight deacy loss is 6.73, training accuracy is 0.468750, time 297.638 samples/sec
+epoch 10, total_step 159740, total loss is 19.85 , inference loss is 13.12, weight deacy loss is 6.73, training accuracy is 0.437500, time 299.090 samples/sec
+epoch 10, total_step 159760, total loss is 13.48 , inference loss is 6.75, weight deacy loss is 6.73, training accuracy is 0.406250, time 217.869 samples/sec
+epoch 10, total_step 159780, total loss is 13.16 , inference loss is 6.43, weight deacy loss is 6.73, training accuracy is 0.437500, time 227.380 samples/sec
+epoch 10, total_step 159800, total loss is 16.52 , inference loss is 9.79, weight deacy loss is 6.73, training accuracy is 0.312500, time 218.749 samples/sec
+epoch 10, total_step 159820, total loss is 16.82 , inference loss is 10.09, weight deacy loss is 6.73, training accuracy is 0.468750, time 234.933 samples/sec
+epoch 10, total_step 159840, total loss is 16.61 , inference loss is 9.88, weight deacy loss is 6.73, training accuracy is 0.218750, time 227.260 samples/sec
+epoch 10, total_step 159860, total loss is 15.37 , inference loss is 8.64, weight deacy loss is 6.73, training accuracy is 0.375000, time 211.726 samples/sec
+epoch 10, total_step 159880, total loss is 15.79 , inference loss is 9.06, weight deacy loss is 6.73, training accuracy is 0.375000, time 214.325 samples/sec
+epoch 10, total_step 159900, total loss is 12.08 , inference loss is 5.35, weight deacy loss is 6.73, training accuracy is 0.500000, time 5.289 samples/sec
+epoch 10, total_step 159920, total loss is 15.34 , inference loss is 8.61, weight deacy loss is 6.73, training accuracy is 0.531250, time 299.200 samples/sec
+epoch 10, total_step 159940, total loss is 17.40 , inference loss is 10.67, weight deacy loss is 6.73, training accuracy is 0.437500, time 298.812 samples/sec
+epoch 10, total_step 159960, total loss is 15.50 , inference loss is 8.77, weight deacy loss is 6.73, training accuracy is 0.312500, time 297.999 samples/sec
+epoch 10, total_step 159980, total loss is 16.22 , inference loss is 9.49, weight deacy loss is 6.73, training accuracy is 0.406250, time 211.463 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.572397
+best_threshold_index 115 0.9901851851851852
+best_threshold_index 115 0.9890740740740741
+best_threshold_index 115 0.9894444444444445
+best_threshold_index 115 0.9905555555555555
+best_threshold_index 115 0.9896296296296296
+best_threshold_index 115 0.9890740740740741
+best_threshold_index 115 0.9898148148148148
+best_threshold_index 115 0.9892592592592593
+best_threshold_index 115 0.9887037037037038
+best_threshold_index 115 0.9892592592592593
+[lfw][160000]XNorm: 21.640497
+[lfw][160000]Accuracy-Flip: 0.98950+-0.00478
+test accuracy is: 0.9894999999999999
+epoch 10, total_step 160000, total loss is 12.54 , inference loss is 5.81, weight deacy loss is 6.73, training accuracy is 0.406250, time 10.623 samples/sec
+epoch 10, total_step 160020, total loss is 20.90 , inference loss is 14.18, weight deacy loss is 6.73, training accuracy is 0.281250, time 214.792 samples/sec
+epoch 10, total_step 160040, total loss is 11.89 , inference loss is 5.16, weight deacy loss is 6.73, training accuracy is 0.468750, time 218.998 samples/sec
+epoch 10, total_step 160060, total loss is 17.24 , inference loss is 10.51, weight deacy loss is 6.73, training accuracy is 0.281250, time 217.183 samples/sec
+epoch 10, total_step 160080, total loss is 15.40 , inference loss is 8.67, weight deacy loss is 6.73, training accuracy is 0.531250, time 219.164 samples/sec
+epoch 10, total_step 160100, total loss is 16.49 , inference loss is 9.76, weight deacy loss is 6.73, training accuracy is 0.312500, time 223.938 samples/sec
+epoch 10, total_step 160120, total loss is 14.31 , inference loss is 7.59, weight deacy loss is 6.73, training accuracy is 0.343750, time 215.470 samples/sec
+epoch 10, total_step 160140, total loss is 15.50 , inference loss is 8.77, weight deacy loss is 6.73, training accuracy is 0.531250, time 209.032 samples/sec
+epoch 10, total_step 160160, total loss is 16.96 , inference loss is 10.23, weight deacy loss is 6.73, training accuracy is 0.468750, time 225.880 samples/sec
+epoch 10, total_step 160180, total loss is 13.40 , inference loss is 6.67, weight deacy loss is 6.73, training accuracy is 0.531250, time 213.173 samples/sec
+epoch 10, total_step 160200, total loss is 14.30 , inference loss is 7.57, weight deacy loss is 6.73, training accuracy is 0.375000, time 8.498 samples/sec
+epoch 10, total_step 160220, total loss is 12.45 , inference loss is 5.72, weight deacy loss is 6.73, training accuracy is 0.531250, time 223.453 samples/sec
+epoch 10, total_step 160240, total loss is 16.71 , inference loss is 9.98, weight deacy loss is 6.73, training accuracy is 0.281250, time 228.178 samples/sec
+epoch 10, total_step 160260, total loss is 15.25 , inference loss is 8.53, weight deacy loss is 6.73, training accuracy is 0.312500, time 221.429 samples/sec
+epoch 10, total_step 160280, total loss is 13.01 , inference loss is 6.28, weight deacy loss is 6.73, training accuracy is 0.531250, time 231.369 samples/sec
+epoch 10, total_step 160300, total loss is 12.66 , inference loss is 5.94, weight deacy loss is 6.73, training accuracy is 0.406250, time 216.540 samples/sec
+epoch 10, total_step 160320, total loss is 13.48 , inference loss is 6.75, weight deacy loss is 6.72, training accuracy is 0.375000, time 205.746 samples/sec
+epoch 10, total_step 160340, total loss is 16.80 , inference loss is 10.08, weight deacy loss is 6.72, training accuracy is 0.468750, time 226.349 samples/sec
+epoch 10, total_step 160360, total loss is 16.08 , inference loss is 9.35, weight deacy loss is 6.72, training accuracy is 0.281250, time 228.241 samples/sec
+epoch 10, total_step 160380, total loss is 13.63 , inference loss is 6.91, weight deacy loss is 6.72, training accuracy is 0.562500, time 213.197 samples/sec
+epoch 10, total_step 160400, total loss is 12.81 , inference loss is 6.08, weight deacy loss is 6.72, training accuracy is 0.375000, time 218.842 samples/sec
+epoch 10, total_step 160420, total loss is 14.94 , inference loss is 8.21, weight deacy loss is 6.72, training accuracy is 0.468750, time 218.003 samples/sec
+epoch 10, total_step 160440, total loss is 17.11 , inference loss is 10.39, weight deacy loss is 6.72, training accuracy is 0.187500, time 218.985 samples/sec
+epoch 10, total_step 160460, total loss is 15.54 , inference loss is 8.82, weight deacy loss is 6.72, training accuracy is 0.406250, time 298.975 samples/sec
+epoch 10, total_step 160480, total loss is 14.22 , inference loss is 7.50, weight deacy loss is 6.72, training accuracy is 0.500000, time 215.794 samples/sec
+epoch 10, total_step 160500, total loss is 13.91 , inference loss is 7.19, weight deacy loss is 6.72, training accuracy is 0.343750, time 11.259 samples/sec
+epoch 10, total_step 160520, total loss is 16.31 , inference loss is 9.59, weight deacy loss is 6.72, training accuracy is 0.406250, time 225.861 samples/sec
+epoch 10, total_step 160540, total loss is 17.12 , inference loss is 10.40, weight deacy loss is 6.72, training accuracy is 0.312500, time 222.183 samples/sec
+epoch 10, total_step 160560, total loss is 14.31 , inference loss is 7.59, weight deacy loss is 6.72, training accuracy is 0.500000, time 217.367 samples/sec
+epoch 10, total_step 160580, total loss is 17.17 , inference loss is 10.45, weight deacy loss is 6.72, training accuracy is 0.250000, time 215.016 samples/sec
+epoch 10, total_step 160600, total loss is 14.69 , inference loss is 7.97, weight deacy loss is 6.72, training accuracy is 0.531250, time 209.493 samples/sec
+epoch 10, total_step 160620, total loss is 18.60 , inference loss is 11.88, weight deacy loss is 6.72, training accuracy is 0.406250, time 221.674 samples/sec
+epoch 10, total_step 160640, total loss is 15.16 , inference loss is 8.43, weight deacy loss is 6.72, training accuracy is 0.437500, time 218.405 samples/sec
+epoch 10, total_step 160660, total loss is 14.54 , inference loss is 7.82, weight deacy loss is 6.72, training accuracy is 0.562500, time 217.128 samples/sec
+epoch 10, total_step 160680, total loss is 17.09 , inference loss is 10.37, weight deacy loss is 6.72, training accuracy is 0.312500, time 218.452 samples/sec
+epoch 10, total_step 160700, total loss is 13.26 , inference loss is 6.54, weight deacy loss is 6.72, training accuracy is 0.343750, time 297.735 samples/sec
+epoch 10, total_step 160720, total loss is 17.14 , inference loss is 10.42, weight deacy loss is 6.72, training accuracy is 0.406250, time 227.226 samples/sec
+epoch 10, total_step 160740, total loss is 14.65 , inference loss is 7.93, weight deacy loss is 6.72, training accuracy is 0.468750, time 213.096 samples/sec
+epoch 10, total_step 160760, total loss is 10.28 , inference loss is 3.56, weight deacy loss is 6.72, training accuracy is 0.625000, time 225.751 samples/sec
+epoch 10, total_step 160780, total loss is 14.92 , inference loss is 8.20, weight deacy loss is 6.72, training accuracy is 0.312500, time 216.996 samples/sec
+epoch 10, total_step 160800, total loss is 14.59 , inference loss is 7.87, weight deacy loss is 6.72, training accuracy is 0.312500, time 12.117 samples/sec
+epoch 10, total_step 160820, total loss is 12.24 , inference loss is 5.53, weight deacy loss is 6.72, training accuracy is 0.375000, time 216.639 samples/sec
+epoch 10, total_step 160840, total loss is 14.63 , inference loss is 7.91, weight deacy loss is 6.72, training accuracy is 0.468750, time 213.038 samples/sec
+epoch 10, total_step 160860, total loss is 12.93 , inference loss is 6.21, weight deacy loss is 6.72, training accuracy is 0.468750, time 219.996 samples/sec
+epoch 10, total_step 160880, total loss is 9.95 , inference loss is 3.24, weight deacy loss is 6.72, training accuracy is 0.593750, time 214.439 samples/sec
+epoch 10, total_step 160900, total loss is 19.85 , inference loss is 13.13, weight deacy loss is 6.72, training accuracy is 0.312500, time 214.063 samples/sec
+epoch 10, total_step 160920, total loss is 13.13 , inference loss is 6.42, weight deacy loss is 6.72, training accuracy is 0.656250, time 211.890 samples/sec
+epoch 10, total_step 160940, total loss is 14.35 , inference loss is 7.63, weight deacy loss is 6.72, training accuracy is 0.531250, time 213.505 samples/sec
+epoch 10, total_step 160960, total loss is 18.70 , inference loss is 11.99, weight deacy loss is 6.72, training accuracy is 0.281250, time 296.610 samples/sec
+epoch 10, total_step 160980, total loss is 15.95 , inference loss is 9.23, weight deacy loss is 6.72, training accuracy is 0.375000, time 213.485 samples/sec
+epoch 10, total_step 161000, total loss is 13.62 , inference loss is 6.91, weight deacy loss is 6.72, training accuracy is 0.406250, time 215.364 samples/sec
+epoch 10, total_step 161020, total loss is 14.75 , inference loss is 8.04, weight deacy loss is 6.72, training accuracy is 0.468750, time 215.024 samples/sec
+epoch 10, total_step 161040, total loss is 14.23 , inference loss is 7.52, weight deacy loss is 6.72, training accuracy is 0.312500, time 227.267 samples/sec
+epoch 10, total_step 161060, total loss is 13.15 , inference loss is 6.43, weight deacy loss is 6.72, training accuracy is 0.593750, time 212.568 samples/sec
+epoch 10, total_step 161080, total loss is 15.47 , inference loss is 8.76, weight deacy loss is 6.71, training accuracy is 0.500000, time 217.523 samples/sec
+epoch 10, total_step 161100, total loss is 16.82 , inference loss is 10.11, weight deacy loss is 6.71, training accuracy is 0.406250, time 9.893 samples/sec
+epoch 10, total_step 161120, total loss is 16.51 , inference loss is 9.79, weight deacy loss is 6.71, training accuracy is 0.500000, time 212.109 samples/sec
+epoch 10, total_step 161140, total loss is 16.11 , inference loss is 9.40, weight deacy loss is 6.71, training accuracy is 0.375000, time 223.619 samples/sec
+epoch 10, total_step 161160, total loss is 13.47 , inference loss is 6.76, weight deacy loss is 6.71, training accuracy is 0.531250, time 225.714 samples/sec
+epoch 10, total_step 161180, total loss is 11.10 , inference loss is 4.39, weight deacy loss is 6.71, training accuracy is 0.531250, time 214.043 samples/sec
+epoch 10, total_step 161200, total loss is 13.88 , inference loss is 7.16, weight deacy loss is 6.71, training accuracy is 0.375000, time 219.376 samples/sec
+epoch 10, total_step 161220, total loss is 12.65 , inference loss is 5.94, weight deacy loss is 6.71, training accuracy is 0.468750, time 217.993 samples/sec
+epoch 10, total_step 161240, total loss is 13.27 , inference loss is 6.56, weight deacy loss is 6.71, training accuracy is 0.500000, time 213.915 samples/sec
+epoch 10, total_step 161260, total loss is 15.25 , inference loss is 8.54, weight deacy loss is 6.71, training accuracy is 0.500000, time 223.682 samples/sec
+epoch 10, total_step 161280, total loss is 19.44 , inference loss is 12.73, weight deacy loss is 6.71, training accuracy is 0.312500, time 212.640 samples/sec
+epoch 10, total_step 161300, total loss is 14.15 , inference loss is 7.44, weight deacy loss is 6.71, training accuracy is 0.343750, time 216.206 samples/sec
+epoch 10, total_step 161320, total loss is 17.58 , inference loss is 10.87, weight deacy loss is 6.71, training accuracy is 0.343750, time 214.542 samples/sec
+epoch 10, total_step 161340, total loss is 13.97 , inference loss is 7.26, weight deacy loss is 6.71, training accuracy is 0.312500, time 211.605 samples/sec
+epoch 10, total_step 161360, total loss is 15.20 , inference loss is 8.49, weight deacy loss is 6.71, training accuracy is 0.343750, time 219.360 samples/sec
+epoch 10, total_step 161380, total loss is 14.20 , inference loss is 7.48, weight deacy loss is 6.71, training accuracy is 0.531250, time 209.970 samples/sec
+epoch 10, total_step 161400, total loss is 12.87 , inference loss is 6.16, weight deacy loss is 6.71, training accuracy is 0.531250, time 11.995 samples/sec
+epoch 10, total_step 161420, total loss is 14.68 , inference loss is 7.97, weight deacy loss is 6.71, training accuracy is 0.468750, time 216.084 samples/sec
+epoch 10, total_step 161440, total loss is 13.43 , inference loss is 6.72, weight deacy loss is 6.71, training accuracy is 0.593750, time 209.733 samples/sec
+epoch 10, total_step 161460, total loss is 14.65 , inference loss is 7.95, weight deacy loss is 6.71, training accuracy is 0.468750, time 230.138 samples/sec
+epoch 10, total_step 161480, total loss is 12.74 , inference loss is 6.03, weight deacy loss is 6.71, training accuracy is 0.500000, time 217.346 samples/sec
+epoch 10, total_step 161500, total loss is 18.04 , inference loss is 11.33, weight deacy loss is 6.71, training accuracy is 0.187500, time 216.625 samples/sec
+epoch 10, total_step 161520, total loss is 13.33 , inference loss is 6.62, weight deacy loss is 6.71, training accuracy is 0.406250, time 209.709 samples/sec
+epoch 10, total_step 161540, total loss is 16.09 , inference loss is 9.38, weight deacy loss is 6.71, training accuracy is 0.437500, time 229.161 samples/sec
+epoch 10, total_step 161560, total loss is 12.97 , inference loss is 6.26, weight deacy loss is 6.71, training accuracy is 0.593750, time 212.269 samples/sec
+epoch 10, total_step 161580, total loss is 14.35 , inference loss is 7.65, weight deacy loss is 6.71, training accuracy is 0.500000, time 220.821 samples/sec
+epoch 10, total_step 161600, total loss is 15.35 , inference loss is 8.64, weight deacy loss is 6.71, training accuracy is 0.468750, time 216.735 samples/sec
+epoch 10, total_step 161620, total loss is 11.40 , inference loss is 4.70, weight deacy loss is 6.71, training accuracy is 0.562500, time 209.379 samples/sec
+epoch 10, total_step 161640, total loss is 13.99 , inference loss is 7.29, weight deacy loss is 6.71, training accuracy is 0.437500, time 213.718 samples/sec
+epoch 10, total_step 161660, total loss is 12.83 , inference loss is 6.13, weight deacy loss is 6.71, training accuracy is 0.437500, time 208.694 samples/sec
+epoch 10, total_step 161680, total loss is 16.37 , inference loss is 9.67, weight deacy loss is 6.71, training accuracy is 0.156250, time 298.506 samples/sec
+epoch 10, total_step 161700, total loss is 13.86 , inference loss is 7.16, weight deacy loss is 6.71, training accuracy is 0.406250, time 12.456 samples/sec
+epoch 10, total_step 161720, total loss is 14.15 , inference loss is 7.44, weight deacy loss is 6.71, training accuracy is 0.437500, time 213.874 samples/sec
+epoch 10, total_step 161740, total loss is 12.28 , inference loss is 5.58, weight deacy loss is 6.71, training accuracy is 0.500000, time 213.991 samples/sec
+epoch 10, total_step 161760, total loss is 19.66 , inference loss is 12.95, weight deacy loss is 6.71, training accuracy is 0.343750, time 215.024 samples/sec
+epoch 10, total_step 161780, total loss is 17.08 , inference loss is 10.37, weight deacy loss is 6.71, training accuracy is 0.343750, time 213.565 samples/sec
+epoch 10, total_step 161800, total loss is 11.58 , inference loss is 4.87, weight deacy loss is 6.71, training accuracy is 0.468750, time 210.066 samples/sec
+epoch 10, total_step 161820, total loss is 12.24 , inference loss is 5.54, weight deacy loss is 6.70, training accuracy is 0.468750, time 212.661 samples/sec
+epoch 10, total_step 161840, total loss is 14.90 , inference loss is 8.19, weight deacy loss is 6.70, training accuracy is 0.312500, time 210.982 samples/sec
+epoch 10, total_step 161860, total loss is 11.85 , inference loss is 5.15, weight deacy loss is 6.70, training accuracy is 0.500000, time 209.798 samples/sec
+epoch 10, total_step 161880, total loss is 14.39 , inference loss is 7.69, weight deacy loss is 6.70, training accuracy is 0.500000, time 214.288 samples/sec
+epoch 10, total_step 161900, total loss is 12.79 , inference loss is 6.08, weight deacy loss is 6.70, training accuracy is 0.500000, time 205.177 samples/sec
+epoch 10, total_step 161920, total loss is 15.94 , inference loss is 9.23, weight deacy loss is 6.70, training accuracy is 0.375000, time 217.814 samples/sec
+epoch 10, total_step 161940, total loss is 17.90 , inference loss is 11.19, weight deacy loss is 6.70, training accuracy is 0.468750, time 299.752 samples/sec
+epoch 10, total_step 161960, total loss is 18.92 , inference loss is 12.22, weight deacy loss is 6.70, training accuracy is 0.312500, time 214.532 samples/sec
+epoch 10, total_step 161980, total loss is 13.14 , inference loss is 6.43, weight deacy loss is 6.70, training accuracy is 0.531250, time 214.213 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.742124999999998
+best_threshold_index 114 0.9879629629629629
+best_threshold_index 112 0.9864814814814815
+best_threshold_index 112 0.9874074074074074
+best_threshold_index 112 0.9883333333333333
+best_threshold_index 112 0.9875925925925926
+best_threshold_index 112 0.9861111111111112
+best_threshold_index 112 0.9877777777777778
+best_threshold_index 114 0.9866666666666667
+best_threshold_index 112 0.9861111111111112
+best_threshold_index 112 0.9862962962962963
+[lfw][162000]XNorm: 22.605432
+[lfw][162000]Accuracy-Flip: 0.98633+-0.00710
+test accuracy is: 0.9863333333333333
+epoch 10, total_step 162000, total loss is 16.78 , inference loss is 10.07, weight deacy loss is 6.70, training accuracy is 0.281250, time 16.141 samples/sec
+epoch 10, total_step 162020, total loss is 16.17 , inference loss is 9.47, weight deacy loss is 6.70, training accuracy is 0.312500, time 216.101 samples/sec
+epoch 10, total_step 162040, total loss is 18.87 , inference loss is 12.17, weight deacy loss is 6.70, training accuracy is 0.343750, time 211.532 samples/sec
+epoch 10, total_step 162060, total loss is 9.62 , inference loss is 2.92, weight deacy loss is 6.70, training accuracy is 0.718750, time 213.664 samples/sec
+epoch 10, total_step 162080, total loss is 10.75 , inference loss is 4.05, weight deacy loss is 6.70, training accuracy is 0.500000, time 214.208 samples/sec
+epoch 10, total_step 162100, total loss is 17.96 , inference loss is 11.26, weight deacy loss is 6.70, training accuracy is 0.437500, time 213.851 samples/sec
+epoch 10, total_step 162120, total loss is 17.30 , inference loss is 10.60, weight deacy loss is 6.70, training accuracy is 0.406250, time 217.095 samples/sec
+epoch 10, total_step 162140, total loss is 12.98 , inference loss is 6.28, weight deacy loss is 6.70, training accuracy is 0.531250, time 211.734 samples/sec
+epoch 10, total_step 162160, total loss is 18.92 , inference loss is 12.21, weight deacy loss is 6.70, training accuracy is 0.343750, time 210.950 samples/sec
+epoch 10, total_step 162180, total loss is 14.93 , inference loss is 8.23, weight deacy loss is 6.70, training accuracy is 0.375000, time 219.016 samples/sec
+epoch 10, total_step 162200, total loss is 13.39 , inference loss is 6.69, weight deacy loss is 6.70, training accuracy is 0.500000, time 208.216 samples/sec
+epoch 10, total_step 162220, total loss is 18.34 , inference loss is 11.64, weight deacy loss is 6.70, training accuracy is 0.343750, time 184.238 samples/sec
+epoch 10, total_step 162240, total loss is 14.35 , inference loss is 7.65, weight deacy loss is 6.70, training accuracy is 0.468750, time 213.911 samples/sec
+epoch 10, total_step 162260, total loss is 19.08 , inference loss is 12.38, weight deacy loss is 6.70, training accuracy is 0.250000, time 206.040 samples/sec
+epoch 10, total_step 162280, total loss is 11.74 , inference loss is 5.04, weight deacy loss is 6.70, training accuracy is 0.406250, time 298.375 samples/sec
+epoch 10, total_step 162300, total loss is 15.85 , inference loss is 9.15, weight deacy loss is 6.70, training accuracy is 0.406250, time 11.750 samples/sec
+epoch 10, total_step 162320, total loss is 13.88 , inference loss is 7.19, weight deacy loss is 6.70, training accuracy is 0.531250, time 219.740 samples/sec
+epoch 10, total_step 162340, total loss is 15.60 , inference loss is 8.90, weight deacy loss is 6.70, training accuracy is 0.343750, time 209.348 samples/sec
+epoch 10, total_step 162360, total loss is 15.86 , inference loss is 9.16, weight deacy loss is 6.70, training accuracy is 0.531250, time 211.017 samples/sec
+epoch 10, total_step 162380, total loss is 15.56 , inference loss is 8.86, weight deacy loss is 6.70, training accuracy is 0.437500, time 209.426 samples/sec
+epoch 10, total_step 162400, total loss is 16.88 , inference loss is 10.18, weight deacy loss is 6.70, training accuracy is 0.406250, time 215.444 samples/sec
+epoch 10, total_step 162420, total loss is 12.15 , inference loss is 5.45, weight deacy loss is 6.70, training accuracy is 0.468750, time 215.995 samples/sec
+epoch 10, total_step 162440, total loss is 16.44 , inference loss is 9.74, weight deacy loss is 6.70, training accuracy is 0.437500, time 209.156 samples/sec
+epoch 10, total_step 162460, total loss is 15.75 , inference loss is 9.05, weight deacy loss is 6.70, training accuracy is 0.468750, time 215.048 samples/sec
+epoch 10, total_step 162480, total loss is 20.62 , inference loss is 13.93, weight deacy loss is 6.70, training accuracy is 0.375000, time 215.086 samples/sec
+epoch 10, total_step 162500, total loss is 14.19 , inference loss is 7.49, weight deacy loss is 6.70, training accuracy is 0.437500, time 204.435 samples/sec
+epoch 10, total_step 162520, total loss is 13.88 , inference loss is 7.18, weight deacy loss is 6.70, training accuracy is 0.500000, time 217.363 samples/sec
+epoch 10, total_step 162540, total loss is 17.79 , inference loss is 11.10, weight deacy loss is 6.70, training accuracy is 0.406250, time 300.421 samples/sec
+epoch 10, total_step 162560, total loss is 14.76 , inference loss is 8.06, weight deacy loss is 6.70, training accuracy is 0.531250, time 228.821 samples/sec
+epoch 10, total_step 162580, total loss is 18.70 , inference loss is 12.00, weight deacy loss is 6.70, training accuracy is 0.437500, time 219.150 samples/sec
+epoch 10, total_step 162600, total loss is 18.29 , inference loss is 11.60, weight deacy loss is 6.70, training accuracy is 0.437500, time 12.990 samples/sec
+epoch 10, total_step 162620, total loss is 14.13 , inference loss is 7.44, weight deacy loss is 6.70, training accuracy is 0.531250, time 222.417 samples/sec
+epoch 10, total_step 162640, total loss is 16.19 , inference loss is 9.49, weight deacy loss is 6.70, training accuracy is 0.406250, time 216.188 samples/sec
+epoch 10, total_step 162660, total loss is 17.14 , inference loss is 10.44, weight deacy loss is 6.70, training accuracy is 0.281250, time 212.161 samples/sec
+epoch 10, total_step 162680, total loss is 13.01 , inference loss is 6.32, weight deacy loss is 6.70, training accuracy is 0.500000, time 216.818 samples/sec
+epoch 10, total_step 162700, total loss is 10.85 , inference loss is 4.15, weight deacy loss is 6.69, training accuracy is 0.656250, time 217.944 samples/sec
+epoch 10, total_step 162720, total loss is 12.80 , inference loss is 6.10, weight deacy loss is 6.69, training accuracy is 0.406250, time 216.484 samples/sec
+epoch 10, total_step 162740, total loss is 14.93 , inference loss is 8.23, weight deacy loss is 6.69, training accuracy is 0.468750, time 217.533 samples/sec
+epoch 10, total_step 162760, total loss is 16.05 , inference loss is 9.35, weight deacy loss is 6.69, training accuracy is 0.468750, time 220.250 samples/sec
+epoch 10, total_step 162780, total loss is 13.17 , inference loss is 6.47, weight deacy loss is 6.69, training accuracy is 0.531250, time 217.386 samples/sec
+epoch 10, total_step 162800, total loss is 12.60 , inference loss is 5.90, weight deacy loss is 6.69, training accuracy is 0.437500, time 213.723 samples/sec
+epoch 10, total_step 162820, total loss is 14.81 , inference loss is 8.11, weight deacy loss is 6.69, training accuracy is 0.468750, time 220.071 samples/sec
+epoch 10, total_step 162840, total loss is 13.16 , inference loss is 6.47, weight deacy loss is 6.69, training accuracy is 0.468750, time 231.532 samples/sec
+epoch 10, total_step 162860, total loss is 14.43 , inference loss is 7.73, weight deacy loss is 6.69, training accuracy is 0.468750, time 300.921 samples/sec
+epoch 10, total_step 162880, total loss is 13.66 , inference loss is 6.97, weight deacy loss is 6.69, training accuracy is 0.406250, time 301.322 samples/sec
+epoch 10, total_step 162900, total loss is 17.68 , inference loss is 10.99, weight deacy loss is 6.69, training accuracy is 0.468750, time 13.279 samples/sec
+epoch 10, total_step 162920, total loss is 17.46 , inference loss is 10.77, weight deacy loss is 6.69, training accuracy is 0.343750, time 218.004 samples/sec
+epoch 10, total_step 162940, total loss is 19.12 , inference loss is 12.42, weight deacy loss is 6.69, training accuracy is 0.343750, time 299.002 samples/sec
+epoch 10, total_step 162960, total loss is 14.45 , inference loss is 7.76, weight deacy loss is 6.69, training accuracy is 0.500000, time 228.821 samples/sec
+epoch 10, total_step 162980, total loss is 11.64 , inference loss is 4.95, weight deacy loss is 6.69, training accuracy is 0.593750, time 217.860 samples/sec
+epoch 10, total_step 163000, total loss is 12.63 , inference loss is 5.94, weight deacy loss is 6.69, training accuracy is 0.562500, time 220.592 samples/sec
+epoch 10, total_step 163020, total loss is 14.41 , inference loss is 7.72, weight deacy loss is 6.69, training accuracy is 0.468750, time 218.808 samples/sec
+epoch 10, total_step 163040, total loss is 14.55 , inference loss is 7.86, weight deacy loss is 6.69, training accuracy is 0.531250, time 207.851 samples/sec
+epoch 10, total_step 163060, total loss is 13.63 , inference loss is 6.94, weight deacy loss is 6.69, training accuracy is 0.562500, time 221.857 samples/sec
+epoch 10, total_step 163080, total loss is 10.13 , inference loss is 3.44, weight deacy loss is 6.69, training accuracy is 0.531250, time 219.843 samples/sec
+epoch 10, total_step 163100, total loss is 11.56 , inference loss is 4.87, weight deacy loss is 6.69, training accuracy is 0.531250, time 215.583 samples/sec
+epoch 10, total_step 163120, total loss is 14.31 , inference loss is 7.62, weight deacy loss is 6.69, training accuracy is 0.531250, time 220.858 samples/sec
+epoch 10, total_step 163140, total loss is 14.92 , inference loss is 8.23, weight deacy loss is 6.69, training accuracy is 0.437500, time 219.747 samples/sec
+epoch 10, total_step 163160, total loss is 17.73 , inference loss is 11.04, weight deacy loss is 6.69, training accuracy is 0.281250, time 194.221 samples/sec
+epoch 10, total_step 163180, total loss is 15.06 , inference loss is 8.37, weight deacy loss is 6.69, training accuracy is 0.312500, time 201.666 samples/sec
+epoch 10, total_step 163200, total loss is 12.27 , inference loss is 5.58, weight deacy loss is 6.69, training accuracy is 0.531250, time 12.658 samples/sec
+epoch 10, total_step 163220, total loss is 12.31 , inference loss is 5.62, weight deacy loss is 6.69, training accuracy is 0.562500, time 217.052 samples/sec
+epoch 10, total_step 163240, total loss is 16.41 , inference loss is 9.72, weight deacy loss is 6.69, training accuracy is 0.468750, time 214.429 samples/sec
+epoch 10, total_step 163260, total loss is 15.88 , inference loss is 9.19, weight deacy loss is 6.69, training accuracy is 0.312500, time 214.070 samples/sec
+epoch 10, total_step 163280, total loss is 13.38 , inference loss is 6.69, weight deacy loss is 6.69, training accuracy is 0.406250, time 215.130 samples/sec
+epoch 10, total_step 163300, total loss is 16.07 , inference loss is 9.38, weight deacy loss is 6.69, training accuracy is 0.500000, time 214.987 samples/sec
+epoch 10, total_step 163320, total loss is 10.16 , inference loss is 3.47, weight deacy loss is 6.69, training accuracy is 0.593750, time 211.760 samples/sec
+epoch 10, total_step 163340, total loss is 12.38 , inference loss is 5.69, weight deacy loss is 6.69, training accuracy is 0.406250, time 211.654 samples/sec
+epoch 10, total_step 163360, total loss is 11.13 , inference loss is 4.45, weight deacy loss is 6.69, training accuracy is 0.625000, time 218.607 samples/sec
+epoch 10, total_step 163380, total loss is 11.99 , inference loss is 5.30, weight deacy loss is 6.69, training accuracy is 0.500000, time 207.202 samples/sec
+epoch 10, total_step 163400, total loss is 12.75 , inference loss is 6.06, weight deacy loss is 6.69, training accuracy is 0.343750, time 212.056 samples/sec
+epoch 10, total_step 163420, total loss is 12.76 , inference loss is 6.08, weight deacy loss is 6.69, training accuracy is 0.468750, time 207.897 samples/sec
+epoch 10, total_step 163440, total loss is 12.29 , inference loss is 5.61, weight deacy loss is 6.69, training accuracy is 0.500000, time 212.450 samples/sec
+epoch 10, total_step 163460, total loss is 15.95 , inference loss is 9.26, weight deacy loss is 6.69, training accuracy is 0.500000, time 230.090 samples/sec
+epoch 10, total_step 163480, total loss is 14.19 , inference loss is 7.51, weight deacy loss is 6.69, training accuracy is 0.375000, time 210.424 samples/sec
+epoch 10, total_step 163500, total loss is 15.48 , inference loss is 8.80, weight deacy loss is 6.69, training accuracy is 0.250000, time 12.424 samples/sec
+epoch 10, total_step 163520, total loss is 15.69 , inference loss is 9.00, weight deacy loss is 6.68, training accuracy is 0.406250, time 209.918 samples/sec
+epoch 10, total_step 163540, total loss is 12.94 , inference loss is 6.25, weight deacy loss is 6.68, training accuracy is 0.500000, time 207.193 samples/sec
+epoch 10, total_step 163560, total loss is 12.07 , inference loss is 5.39, weight deacy loss is 6.68, training accuracy is 0.468750, time 222.850 samples/sec
+epoch 10, total_step 163580, total loss is 14.56 , inference loss is 7.88, weight deacy loss is 6.68, training accuracy is 0.468750, time 233.832 samples/sec
+epoch 10, total_step 163600, total loss is 13.13 , inference loss is 6.45, weight deacy loss is 6.68, training accuracy is 0.593750, time 226.185 samples/sec
+epoch 10, total_step 163620, total loss is 15.84 , inference loss is 9.16, weight deacy loss is 6.68, training accuracy is 0.468750, time 216.266 samples/sec
+epoch 10, total_step 163640, total loss is 14.93 , inference loss is 8.24, weight deacy loss is 6.68, training accuracy is 0.500000, time 214.703 samples/sec
+epoch 10, total_step 163660, total loss is 19.51 , inference loss is 12.82, weight deacy loss is 6.68, training accuracy is 0.281250, time 211.446 samples/sec
+epoch 10, total_step 163680, total loss is 11.76 , inference loss is 5.08, weight deacy loss is 6.68, training accuracy is 0.625000, time 212.669 samples/sec
+epoch 10, total_step 163700, total loss is 13.35 , inference loss is 6.67, weight deacy loss is 6.68, training accuracy is 0.593750, time 217.434 samples/sec
+epoch 10, total_step 163720, total loss is 13.94 , inference loss is 7.26, weight deacy loss is 6.68, training accuracy is 0.468750, time 200.606 samples/sec
+epoch 10, total_step 163740, total loss is 18.38 , inference loss is 11.69, weight deacy loss is 6.68, training accuracy is 0.343750, time 209.110 samples/sec
+epoch 10, total_step 163760, total loss is 16.88 , inference loss is 10.20, weight deacy loss is 6.68, training accuracy is 0.312500, time 216.126 samples/sec
+epoch 10, total_step 163780, total loss is 12.04 , inference loss is 5.36, weight deacy loss is 6.68, training accuracy is 0.500000, time 298.195 samples/sec
+epoch 10, total_step 163800, total loss is 11.77 , inference loss is 5.09, weight deacy loss is 6.68, training accuracy is 0.593750, time 12.579 samples/sec
+epoch 10, total_step 163820, total loss is 12.11 , inference loss is 5.42, weight deacy loss is 6.68, training accuracy is 0.593750, time 214.543 samples/sec
+epoch 10, total_step 163840, total loss is 9.95 , inference loss is 3.27, weight deacy loss is 6.68, training accuracy is 0.593750, time 213.848 samples/sec
+epoch 10, total_step 163860, total loss is 13.35 , inference loss is 6.67, weight deacy loss is 6.68, training accuracy is 0.437500, time 215.407 samples/sec
+epoch 10, total_step 163880, total loss is 13.46 , inference loss is 6.78, weight deacy loss is 6.68, training accuracy is 0.500000, time 218.233 samples/sec
+epoch 10, total_step 163900, total loss is 12.32 , inference loss is 5.64, weight deacy loss is 6.68, training accuracy is 0.500000, time 217.887 samples/sec
+epoch 10, total_step 163920, total loss is 10.49 , inference loss is 3.81, weight deacy loss is 6.68, training accuracy is 0.593750, time 212.871 samples/sec
+epoch 10, total_step 163940, total loss is 12.72 , inference loss is 6.04, weight deacy loss is 6.68, training accuracy is 0.531250, time 212.049 samples/sec
+epoch 10, total_step 163960, total loss is 14.77 , inference loss is 8.09, weight deacy loss is 6.68, training accuracy is 0.468750, time 211.949 samples/sec
+epoch 10, total_step 163980, total loss is 15.21 , inference loss is 8.53, weight deacy loss is 6.68, training accuracy is 0.500000, time 225.426 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.088353000000037
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 111 0.9857407407407407
+best_threshold_index 111 0.9862962962962963
+best_threshold_index 108 0.9864814814814815
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 111 0.985
+best_threshold_index 111 0.9861111111111112
+best_threshold_index 111 0.9857407407407407
+best_threshold_index 111 0.9844444444444445
+best_threshold_index 111 0.9851851851851852
+[lfw][164000]XNorm: 21.339537
+[lfw][164000]Accuracy-Flip: 0.98550+-0.00683
+test accuracy is: 0.9855
+epoch 10, total_step 164000, total loss is 10.57 , inference loss is 3.89, weight deacy loss is 6.68, training accuracy is 0.531250, time 15.878 samples/sec
+epoch 10, total_step 164020, total loss is 11.70 , inference loss is 5.03, weight deacy loss is 6.68, training accuracy is 0.406250, time 212.744 samples/sec
+epoch 10, total_step 164040, total loss is 18.26 , inference loss is 11.58, weight deacy loss is 6.68, training accuracy is 0.312500, time 214.105 samples/sec
+epoch 10, total_step 164060, total loss is 14.88 , inference loss is 8.20, weight deacy loss is 6.68, training accuracy is 0.500000, time 221.216 samples/sec
+epoch 10, total_step 164080, total loss is 19.93 , inference loss is 13.25, weight deacy loss is 6.68, training accuracy is 0.500000, time 219.825 samples/sec
+epoch 10, total_step 164100, total loss is 12.78 , inference loss is 6.10, weight deacy loss is 6.68, training accuracy is 0.468750, time 12.509 samples/sec
+epoch 10, total_step 164120, total loss is 15.59 , inference loss is 8.91, weight deacy loss is 6.68, training accuracy is 0.375000, time 217.595 samples/sec
+epoch 10, total_step 164140, total loss is 12.93 , inference loss is 6.25, weight deacy loss is 6.68, training accuracy is 0.406250, time 216.415 samples/sec
+epoch 10, total_step 164160, total loss is 15.77 , inference loss is 9.09, weight deacy loss is 6.68, training accuracy is 0.406250, time 216.545 samples/sec
+epoch 10, total_step 164180, total loss is 13.97 , inference loss is 7.29, weight deacy loss is 6.68, training accuracy is 0.531250, time 214.443 samples/sec
+epoch 10, total_step 164200, total loss is 12.83 , inference loss is 6.15, weight deacy loss is 6.68, training accuracy is 0.500000, time 213.885 samples/sec
+epoch 10, total_step 164220, total loss is 16.58 , inference loss is 9.90, weight deacy loss is 6.68, training accuracy is 0.343750, time 213.375 samples/sec
+epoch 10, total_step 164240, total loss is 14.37 , inference loss is 7.70, weight deacy loss is 6.68, training accuracy is 0.375000, time 212.790 samples/sec
+epoch 10, total_step 164260, total loss is 12.98 , inference loss is 6.31, weight deacy loss is 6.68, training accuracy is 0.437500, time 223.597 samples/sec
+epoch 10, total_step 164280, total loss is 12.34 , inference loss is 5.67, weight deacy loss is 6.68, training accuracy is 0.531250, time 221.063 samples/sec
+epoch 10, total_step 164300, total loss is 15.62 , inference loss is 8.95, weight deacy loss is 6.68, training accuracy is 0.437500, time 237.216 samples/sec
+epoch 10, total_step 164320, total loss is 17.71 , inference loss is 11.03, weight deacy loss is 6.68, training accuracy is 0.406250, time 208.637 samples/sec
+epoch 10, total_step 164340, total loss is 18.68 , inference loss is 12.01, weight deacy loss is 6.68, training accuracy is 0.281250, time 222.000 samples/sec
+epoch 10, total_step 164360, total loss is 12.39 , inference loss is 5.71, weight deacy loss is 6.67, training accuracy is 0.406250, time 223.025 samples/sec
+epoch 10, total_step 164380, total loss is 13.13 , inference loss is 6.45, weight deacy loss is 6.67, training accuracy is 0.593750, time 246.661 samples/sec
+epoch 10, total_step 164400, total loss is 14.27 , inference loss is 7.60, weight deacy loss is 6.67, training accuracy is 0.375000, time 12.065 samples/sec
+epoch 10, total_step 164420, total loss is 15.44 , inference loss is 8.77, weight deacy loss is 6.67, training accuracy is 0.468750, time 215.053 samples/sec
+epoch 10, total_step 164440, total loss is 12.23 , inference loss is 5.55, weight deacy loss is 6.67, training accuracy is 0.375000, time 299.271 samples/sec
+epoch 10, total_step 164460, total loss is 13.75 , inference loss is 7.08, weight deacy loss is 6.67, training accuracy is 0.531250, time 214.573 samples/sec
+epoch 10, total_step 164480, total loss is 16.15 , inference loss is 9.47, weight deacy loss is 6.67, training accuracy is 0.437500, time 216.753 samples/sec
+epoch 10, total_step 164500, total loss is 14.29 , inference loss is 7.62, weight deacy loss is 6.67, training accuracy is 0.406250, time 225.439 samples/sec
+epoch 10, total_step 164520, total loss is 14.28 , inference loss is 7.61, weight deacy loss is 6.67, training accuracy is 0.281250, time 217.760 samples/sec
+epoch 10, total_step 164540, total loss is 15.19 , inference loss is 8.52, weight deacy loss is 6.67, training accuracy is 0.468750, time 215.237 samples/sec
+epoch 10, total_step 164560, total loss is 10.84 , inference loss is 4.17, weight deacy loss is 6.67, training accuracy is 0.531250, time 214.454 samples/sec
+epoch 10, total_step 164580, total loss is 13.68 , inference loss is 7.01, weight deacy loss is 6.67, training accuracy is 0.468750, time 220.706 samples/sec
+epoch 10, total_step 164600, total loss is 12.29 , inference loss is 5.62, weight deacy loss is 6.67, training accuracy is 0.468750, time 296.519 samples/sec
+epoch 10, total_step 164620, total loss is 15.07 , inference loss is 8.40, weight deacy loss is 6.67, training accuracy is 0.343750, time 299.158 samples/sec
+epoch 10, total_step 164640, total loss is 14.49 , inference loss is 7.82, weight deacy loss is 6.67, training accuracy is 0.437500, time 211.817 samples/sec
+epoch 10, total_step 164660, total loss is 12.71 , inference loss is 6.04, weight deacy loss is 6.67, training accuracy is 0.437500, time 215.891 samples/sec
+epoch 10, total_step 164680, total loss is 14.77 , inference loss is 8.10, weight deacy loss is 6.67, training accuracy is 0.437500, time 221.258 samples/sec
+epoch 10, total_step 164700, total loss is 15.46 , inference loss is 8.78, weight deacy loss is 6.67, training accuracy is 0.531250, time 13.323 samples/sec
+epoch 10, total_step 164720, total loss is 16.23 , inference loss is 9.56, weight deacy loss is 6.67, training accuracy is 0.375000, time 213.369 samples/sec
+epoch 10, total_step 164740, total loss is 12.00 , inference loss is 5.33, weight deacy loss is 6.67, training accuracy is 0.468750, time 228.357 samples/sec
+epoch 10, total_step 164760, total loss is 12.84 , inference loss is 6.17, weight deacy loss is 6.67, training accuracy is 0.437500, time 208.414 samples/sec
+epoch 10, total_step 164780, total loss is 14.40 , inference loss is 7.73, weight deacy loss is 6.67, training accuracy is 0.343750, time 205.239 samples/sec
+epoch 10, total_step 164800, total loss is 16.16 , inference loss is 9.49, weight deacy loss is 6.67, training accuracy is 0.312500, time 275.906 samples/sec
+epoch 10, total_step 164820, total loss is 18.41 , inference loss is 11.74, weight deacy loss is 6.67, training accuracy is 0.312500, time 209.902 samples/sec
+epoch 10, total_step 164840, total loss is 16.27 , inference loss is 9.60, weight deacy loss is 6.67, training accuracy is 0.343750, time 224.999 samples/sec
+epoch 10, total_step 164860, total loss is 16.42 , inference loss is 9.75, weight deacy loss is 6.67, training accuracy is 0.281250, time 212.809 samples/sec
+epoch 10, total_step 164880, total loss is 15.27 , inference loss is 8.60, weight deacy loss is 6.67, training accuracy is 0.343750, time 221.702 samples/sec
+epoch 10, total_step 164900, total loss is 14.52 , inference loss is 7.86, weight deacy loss is 6.67, training accuracy is 0.437500, time 214.596 samples/sec
+epoch 10, total_step 164920, total loss is 15.30 , inference loss is 8.63, weight deacy loss is 6.67, training accuracy is 0.312500, time 219.632 samples/sec
+epoch 10, total_step 164940, total loss is 15.29 , inference loss is 8.62, weight deacy loss is 6.67, training accuracy is 0.437500, time 219.775 samples/sec
+epoch 10, total_step 164960, total loss is 16.87 , inference loss is 10.20, weight deacy loss is 6.67, training accuracy is 0.218750, time 217.202 samples/sec
+epoch 10, total_step 164980, total loss is 13.82 , inference loss is 7.15, weight deacy loss is 6.67, training accuracy is 0.406250, time 212.031 samples/sec
+epoch 10, total_step 165000, total loss is 11.68 , inference loss is 5.01, weight deacy loss is 6.67, training accuracy is 0.531250, time 12.693 samples/sec
+epoch 10, total_step 165020, total loss is 16.35 , inference loss is 9.68, weight deacy loss is 6.67, training accuracy is 0.562500, time 230.223 samples/sec
+epoch 10, total_step 165040, total loss is 16.52 , inference loss is 9.85, weight deacy loss is 6.67, training accuracy is 0.468750, time 214.795 samples/sec
+epoch 10, total_step 165060, total loss is 18.02 , inference loss is 11.35, weight deacy loss is 6.67, training accuracy is 0.343750, time 299.659 samples/sec
+epoch 10, total_step 165080, total loss is 16.30 , inference loss is 9.63, weight deacy loss is 6.67, training accuracy is 0.343750, time 223.795 samples/sec
+epoch 10, total_step 165100, total loss is 18.01 , inference loss is 11.34, weight deacy loss is 6.67, training accuracy is 0.468750, time 223.248 samples/sec
+epoch 10, total_step 165120, total loss is 15.75 , inference loss is 9.09, weight deacy loss is 6.67, training accuracy is 0.437500, time 220.236 samples/sec
+epoch 10, total_step 165140, total loss is 14.68 , inference loss is 8.01, weight deacy loss is 6.66, training accuracy is 0.343750, time 212.953 samples/sec
+epoch 10, total_step 165160, total loss is 15.58 , inference loss is 8.92, weight deacy loss is 6.66, training accuracy is 0.343750, time 214.124 samples/sec
+epoch 10, total_step 165180, total loss is 16.64 , inference loss is 9.97, weight deacy loss is 6.66, training accuracy is 0.375000, time 225.327 samples/sec
+epoch 10, total_step 165200, total loss is 14.99 , inference loss is 8.33, weight deacy loss is 6.66, training accuracy is 0.406250, time 222.875 samples/sec
+epoch 10, total_step 165220, total loss is 14.02 , inference loss is 7.36, weight deacy loss is 6.66, training accuracy is 0.281250, time 221.074 samples/sec
+epoch 10, total_step 165240, total loss is 12.13 , inference loss is 5.46, weight deacy loss is 6.66, training accuracy is 0.468750, time 214.634 samples/sec
+epoch 10, total_step 165260, total loss is 15.97 , inference loss is 9.31, weight deacy loss is 6.66, training accuracy is 0.375000, time 213.402 samples/sec
+epoch 10, total_step 165280, total loss is 18.44 , inference loss is 11.78, weight deacy loss is 6.66, training accuracy is 0.312500, time 212.668 samples/sec
+epoch 10, total_step 165300, total loss is 13.59 , inference loss is 6.92, weight deacy loss is 6.66, training accuracy is 0.406250, time 12.865 samples/sec
+epoch 10, total_step 165320, total loss is 13.58 , inference loss is 6.92, weight deacy loss is 6.66, training accuracy is 0.375000, time 299.799 samples/sec
+epoch 10, total_step 165340, total loss is 15.88 , inference loss is 9.22, weight deacy loss is 6.66, training accuracy is 0.375000, time 217.375 samples/sec
+epoch 10, total_step 165360, total loss is 20.44 , inference loss is 13.78, weight deacy loss is 6.66, training accuracy is 0.281250, time 216.969 samples/sec
+epoch 10, total_step 165380, total loss is 15.79 , inference loss is 9.12, weight deacy loss is 6.66, training accuracy is 0.281250, time 224.590 samples/sec
+epoch 10, total_step 165400, total loss is 18.30 , inference loss is 11.64, weight deacy loss is 6.66, training accuracy is 0.312500, time 222.620 samples/sec
+epoch 10, total_step 165420, total loss is 16.49 , inference loss is 9.83, weight deacy loss is 6.66, training accuracy is 0.406250, time 221.967 samples/sec
+epoch 10, total_step 165440, total loss is 10.64 , inference loss is 3.98, weight deacy loss is 6.66, training accuracy is 0.437500, time 214.067 samples/sec
+epoch 10, total_step 165460, total loss is 18.66 , inference loss is 12.00, weight deacy loss is 6.66, training accuracy is 0.156250, time 211.888 samples/sec
+epoch 10, total_step 165480, total loss is 16.98 , inference loss is 10.32, weight deacy loss is 6.66, training accuracy is 0.312500, time 216.532 samples/sec
+epoch 10, total_step 165500, total loss is 15.40 , inference loss is 8.74, weight deacy loss is 6.66, training accuracy is 0.437500, time 225.675 samples/sec
+epoch 10, total_step 165520, total loss is 13.58 , inference loss is 6.92, weight deacy loss is 6.66, training accuracy is 0.312500, time 220.190 samples/sec
+epoch 10, total_step 165540, total loss is 11.92 , inference loss is 5.26, weight deacy loss is 6.66, training accuracy is 0.437500, time 228.200 samples/sec
+epoch 10, total_step 165560, total loss is 12.67 , inference loss is 6.01, weight deacy loss is 6.66, training accuracy is 0.468750, time 219.421 samples/sec
+epoch 10, total_step 165580, total loss is 14.81 , inference loss is 8.15, weight deacy loss is 6.66, training accuracy is 0.312500, time 209.808 samples/sec
+epoch 10, total_step 165600, total loss is 17.51 , inference loss is 10.85, weight deacy loss is 6.66, training accuracy is 0.312500, time 12.207 samples/sec
+epoch 10, total_step 165620, total loss is 13.40 , inference loss is 6.74, weight deacy loss is 6.66, training accuracy is 0.437500, time 221.638 samples/sec
+epoch 10, total_step 165640, total loss is 17.89 , inference loss is 11.23, weight deacy loss is 6.66, training accuracy is 0.250000, time 217.411 samples/sec
+epoch 10, total_step 165660, total loss is 17.76 , inference loss is 11.10, weight deacy loss is 6.66, training accuracy is 0.406250, time 216.143 samples/sec
+epoch 10, total_step 165680, total loss is 11.49 , inference loss is 4.83, weight deacy loss is 6.66, training accuracy is 0.406250, time 213.493 samples/sec
+epoch 10, total_step 165700, total loss is 13.25 , inference loss is 6.60, weight deacy loss is 6.66, training accuracy is 0.437500, time 215.612 samples/sec
+epoch 10, total_step 165720, total loss is 13.39 , inference loss is 6.73, weight deacy loss is 6.66, training accuracy is 0.375000, time 221.136 samples/sec
+epoch 10, total_step 165740, total loss is 13.44 , inference loss is 6.78, weight deacy loss is 6.66, training accuracy is 0.437500, time 221.855 samples/sec
+epoch 10, total_step 165760, total loss is 16.40 , inference loss is 9.74, weight deacy loss is 6.66, training accuracy is 0.343750, time 220.990 samples/sec
+epoch 10, total_step 165780, total loss is 11.04 , inference loss is 4.39, weight deacy loss is 6.66, training accuracy is 0.562500, time 212.983 samples/sec
+epoch 10, total_step 165800, total loss is 16.78 , inference loss is 10.13, weight deacy loss is 6.66, training accuracy is 0.375000, time 217.498 samples/sec
+epoch 10, total_step 165820, total loss is 14.18 , inference loss is 7.52, weight deacy loss is 6.66, training accuracy is 0.375000, time 197.570 samples/sec
+epoch 10, total_step 165840, total loss is 14.66 , inference loss is 8.00, weight deacy loss is 6.66, training accuracy is 0.468750, time 203.397 samples/sec
+epoch 10, total_step 165860, total loss is 13.31 , inference loss is 6.65, weight deacy loss is 6.66, training accuracy is 0.281250, time 209.082 samples/sec
+epoch 10, total_step 165880, total loss is 13.48 , inference loss is 6.82, weight deacy loss is 6.66, training accuracy is 0.250000, time 206.145 samples/sec
+epoch 10, total_step 165900, total loss is 16.71 , inference loss is 10.06, weight deacy loss is 6.66, training accuracy is 0.281250, time 13.361 samples/sec
+epoch 10, total_step 165920, total loss is 18.76 , inference loss is 12.10, weight deacy loss is 6.65, training accuracy is 0.281250, time 214.872 samples/sec
+epoch 10, total_step 165940, total loss is 12.77 , inference loss is 6.12, weight deacy loss is 6.65, training accuracy is 0.343750, time 219.485 samples/sec
+epoch 10, total_step 165960, total loss is 14.54 , inference loss is 7.88, weight deacy loss is 6.65, training accuracy is 0.406250, time 221.739 samples/sec
+epoch 10, total_step 165980, total loss is 16.38 , inference loss is 9.73, weight deacy loss is 6.65, training accuracy is 0.343750, time 222.082 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.568762000000017
+best_threshold_index 110 0.9840740740740741
+best_threshold_index 110 0.9833333333333333
+best_threshold_index 110 0.9838888888888889
+best_threshold_index 110 0.9848148148148148
+best_threshold_index 110 0.9844444444444445
+best_threshold_index 110 0.9831481481481481
+best_threshold_index 110 0.9844444444444445
+best_threshold_index 110 0.9835185185185186
+best_threshold_index 110 0.9822222222222222
+best_threshold_index 110 0.9827777777777778
+[lfw][166000]XNorm: 23.164252
+[lfw][166000]Accuracy-Flip: 0.98367+-0.00698
+test accuracy is: 0.9836666666666668
+epoch 10, total_step 166000, total loss is 16.22 , inference loss is 9.57, weight deacy loss is 6.65, training accuracy is 0.250000, time 16.454 samples/sec
+epoch 10, total_step 166020, total loss is 13.89 , inference loss is 7.24, weight deacy loss is 6.65, training accuracy is 0.218750, time 214.271 samples/sec
+epoch 10, total_step 166040, total loss is 18.16 , inference loss is 11.51, weight deacy loss is 6.65, training accuracy is 0.218750, time 216.275 samples/sec
+epoch 10, total_step 166060, total loss is 17.73 , inference loss is 11.08, weight deacy loss is 6.65, training accuracy is 0.375000, time 224.442 samples/sec
+epoch 10, total_step 166080, total loss is 17.00 , inference loss is 10.35, weight deacy loss is 6.65, training accuracy is 0.343750, time 217.297 samples/sec
+epoch 10, total_step 166100, total loss is 12.78 , inference loss is 6.13, weight deacy loss is 6.65, training accuracy is 0.375000, time 211.261 samples/sec
+epoch 10, total_step 166120, total loss is 12.27 , inference loss is 5.62, weight deacy loss is 6.65, training accuracy is 0.343750, time 228.500 samples/sec
+epoch 10, total_step 166140, total loss is 18.75 , inference loss is 12.10, weight deacy loss is 6.65, training accuracy is 0.281250, time 209.311 samples/sec
+epoch 10, total_step 166160, total loss is 14.83 , inference loss is 8.18, weight deacy loss is 6.65, training accuracy is 0.468750, time 181.227 samples/sec
+epoch 10, total_step 166180, total loss is 17.68 , inference loss is 11.03, weight deacy loss is 6.65, training accuracy is 0.437500, time 225.574 samples/sec
+epoch 10, total_step 166200, total loss is 15.21 , inference loss is 8.56, weight deacy loss is 6.65, training accuracy is 0.312500, time 14.889 samples/sec
+epoch 10, total_step 166220, total loss is 15.52 , inference loss is 8.87, weight deacy loss is 6.65, training accuracy is 0.312500, time 283.772 samples/sec
+epoch 10, total_step 166240, total loss is 16.08 , inference loss is 9.42, weight deacy loss is 6.65, training accuracy is 0.281250, time 218.192 samples/sec
+epoch 10, total_step 166260, total loss is 15.76 , inference loss is 9.11, weight deacy loss is 6.65, training accuracy is 0.375000, time 214.729 samples/sec
+epoch 10, total_step 166280, total loss is 15.01 , inference loss is 8.36, weight deacy loss is 6.65, training accuracy is 0.406250, time 217.393 samples/sec
+epoch 10, total_step 166300, total loss is 13.19 , inference loss is 6.54, weight deacy loss is 6.65, training accuracy is 0.281250, time 216.502 samples/sec
+epoch 10, total_step 166320, total loss is 12.61 , inference loss is 5.96, weight deacy loss is 6.65, training accuracy is 0.437500, time 219.366 samples/sec
+epoch 10, total_step 166340, total loss is 16.68 , inference loss is 10.03, weight deacy loss is 6.65, training accuracy is 0.312500, time 214.652 samples/sec
+epoch 10, total_step 166360, total loss is 17.32 , inference loss is 10.67, weight deacy loss is 6.65, training accuracy is 0.312500, time 217.386 samples/sec
+epoch 10, total_step 166380, total loss is 18.45 , inference loss is 11.80, weight deacy loss is 6.65, training accuracy is 0.187500, time 297.857 samples/sec
+epoch 10, total_step 166400, total loss is 18.17 , inference loss is 11.52, weight deacy loss is 6.65, training accuracy is 0.250000, time 299.433 samples/sec
+epoch 10, total_step 166420, total loss is 13.87 , inference loss is 7.22, weight deacy loss is 6.65, training accuracy is 0.375000, time 221.050 samples/sec
+epoch 10, total_step 166440, total loss is 15.57 , inference loss is 8.92, weight deacy loss is 6.65, training accuracy is 0.312500, time 213.933 samples/sec
+epoch 10, total_step 166460, total loss is 17.02 , inference loss is 10.38, weight deacy loss is 6.65, training accuracy is 0.343750, time 217.443 samples/sec
+epoch 10, total_step 166480, total loss is 14.50 , inference loss is 7.85, weight deacy loss is 6.65, training accuracy is 0.437500, time 213.820 samples/sec
+epoch 10, total_step 166500, total loss is 14.87 , inference loss is 8.22, weight deacy loss is 6.65, training accuracy is 0.437500, time 13.081 samples/sec
+epoch 10, total_step 166520, total loss is 13.90 , inference loss is 7.25, weight deacy loss is 6.65, training accuracy is 0.375000, time 210.695 samples/sec
+epoch 10, total_step 166540, total loss is 16.98 , inference loss is 10.33, weight deacy loss is 6.65, training accuracy is 0.468750, time 223.475 samples/sec
+epoch 10, total_step 166560, total loss is 14.23 , inference loss is 7.58, weight deacy loss is 6.65, training accuracy is 0.343750, time 212.351 samples/sec
+epoch 10, total_step 166580, total loss is 12.99 , inference loss is 6.34, weight deacy loss is 6.65, training accuracy is 0.468750, time 215.223 samples/sec
+epoch 10, total_step 166600, total loss is 17.25 , inference loss is 10.60, weight deacy loss is 6.65, training accuracy is 0.250000, time 211.531 samples/sec
+epoch 10, total_step 166620, total loss is 13.19 , inference loss is 6.54, weight deacy loss is 6.65, training accuracy is 0.312500, time 228.526 samples/sec
+epoch 10, total_step 166640, total loss is 16.30 , inference loss is 9.66, weight deacy loss is 6.65, training accuracy is 0.281250, time 221.064 samples/sec
+epoch 10, total_step 166660, total loss is 18.48 , inference loss is 11.84, weight deacy loss is 6.64, training accuracy is 0.250000, time 296.849 samples/sec
+epoch 10, total_step 166680, total loss is 19.58 , inference loss is 12.94, weight deacy loss is 6.64, training accuracy is 0.218750, time 218.717 samples/sec
+epoch 10, total_step 166700, total loss is 13.88 , inference loss is 7.24, weight deacy loss is 6.64, training accuracy is 0.375000, time 216.206 samples/sec
+epoch 10, total_step 166720, total loss is 15.83 , inference loss is 9.19, weight deacy loss is 6.64, training accuracy is 0.218750, time 216.287 samples/sec
+epoch 10, total_step 166740, total loss is 14.78 , inference loss is 8.14, weight deacy loss is 6.64, training accuracy is 0.468750, time 216.183 samples/sec
+epoch 10, total_step 166760, total loss is 17.41 , inference loss is 10.77, weight deacy loss is 6.64, training accuracy is 0.218750, time 216.862 samples/sec
+epoch 10, total_step 166780, total loss is 14.07 , inference loss is 7.43, weight deacy loss is 6.64, training accuracy is 0.312500, time 229.498 samples/sec
+epoch 10, total_step 166800, total loss is 15.94 , inference loss is 9.30, weight deacy loss is 6.64, training accuracy is 0.375000, time 12.784 samples/sec
+epoch 10, total_step 166820, total loss is 16.23 , inference loss is 9.59, weight deacy loss is 6.64, training accuracy is 0.281250, time 223.356 samples/sec
+epoch 10, total_step 166840, total loss is 18.00 , inference loss is 11.36, weight deacy loss is 6.64, training accuracy is 0.312500, time 220.598 samples/sec
+epoch 10, total_step 166860, total loss is 17.09 , inference loss is 10.44, weight deacy loss is 6.64, training accuracy is 0.437500, time 230.109 samples/sec
+epoch 10, total_step 166880, total loss is 16.97 , inference loss is 10.32, weight deacy loss is 6.64, training accuracy is 0.218750, time 218.049 samples/sec
+epoch 10, total_step 166900, total loss is 16.33 , inference loss is 9.68, weight deacy loss is 6.64, training accuracy is 0.281250, time 224.075 samples/sec
+epoch 10, total_step 166920, total loss is 17.34 , inference loss is 10.69, weight deacy loss is 6.64, training accuracy is 0.250000, time 297.219 samples/sec
+epoch 10, total_step 166940, total loss is 16.59 , inference loss is 9.95, weight deacy loss is 6.64, training accuracy is 0.375000, time 224.005 samples/sec
+epoch 10, total_step 166960, total loss is 15.42 , inference loss is 8.78, weight deacy loss is 6.64, training accuracy is 0.406250, time 213.982 samples/sec
+epoch 10, total_step 166980, total loss is 15.79 , inference loss is 9.15, weight deacy loss is 6.64, training accuracy is 0.343750, time 221.293 samples/sec
+epoch 10, total_step 167000, total loss is 16.03 , inference loss is 9.39, weight deacy loss is 6.64, training accuracy is 0.406250, time 218.928 samples/sec
+epoch 10, total_step 167020, total loss is 13.37 , inference loss is 6.73, weight deacy loss is 6.64, training accuracy is 0.312500, time 221.374 samples/sec
+epoch 10, total_step 167040, total loss is 14.22 , inference loss is 7.58, weight deacy loss is 6.64, training accuracy is 0.375000, time 216.559 samples/sec
+epoch 10, total_step 167060, total loss is 13.75 , inference loss is 7.11, weight deacy loss is 6.64, training accuracy is 0.406250, time 221.574 samples/sec
+epoch 10, total_step 167080, total loss is 11.80 , inference loss is 5.17, weight deacy loss is 6.64, training accuracy is 0.406250, time 227.336 samples/sec
+epoch 10, total_step 167100, total loss is 11.94 , inference loss is 5.30, weight deacy loss is 6.64, training accuracy is 0.531250, time 11.785 samples/sec
+epoch 10, total_step 167120, total loss is 13.58 , inference loss is 6.94, weight deacy loss is 6.64, training accuracy is 0.375000, time 217.769 samples/sec
+epoch 10, total_step 167140, total loss is 15.06 , inference loss is 8.42, weight deacy loss is 6.64, training accuracy is 0.343750, time 222.836 samples/sec
+epoch 10, total_step 167160, total loss is 12.44 , inference loss is 5.80, weight deacy loss is 6.64, training accuracy is 0.312500, time 213.418 samples/sec
+epoch 10, total_step 167180, total loss is 18.67 , inference loss is 12.04, weight deacy loss is 6.64, training accuracy is 0.281250, time 299.401 samples/sec
+epoch 10, total_step 167200, total loss is 18.92 , inference loss is 12.28, weight deacy loss is 6.64, training accuracy is 0.312500, time 212.146 samples/sec
+epoch 10, total_step 167220, total loss is 16.68 , inference loss is 10.05, weight deacy loss is 6.64, training accuracy is 0.312500, time 207.759 samples/sec
+epoch 10, total_step 167240, total loss is 16.96 , inference loss is 10.32, weight deacy loss is 6.64, training accuracy is 0.312500, time 223.916 samples/sec
+epoch 10, total_step 167260, total loss is 15.33 , inference loss is 8.69, weight deacy loss is 6.64, training accuracy is 0.312500, time 234.296 samples/sec
+epoch 10, total_step 167280, total loss is 15.93 , inference loss is 9.29, weight deacy loss is 6.64, training accuracy is 0.218750, time 221.422 samples/sec
+epoch 10, total_step 167300, total loss is 14.18 , inference loss is 7.54, weight deacy loss is 6.64, training accuracy is 0.343750, time 214.563 samples/sec
+epoch 10, total_step 167320, total loss is 15.69 , inference loss is 9.05, weight deacy loss is 6.64, training accuracy is 0.312500, time 213.155 samples/sec
+epoch 10, total_step 167340, total loss is 16.68 , inference loss is 10.04, weight deacy loss is 6.64, training accuracy is 0.250000, time 213.140 samples/sec
+epoch 10, total_step 167360, total loss is 15.32 , inference loss is 8.68, weight deacy loss is 6.64, training accuracy is 0.312500, time 220.788 samples/sec
+epoch 10, total_step 167380, total loss is 13.76 , inference loss is 7.12, weight deacy loss is 6.63, training accuracy is 0.375000, time 209.646 samples/sec
+epoch 10, total_step 167400, total loss is 15.27 , inference loss is 8.64, weight deacy loss is 6.63, training accuracy is 0.187500, time 6.799 samples/sec
+epoch 10, total_step 167420, total loss is 14.40 , inference loss is 7.77, weight deacy loss is 6.63, training accuracy is 0.375000, time 197.002 samples/sec
+epoch 10, total_step 167440, total loss is 12.11 , inference loss is 5.47, weight deacy loss is 6.63, training accuracy is 0.375000, time 298.707 samples/sec
+epoch 10, total_step 167460, total loss is 14.54 , inference loss is 7.90, weight deacy loss is 6.63, training accuracy is 0.187500, time 205.062 samples/sec
+epoch 10, total_step 167480, total loss is 16.31 , inference loss is 9.67, weight deacy loss is 6.63, training accuracy is 0.312500, time 205.063 samples/sec
+epoch 10, total_step 167500, total loss is 21.38 , inference loss is 14.75, weight deacy loss is 6.63, training accuracy is 0.281250, time 236.534 samples/sec
+epoch 10, total_step 167520, total loss is 15.11 , inference loss is 8.47, weight deacy loss is 6.63, training accuracy is 0.406250, time 219.677 samples/sec
+epoch 10, total_step 167540, total loss is 15.90 , inference loss is 9.26, weight deacy loss is 6.63, training accuracy is 0.375000, time 221.100 samples/sec
+epoch 10, total_step 167560, total loss is 18.88 , inference loss is 12.25, weight deacy loss is 6.63, training accuracy is 0.312500, time 216.712 samples/sec
+epoch 10, total_step 167580, total loss is 14.78 , inference loss is 8.14, weight deacy loss is 6.63, training accuracy is 0.437500, time 215.605 samples/sec
+epoch 10, total_step 167600, total loss is 22.97 , inference loss is 16.34, weight deacy loss is 6.63, training accuracy is 0.156250, time 219.809 samples/sec
+epoch 10, total_step 167620, total loss is 18.73 , inference loss is 12.10, weight deacy loss is 6.63, training accuracy is 0.218750, time 214.352 samples/sec
+epoch 10, total_step 167640, total loss is 15.19 , inference loss is 8.56, weight deacy loss is 6.63, training accuracy is 0.437500, time 221.065 samples/sec
+epoch 10, total_step 167660, total loss is 14.93 , inference loss is 8.30, weight deacy loss is 6.63, training accuracy is 0.406250, time 217.956 samples/sec
+epoch 10, total_step 167680, total loss is 18.26 , inference loss is 11.63, weight deacy loss is 6.63, training accuracy is 0.281250, time 222.747 samples/sec
+epoch 10, total_step 167700, total loss is 18.91 , inference loss is 12.28, weight deacy loss is 6.63, training accuracy is 0.312500, time 13.144 samples/sec
+epoch 10, total_step 167720, total loss is 13.02 , inference loss is 6.39, weight deacy loss is 6.63, training accuracy is 0.281250, time 231.040 samples/sec
+epoch 10, total_step 167740, total loss is 13.25 , inference loss is 6.62, weight deacy loss is 6.63, training accuracy is 0.281250, time 224.427 samples/sec
+epoch 10, total_step 167760, total loss is 17.85 , inference loss is 11.22, weight deacy loss is 6.63, training accuracy is 0.312500, time 223.937 samples/sec
+epoch 10, total_step 167780, total loss is 16.67 , inference loss is 10.04, weight deacy loss is 6.63, training accuracy is 0.312500, time 221.134 samples/sec
+epoch 10, total_step 167800, total loss is 19.56 , inference loss is 12.93, weight deacy loss is 6.63, training accuracy is 0.281250, time 223.134 samples/sec
+epoch 10, total_step 167820, total loss is 13.50 , inference loss is 6.88, weight deacy loss is 6.63, training accuracy is 0.250000, time 202.210 samples/sec
+epoch 10, total_step 167840, total loss is 13.15 , inference loss is 6.52, weight deacy loss is 6.63, training accuracy is 0.406250, time 206.269 samples/sec
+epoch 10, total_step 167860, total loss is 12.49 , inference loss is 5.86, weight deacy loss is 6.63, training accuracy is 0.375000, time 223.117 samples/sec
+epoch 10, total_step 167880, total loss is 16.47 , inference loss is 9.84, weight deacy loss is 6.63, training accuracy is 0.218750, time 220.488 samples/sec
+epoch 10, total_step 167900, total loss is 16.33 , inference loss is 9.70, weight deacy loss is 6.63, training accuracy is 0.250000, time 220.016 samples/sec
+epoch 10, total_step 167920, total loss is 18.72 , inference loss is 12.09, weight deacy loss is 6.63, training accuracy is 0.250000, time 216.698 samples/sec
+epoch 10, total_step 167940, total loss is 19.76 , inference loss is 13.13, weight deacy loss is 6.63, training accuracy is 0.250000, time 230.195 samples/sec
+epoch 10, total_step 167960, total loss is 16.05 , inference loss is 9.43, weight deacy loss is 6.63, training accuracy is 0.375000, time 218.474 samples/sec
+epoch 10, total_step 167980, total loss is 20.06 , inference loss is 13.43, weight deacy loss is 6.63, training accuracy is 0.406250, time 224.789 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.377277000000028
+best_threshold_index 107 0.9840740740740741
+best_threshold_index 105 0.9827777777777778
+best_threshold_index 105 0.9840740740740741
+best_threshold_index 109 0.9855555555555555
+best_threshold_index 107 0.9835185185185186
+best_threshold_index 107 0.9824074074074074
+best_threshold_index 107 0.9840740740740741
+best_threshold_index 105 0.9829629629629629
+best_threshold_index 107 0.9816666666666667
+best_threshold_index 107 0.9829629629629629
+[lfw][168000]XNorm: 23.014036
+[lfw][168000]Accuracy-Flip: 0.98217+-0.00931
+test accuracy is: 0.9821666666666667
+epoch 10, total_step 168000, total loss is 19.10 , inference loss is 12.47, weight deacy loss is 6.63, training accuracy is 0.343750, time 16.866 samples/sec2022-12-01 11:29:29.108516: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 6939 of 10000
+2022-12-01 11:29:33.585448: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 10, total_step 168020, total loss is 14.40 , inference loss is 7.78, weight deacy loss is 6.63, training accuracy is 0.343750, time 213.768 samples/sec
+epoch 10, total_step 168040, total loss is 14.75 , inference loss is 8.13, weight deacy loss is 6.63, training accuracy is 0.375000, time 225.806 samples/sec
+epoch 10, total_step 168060, total loss is 16.01 , inference loss is 9.38, weight deacy loss is 6.63, training accuracy is 0.250000, time 215.661 samples/sec
+epoch 10, total_step 168080, total loss is 16.00 , inference loss is 9.38, weight deacy loss is 6.63, training accuracy is 0.312500, time 213.749 samples/sec
+epoch 10, total_step 168100, total loss is 15.33 , inference loss is 8.70, weight deacy loss is 6.63, training accuracy is 0.406250, time 215.840 samples/sec
+epoch 10, total_step 168120, total loss is 17.07 , inference loss is 10.45, weight deacy loss is 6.63, training accuracy is 0.375000, time 222.788 samples/sec
+epoch 10, total_step 168140, total loss is 13.86 , inference loss is 7.23, weight deacy loss is 6.63, training accuracy is 0.437500, time 297.948 samples/sec
+epoch 10, total_step 168160, total loss is 18.48 , inference loss is 11.86, weight deacy loss is 6.62, training accuracy is 0.343750, time 215.734 samples/sec
+epoch 10, total_step 168180, total loss is 19.92 , inference loss is 13.30, weight deacy loss is 6.62, training accuracy is 0.250000, time 212.883 samples/sec
+epoch 10, total_step 168200, total loss is 17.65 , inference loss is 11.03, weight deacy loss is 6.62, training accuracy is 0.312500, time 222.168 samples/sec
+epoch 10, total_step 168220, total loss is 14.26 , inference loss is 7.64, weight deacy loss is 6.62, training accuracy is 0.468750, time 216.120 samples/sec
+epoch 10, total_step 168240, total loss is 18.88 , inference loss is 12.26, weight deacy loss is 6.62, training accuracy is 0.250000, time 215.052 samples/sec
+epoch 10, total_step 168260, total loss is 17.16 , inference loss is 10.54, weight deacy loss is 6.62, training accuracy is 0.312500, time 298.613 samples/sec
+epoch 10, total_step 168280, total loss is 14.01 , inference loss is 7.39, weight deacy loss is 6.62, training accuracy is 0.437500, time 218.935 samples/sec
+epoch 10, total_step 168300, total loss is 15.70 , inference loss is 9.08, weight deacy loss is 6.62, training accuracy is 0.343750, time 12.668 samples/sec
+epoch 10, total_step 168320, total loss is 15.45 , inference loss is 8.83, weight deacy loss is 6.62, training accuracy is 0.187500, time 216.124 samples/sec
+epoch 10, total_step 168340, total loss is 17.07 , inference loss is 10.45, weight deacy loss is 6.62, training accuracy is 0.312500, time 187.895 samples/sec
+epoch 10, total_step 168360, total loss is 13.44 , inference loss is 6.82, weight deacy loss is 6.62, training accuracy is 0.468750, time 183.063 samples/sec
+epoch 10, total_step 168380, total loss is 20.09 , inference loss is 13.46, weight deacy loss is 6.62, training accuracy is 0.187500, time 203.905 samples/sec
+epoch 10, total_step 168400, total loss is 17.08 , inference loss is 10.46, weight deacy loss is 6.62, training accuracy is 0.343750, time 203.846 samples/sec
+epoch 10, total_step 168420, total loss is 11.09 , inference loss is 4.47, weight deacy loss is 6.62, training accuracy is 0.531250, time 225.281 samples/sec
+epoch 10, total_step 168440, total loss is 10.82 , inference loss is 4.20, weight deacy loss is 6.62, training accuracy is 0.375000, time 206.667 samples/sec
+epoch 10, total_step 168460, total loss is 14.48 , inference loss is 7.85, weight deacy loss is 6.62, training accuracy is 0.312500, time 203.961 samples/sec
+epoch 10, total_step 168480, total loss is 16.71 , inference loss is 10.09, weight deacy loss is 6.62, training accuracy is 0.125000, time 181.332 samples/sec
+epoch 10, total_step 168500, total loss is 14.41 , inference loss is 7.78, weight deacy loss is 6.62, training accuracy is 0.312500, time 206.322 samples/sec
+epoch 10, total_step 168520, total loss is 14.86 , inference loss is 8.24, weight deacy loss is 6.62, training accuracy is 0.500000, time 211.448 samples/sec
+epoch 10, total_step 168540, total loss is 15.95 , inference loss is 9.33, weight deacy loss is 6.62, training accuracy is 0.343750, time 219.373 samples/sec
+epoch 10, total_step 168560, total loss is 16.64 , inference loss is 10.02, weight deacy loss is 6.62, training accuracy is 0.281250, time 181.639 samples/sec
+epoch 10, total_step 168580, total loss is 16.26 , inference loss is 9.64, weight deacy loss is 6.62, training accuracy is 0.343750, time 298.100 samples/sec
+epoch 10, total_step 168600, total loss is 14.49 , inference loss is 7.87, weight deacy loss is 6.62, training accuracy is 0.406250, time 13.878 samples/sec
+epoch 10, total_step 168620, total loss is 15.72 , inference loss is 9.10, weight deacy loss is 6.62, training accuracy is 0.250000, time 227.177 samples/sec
+epoch 10, total_step 168640, total loss is 13.53 , inference loss is 6.91, weight deacy loss is 6.62, training accuracy is 0.375000, time 197.613 samples/sec
+End of epoch 10
+epoch 11, total_step 168660, total loss is 11.06 , inference loss is 4.45, weight deacy loss is 6.62, training accuracy is 0.593750, time 218.967 samples/sec
+epoch 11, total_step 168680, total loss is 9.74 , inference loss is 3.12, weight deacy loss is 6.62, training accuracy is 0.656250, time 225.902 samples/sec
+epoch 11, total_step 168700, total loss is 11.27 , inference loss is 4.65, weight deacy loss is 6.62, training accuracy is 0.718750, time 230.515 samples/sec
+epoch 11, total_step 168720, total loss is 10.99 , inference loss is 4.37, weight deacy loss is 6.62, training accuracy is 0.625000, time 222.531 samples/sec
+epoch 11, total_step 168740, total loss is 9.80 , inference loss is 3.18, weight deacy loss is 6.62, training accuracy is 0.718750, time 227.462 samples/sec
+epoch 11, total_step 168760, total loss is 12.39 , inference loss is 5.78, weight deacy loss is 6.62, training accuracy is 0.687500, time 201.689 samples/sec
+epoch 11, total_step 168780, total loss is 10.24 , inference loss is 3.63, weight deacy loss is 6.62, training accuracy is 0.812500, time 202.851 samples/sec
+epoch 11, total_step 168800, total loss is 9.93 , inference loss is 3.32, weight deacy loss is 6.62, training accuracy is 0.656250, time 245.856 samples/sec
+epoch 11, total_step 168820, total loss is 9.85 , inference loss is 3.24, weight deacy loss is 6.61, training accuracy is 0.625000, time 206.155 samples/sec
+epoch 11, total_step 168840, total loss is 12.53 , inference loss is 5.92, weight deacy loss is 6.61, training accuracy is 0.687500, time 202.499 samples/sec
+epoch 11, total_step 168860, total loss is 12.98 , inference loss is 6.37, weight deacy loss is 6.61, training accuracy is 0.750000, time 224.125 samples/sec
+epoch 11, total_step 168880, total loss is 9.13 , inference loss is 2.51, weight deacy loss is 6.61, training accuracy is 0.656250, time 249.387 samples/sec
+epoch 11, total_step 168900, total loss is 11.35 , inference loss is 4.74, weight deacy loss is 6.61, training accuracy is 0.593750, time 13.530 samples/sec
+epoch 11, total_step 168920, total loss is 11.55 , inference loss is 4.94, weight deacy loss is 6.61, training accuracy is 0.625000, time 225.062 samples/sec
+epoch 11, total_step 168940, total loss is 13.68 , inference loss is 7.07, weight deacy loss is 6.61, training accuracy is 0.625000, time 227.424 samples/sec
+epoch 11, total_step 168960, total loss is 8.26 , inference loss is 1.65, weight deacy loss is 6.61, training accuracy is 0.812500, time 214.481 samples/sec
+epoch 11, total_step 168980, total loss is 10.79 , inference loss is 4.18, weight deacy loss is 6.61, training accuracy is 0.593750, time 225.427 samples/sec
+epoch 11, total_step 169000, total loss is 13.29 , inference loss is 6.68, weight deacy loss is 6.61, training accuracy is 0.625000, time 246.164 samples/sec
+epoch 11, total_step 169020, total loss is 9.87 , inference loss is 3.26, weight deacy loss is 6.61, training accuracy is 0.625000, time 298.319 samples/sec
+epoch 11, total_step 169040, total loss is 9.21 , inference loss is 2.60, weight deacy loss is 6.61, training accuracy is 0.625000, time 232.915 samples/sec
+epoch 11, total_step 169060, total loss is 12.81 , inference loss is 6.20, weight deacy loss is 6.61, training accuracy is 0.593750, time 224.108 samples/sec
+epoch 11, total_step 169080, total loss is 12.14 , inference loss is 5.53, weight deacy loss is 6.61, training accuracy is 0.593750, time 183.836 samples/sec
+epoch 11, total_step 169100, total loss is 13.25 , inference loss is 6.63, weight deacy loss is 6.61, training accuracy is 0.593750, time 231.201 samples/sec
+epoch 11, total_step 169120, total loss is 15.84 , inference loss is 9.23, weight deacy loss is 6.61, training accuracy is 0.437500, time 218.530 samples/sec
+epoch 11, total_step 169140, total loss is 10.38 , inference loss is 3.77, weight deacy loss is 6.61, training accuracy is 0.718750, time 199.913 samples/sec
+epoch 11, total_step 169160, total loss is 10.93 , inference loss is 4.31, weight deacy loss is 6.61, training accuracy is 0.531250, time 219.536 samples/sec
+epoch 11, total_step 169180, total loss is 14.66 , inference loss is 8.04, weight deacy loss is 6.61, training accuracy is 0.593750, time 227.313 samples/sec
+epoch 11, total_step 169200, total loss is 16.46 , inference loss is 9.85, weight deacy loss is 6.61, training accuracy is 0.531250, time 13.040 samples/sec
+epoch 11, total_step 169220, total loss is 8.77 , inference loss is 2.16, weight deacy loss is 6.61, training accuracy is 0.750000, time 229.091 samples/sec
+epoch 11, total_step 169240, total loss is 18.40 , inference loss is 11.79, weight deacy loss is 6.61, training accuracy is 0.531250, time 223.699 samples/sec
+epoch 11, total_step 169260, total loss is 8.25 , inference loss is 1.64, weight deacy loss is 6.61, training accuracy is 0.812500, time 214.415 samples/sec
+epoch 11, total_step 169280, total loss is 11.55 , inference loss is 4.93, weight deacy loss is 6.61, training accuracy is 0.562500, time 300.658 samples/sec
+epoch 11, total_step 169300, total loss is 16.26 , inference loss is 9.64, weight deacy loss is 6.61, training accuracy is 0.656250, time 220.694 samples/sec
+epoch 11, total_step 169320, total loss is 13.23 , inference loss is 6.62, weight deacy loss is 6.61, training accuracy is 0.593750, time 225.327 samples/sec
+epoch 11, total_step 169340, total loss is 12.92 , inference loss is 6.30, weight deacy loss is 6.61, training accuracy is 0.687500, time 225.828 samples/sec
+epoch 11, total_step 169360, total loss is 11.66 , inference loss is 5.05, weight deacy loss is 6.61, training accuracy is 0.656250, time 227.155 samples/sec
+epoch 11, total_step 169380, total loss is 13.93 , inference loss is 7.31, weight deacy loss is 6.61, training accuracy is 0.593750, time 226.509 samples/sec
+epoch 11, total_step 169400, total loss is 11.75 , inference loss is 5.14, weight deacy loss is 6.61, training accuracy is 0.562500, time 225.421 samples/sec
+epoch 11, total_step 169420, total loss is 10.56 , inference loss is 3.95, weight deacy loss is 6.61, training accuracy is 0.656250, time 214.634 samples/sec
+epoch 11, total_step 169440, total loss is 13.41 , inference loss is 6.80, weight deacy loss is 6.61, training accuracy is 0.500000, time 227.744 samples/sec
+epoch 11, total_step 169460, total loss is 13.60 , inference loss is 6.99, weight deacy loss is 6.61, training accuracy is 0.468750, time 226.746 samples/sec
+epoch 11, total_step 169480, total loss is 9.17 , inference loss is 2.56, weight deacy loss is 6.61, training accuracy is 0.750000, time 226.746 samples/sec
+epoch 11, total_step 169500, total loss is 11.00 , inference loss is 4.38, weight deacy loss is 6.61, training accuracy is 0.656250, time 12.049 samples/sec
+epoch 11, total_step 169520, total loss is 13.04 , inference loss is 6.43, weight deacy loss is 6.61, training accuracy is 0.593750, time 219.348 samples/sec
+epoch 11, total_step 169540, total loss is 15.99 , inference loss is 9.37, weight deacy loss is 6.61, training accuracy is 0.500000, time 296.595 samples/sec
+epoch 11, total_step 169560, total loss is 10.97 , inference loss is 4.35, weight deacy loss is 6.61, training accuracy is 0.593750, time 237.775 samples/sec
+epoch 11, total_step 169580, total loss is 13.89 , inference loss is 7.28, weight deacy loss is 6.61, training accuracy is 0.562500, time 225.074 samples/sec
+epoch 11, total_step 169600, total loss is 11.89 , inference loss is 5.27, weight deacy loss is 6.61, training accuracy is 0.656250, time 225.994 samples/sec
+epoch 11, total_step 169620, total loss is 14.81 , inference loss is 8.19, weight deacy loss is 6.61, training accuracy is 0.468750, time 208.140 samples/sec
+epoch 11, total_step 169640, total loss is 12.61 , inference loss is 6.00, weight deacy loss is 6.61, training accuracy is 0.625000, time 239.937 samples/sec
+epoch 11, total_step 169660, total loss is 12.37 , inference loss is 5.76, weight deacy loss is 6.61, training accuracy is 0.593750, time 221.938 samples/sec
+epoch 11, total_step 169680, total loss is 15.23 , inference loss is 8.62, weight deacy loss is 6.61, training accuracy is 0.437500, time 221.184 samples/sec
+epoch 11, total_step 169700, total loss is 22.95 , inference loss is 16.34, weight deacy loss is 6.61, training accuracy is 0.437500, time 223.142 samples/sec
+epoch 11, total_step 169720, total loss is 14.93 , inference loss is 8.31, weight deacy loss is 6.61, training accuracy is 0.625000, time 225.109 samples/sec
+epoch 11, total_step 169740, total loss is 13.98 , inference loss is 7.37, weight deacy loss is 6.61, training accuracy is 0.562500, time 217.927 samples/sec
+epoch 11, total_step 169760, total loss is 13.40 , inference loss is 6.79, weight deacy loss is 6.61, training accuracy is 0.531250, time 222.802 samples/sec
+epoch 11, total_step 169780, total loss is 9.98 , inference loss is 3.37, weight deacy loss is 6.61, training accuracy is 0.562500, time 222.511 samples/sec
+epoch 11, total_step 169800, total loss is 13.83 , inference loss is 7.22, weight deacy loss is 6.61, training accuracy is 0.593750, time 12.593 samples/sec
+epoch 11, total_step 169820, total loss is 14.84 , inference loss is 8.22, weight deacy loss is 6.61, training accuracy is 0.468750, time 226.671 samples/sec
+epoch 11, total_step 169840, total loss is 10.84 , inference loss is 4.23, weight deacy loss is 6.61, training accuracy is 0.625000, time 219.762 samples/sec
+epoch 11, total_step 169860, total loss is 12.38 , inference loss is 5.77, weight deacy loss is 6.61, training accuracy is 0.656250, time 225.712 samples/sec
+epoch 11, total_step 169880, total loss is 12.45 , inference loss is 5.83, weight deacy loss is 6.61, training accuracy is 0.593750, time 228.021 samples/sec
+epoch 11, total_step 169900, total loss is 10.58 , inference loss is 3.97, weight deacy loss is 6.61, training accuracy is 0.593750, time 235.604 samples/sec
+epoch 11, total_step 169920, total loss is 12.81 , inference loss is 6.19, weight deacy loss is 6.61, training accuracy is 0.531250, time 212.333 samples/sec
+epoch 11, total_step 169940, total loss is 10.94 , inference loss is 4.33, weight deacy loss is 6.61, training accuracy is 0.531250, time 220.981 samples/sec
+epoch 11, total_step 169960, total loss is 13.31 , inference loss is 6.69, weight deacy loss is 6.62, training accuracy is 0.562500, time 216.133 samples/sec
+epoch 11, total_step 169980, total loss is 14.76 , inference loss is 8.15, weight deacy loss is 6.62, training accuracy is 0.531250, time 235.228 samples/sec
+testing verification..
+(12000, 512)
+infer time 21.827897000000025
+best_threshold_index 114 0.9842592592592593
+best_threshold_index 114 0.9835185185185186
+best_threshold_index 114 0.9840740740740741
+best_threshold_index 114 0.9844444444444445
+best_threshold_index 120 0.9840740740740741
+best_threshold_index 120 0.9835185185185186
+best_threshold_index 115 0.9846296296296296
+best_threshold_index 114 0.9844444444444445
+best_threshold_index 114 0.9825925925925926
+best_threshold_index 114 0.9835185185185186
+[lfw][170000]XNorm: 22.310656
+[lfw][170000]Accuracy-Flip: 0.98283+-0.00527
+test accuracy is: 0.9828333333333333
+epoch 11, total_step 170000, total loss is 10.97 , inference loss is 4.35, weight deacy loss is 6.62, training accuracy is 0.625000, time 16.808 samples/sec
+epoch 11, total_step 170020, total loss is 11.94 , inference loss is 5.33, weight deacy loss is 6.62, training accuracy is 0.625000, time 226.283 samples/sec
+epoch 11, total_step 170040, total loss is 16.19 , inference loss is 9.58, weight deacy loss is 6.62, training accuracy is 0.562500, time 226.022 samples/sec
+epoch 11, total_step 170060, total loss is 12.56 , inference loss is 5.95, weight deacy loss is 6.62, training accuracy is 0.625000, time 299.865 samples/sec
+epoch 11, total_step 170080, total loss is 14.03 , inference loss is 7.41, weight deacy loss is 6.62, training accuracy is 0.531250, time 218.022 samples/sec
+epoch 11, total_step 170100, total loss is 14.02 , inference loss is 7.40, weight deacy loss is 6.62, training accuracy is 0.593750, time 13.190 samples/sec
+epoch 11, total_step 170120, total loss is 15.35 , inference loss is 8.73, weight deacy loss is 6.62, training accuracy is 0.468750, time 222.246 samples/sec
+epoch 11, total_step 170140, total loss is 13.27 , inference loss is 6.66, weight deacy loss is 6.62, training accuracy is 0.468750, time 220.598 samples/sec
+epoch 11, total_step 170160, total loss is 16.46 , inference loss is 9.84, weight deacy loss is 6.62, training accuracy is 0.406250, time 222.286 samples/sec
+epoch 11, total_step 170180, total loss is 11.74 , inference loss is 5.12, weight deacy loss is 6.62, training accuracy is 0.500000, time 224.882 samples/sec
+epoch 11, total_step 170200, total loss is 13.17 , inference loss is 6.55, weight deacy loss is 6.62, training accuracy is 0.500000, time 222.308 samples/sec
+epoch 11, total_step 170220, total loss is 11.53 , inference loss is 4.92, weight deacy loss is 6.62, training accuracy is 0.562500, time 221.441 samples/sec
+epoch 11, total_step 170240, total loss is 16.62 , inference loss is 10.00, weight deacy loss is 6.62, training accuracy is 0.343750, time 227.615 samples/sec
+epoch 11, total_step 170260, total loss is 13.98 , inference loss is 7.37, weight deacy loss is 6.62, training accuracy is 0.500000, time 226.279 samples/sec
+epoch 11, total_step 170280, total loss is 14.11 , inference loss is 7.50, weight deacy loss is 6.62, training accuracy is 0.406250, time 226.464 samples/sec
+epoch 11, total_step 170300, total loss is 12.35 , inference loss is 5.74, weight deacy loss is 6.62, training accuracy is 0.500000, time 228.813 samples/sec
+epoch 11, total_step 170320, total loss is 12.85 , inference loss is 6.24, weight deacy loss is 6.62, training accuracy is 0.593750, time 298.203 samples/sec
+epoch 11, total_step 170340, total loss is 15.96 , inference loss is 9.34, weight deacy loss is 6.62, training accuracy is 0.531250, time 216.049 samples/sec
+epoch 11, total_step 170360, total loss is 16.11 , inference loss is 9.50, weight deacy loss is 6.62, training accuracy is 0.468750, time 225.045 samples/sec
+epoch 11, total_step 170380, total loss is 16.38 , inference loss is 9.77, weight deacy loss is 6.62, training accuracy is 0.343750, time 204.649 samples/sec
+epoch 11, total_step 170400, total loss is 12.92 , inference loss is 6.30, weight deacy loss is 6.62, training accuracy is 0.562500, time 12.392 samples/sec
+epoch 11, total_step 170420, total loss is 16.17 , inference loss is 9.55, weight deacy loss is 6.62, training accuracy is 0.312500, time 225.699 samples/sec
+epoch 11, total_step 170440, total loss is 16.56 , inference loss is 9.94, weight deacy loss is 6.62, training accuracy is 0.468750, time 235.966 samples/sec
+epoch 11, total_step 170460, total loss is 14.75 , inference loss is 8.14, weight deacy loss is 6.62, training accuracy is 0.375000, time 226.906 samples/sec
+epoch 11, total_step 170480, total loss is 12.80 , inference loss is 6.19, weight deacy loss is 6.62, training accuracy is 0.500000, time 219.071 samples/sec
+epoch 11, total_step 170500, total loss is 13.09 , inference loss is 6.47, weight deacy loss is 6.62, training accuracy is 0.562500, time 205.226 samples/sec
+epoch 11, total_step 170520, total loss is 11.91 , inference loss is 5.30, weight deacy loss is 6.62, training accuracy is 0.718750, time 227.305 samples/sec
+epoch 11, total_step 170540, total loss is 13.69 , inference loss is 7.07, weight deacy loss is 6.62, training accuracy is 0.531250, time 221.401 samples/sec
+epoch 11, total_step 170560, total loss is 12.50 , inference loss is 5.88, weight deacy loss is 6.62, training accuracy is 0.500000, time 228.059 samples/sec
+epoch 11, total_step 170580, total loss is 11.89 , inference loss is 5.28, weight deacy loss is 6.62, training accuracy is 0.593750, time 298.340 samples/sec
+epoch 11, total_step 170600, total loss is 11.83 , inference loss is 5.21, weight deacy loss is 6.62, training accuracy is 0.500000, time 219.679 samples/sec
+epoch 11, total_step 170620, total loss is 9.83 , inference loss is 3.21, weight deacy loss is 6.62, training accuracy is 0.718750, time 227.283 samples/sec
+epoch 11, total_step 170640, total loss is 11.50 , inference loss is 4.89, weight deacy loss is 6.62, training accuracy is 0.593750, time 228.623 samples/sec
+epoch 11, total_step 170660, total loss is 13.81 , inference loss is 7.19, weight deacy loss is 6.62, training accuracy is 0.531250, time 219.095 samples/sec
+epoch 11, total_step 170680, total loss is 17.45 , inference loss is 10.84, weight deacy loss is 6.62, training accuracy is 0.375000, time 231.284 samples/sec
+epoch 11, total_step 170700, total loss is 12.70 , inference loss is 6.08, weight deacy loss is 6.62, training accuracy is 0.625000, time 12.681 samples/sec
+epoch 11, total_step 170720, total loss is 14.63 , inference loss is 8.01, weight deacy loss is 6.61, training accuracy is 0.656250, time 218.344 samples/sec
+epoch 11, total_step 170740, total loss is 11.71 , inference loss is 5.10, weight deacy loss is 6.61, training accuracy is 0.718750, time 216.088 samples/sec
+epoch 11, total_step 170760, total loss is 14.31 , inference loss is 7.70, weight deacy loss is 6.61, training accuracy is 0.500000, time 225.543 samples/sec
+epoch 11, total_step 170780, total loss is 10.64 , inference loss is 4.03, weight deacy loss is 6.61, training accuracy is 0.593750, time 226.465 samples/sec
+epoch 11, total_step 170800, total loss is 13.20 , inference loss is 6.59, weight deacy loss is 6.61, training accuracy is 0.593750, time 223.556 samples/sec
+epoch 11, total_step 170820, total loss is 9.84 , inference loss is 3.22, weight deacy loss is 6.61, training accuracy is 0.656250, time 222.587 samples/sec
+epoch 11, total_step 170840, total loss is 10.05 , inference loss is 3.44, weight deacy loss is 6.61, training accuracy is 0.562500, time 296.777 samples/sec
+epoch 11, total_step 170860, total loss is 11.31 , inference loss is 4.69, weight deacy loss is 6.61, training accuracy is 0.593750, time 219.913 samples/sec
+epoch 11, total_step 170880, total loss is 13.05 , inference loss is 6.43, weight deacy loss is 6.61, training accuracy is 0.562500, time 226.442 samples/sec
+epoch 11, total_step 170900, total loss is 16.25 , inference loss is 9.64, weight deacy loss is 6.61, training accuracy is 0.468750, time 238.675 samples/sec
+epoch 11, total_step 170920, total loss is 10.97 , inference loss is 4.36, weight deacy loss is 6.61, training accuracy is 0.562500, time 223.181 samples/sec
+epoch 11, total_step 170940, total loss is 11.30 , inference loss is 4.69, weight deacy loss is 6.61, training accuracy is 0.656250, time 223.966 samples/sec
+epoch 11, total_step 170960, total loss is 13.67 , inference loss is 7.06, weight deacy loss is 6.61, training accuracy is 0.562500, time 213.024 samples/sec
+epoch 11, total_step 170980, total loss is 15.44 , inference loss is 8.83, weight deacy loss is 6.61, training accuracy is 0.656250, time 215.702 samples/sec
+epoch 11, total_step 171000, total loss is 14.05 , inference loss is 7.44, weight deacy loss is 6.61, training accuracy is 0.625000, time 12.125 samples/sec
+epoch 11, total_step 171020, total loss is 10.86 , inference loss is 4.25, weight deacy loss is 6.61, training accuracy is 0.500000, time 224.145 samples/sec
+epoch 11, total_step 171040, total loss is 13.31 , inference loss is 6.69, weight deacy loss is 6.61, training accuracy is 0.562500, time 217.807 samples/sec
+epoch 11, total_step 171060, total loss is 11.72 , inference loss is 5.11, weight deacy loss is 6.61, training accuracy is 0.562500, time 219.997 samples/sec
+epoch 11, total_step 171080, total loss is 9.72 , inference loss is 3.11, weight deacy loss is 6.61, training accuracy is 0.718750, time 230.344 samples/sec
+epoch 11, total_step 171100, total loss is 13.39 , inference loss is 6.78, weight deacy loss is 6.61, training accuracy is 0.468750, time 299.990 samples/sec
+epoch 11, total_step 171120, total loss is 11.85 , inference loss is 5.24, weight deacy loss is 6.61, training accuracy is 0.531250, time 221.399 samples/sec
+epoch 11, total_step 171140, total loss is 10.22 , inference loss is 3.61, weight deacy loss is 6.61, training accuracy is 0.625000, time 218.393 samples/sec
+epoch 11, total_step 171160, total loss is 13.66 , inference loss is 7.05, weight deacy loss is 6.61, training accuracy is 0.500000, time 211.170 samples/sec
+epoch 11, total_step 171180, total loss is 11.86 , inference loss is 5.25, weight deacy loss is 6.61, training accuracy is 0.656250, time 182.522 samples/sec
+epoch 11, total_step 171200, total loss is 13.58 , inference loss is 6.97, weight deacy loss is 6.61, training accuracy is 0.500000, time 220.269 samples/sec
+epoch 11, total_step 171220, total loss is 12.08 , inference loss is 5.47, weight deacy loss is 6.61, training accuracy is 0.500000, time 227.447 samples/sec
+epoch 11, total_step 171240, total loss is 10.79 , inference loss is 4.18, weight deacy loss is 6.61, training accuracy is 0.625000, time 224.988 samples/sec
+epoch 11, total_step 171260, total loss is 11.02 , inference loss is 4.41, weight deacy loss is 6.61, training accuracy is 0.656250, time 205.102 samples/sec
+epoch 11, total_step 171280, total loss is 10.63 , inference loss is 4.02, weight deacy loss is 6.61, training accuracy is 0.625000, time 230.919 samples/sec
+epoch 11, total_step 171300, total loss is 13.47 , inference loss is 6.86, weight deacy loss is 6.61, training accuracy is 0.468750, time 9.243 samples/sec
+epoch 11, total_step 171320, total loss is 11.43 , inference loss is 4.82, weight deacy loss is 6.61, training accuracy is 0.562500, time 209.167 samples/sec
+epoch 11, total_step 171340, total loss is 10.29 , inference loss is 3.68, weight deacy loss is 6.61, training accuracy is 0.593750, time 218.725 samples/sec
+epoch 11, total_step 171360, total loss is 11.02 , inference loss is 4.41, weight deacy loss is 6.61, training accuracy is 0.687500, time 300.595 samples/sec
+epoch 11, total_step 171380, total loss is 9.71 , inference loss is 3.10, weight deacy loss is 6.61, training accuracy is 0.687500, time 231.220 samples/sec
+epoch 11, total_step 171400, total loss is 12.61 , inference loss is 6.01, weight deacy loss is 6.61, training accuracy is 0.593750, time 197.034 samples/sec
+epoch 11, total_step 171420, total loss is 14.50 , inference loss is 7.89, weight deacy loss is 6.61, training accuracy is 0.562500, time 233.961 samples/sec
+epoch 11, total_step 171440, total loss is 13.33 , inference loss is 6.72, weight deacy loss is 6.61, training accuracy is 0.375000, time 218.554 samples/sec
+epoch 11, total_step 171460, total loss is 11.33 , inference loss is 4.72, weight deacy loss is 6.61, training accuracy is 0.531250, time 226.627 samples/sec
+epoch 11, total_step 171480, total loss is 10.97 , inference loss is 4.37, weight deacy loss is 6.61, training accuracy is 0.687500, time 223.633 samples/sec
+epoch 11, total_step 171500, total loss is 11.23 , inference loss is 4.62, weight deacy loss is 6.61, training accuracy is 0.687500, time 202.886 samples/sec
+epoch 11, total_step 171520, total loss is 11.76 , inference loss is 5.15, weight deacy loss is 6.61, training accuracy is 0.562500, time 218.400 samples/sec
+epoch 11, total_step 171540, total loss is 12.23 , inference loss is 5.62, weight deacy loss is 6.61, training accuracy is 0.531250, time 226.769 samples/sec
+epoch 11, total_step 171560, total loss is 9.78 , inference loss is 3.17, weight deacy loss is 6.61, training accuracy is 0.750000, time 219.859 samples/sec
+epoch 11, total_step 171580, total loss is 11.13 , inference loss is 4.53, weight deacy loss is 6.61, training accuracy is 0.593750, time 227.107 samples/sec
+epoch 11, total_step 171600, total loss is 10.87 , inference loss is 4.27, weight deacy loss is 6.61, training accuracy is 0.625000, time 12.277 samples/sec
+epoch 11, total_step 171620, total loss is 15.94 , inference loss is 9.33, weight deacy loss is 6.60, training accuracy is 0.562500, time 243.251 samples/sec
+epoch 11, total_step 171640, total loss is 10.31 , inference loss is 3.71, weight deacy loss is 6.60, training accuracy is 0.593750, time 217.904 samples/sec
+epoch 11, total_step 171660, total loss is 10.15 , inference loss is 3.54, weight deacy loss is 6.60, training accuracy is 0.718750, time 218.828 samples/sec
+epoch 11, total_step 171680, total loss is 12.10 , inference loss is 5.49, weight deacy loss is 6.60, training accuracy is 0.718750, time 225.224 samples/sec
+epoch 11, total_step 171700, total loss is 9.73 , inference loss is 3.13, weight deacy loss is 6.60, training accuracy is 0.812500, time 219.298 samples/sec
+epoch 11, total_step 171720, total loss is 9.99 , inference loss is 3.38, weight deacy loss is 6.60, training accuracy is 0.750000, time 216.082 samples/sec
+epoch 11, total_step 171740, total loss is 14.52 , inference loss is 7.91, weight deacy loss is 6.60, training accuracy is 0.593750, time 219.116 samples/sec
+epoch 11, total_step 171760, total loss is 14.73 , inference loss is 8.12, weight deacy loss is 6.60, training accuracy is 0.437500, time 223.913 samples/sec
+epoch 11, total_step 171780, total loss is 12.99 , inference loss is 6.39, weight deacy loss is 6.60, training accuracy is 0.625000, time 295.785 samples/sec
+epoch 11, total_step 171800, total loss is 11.54 , inference loss is 4.94, weight deacy loss is 6.60, training accuracy is 0.593750, time 199.725 samples/sec
+epoch 11, total_step 171820, total loss is 11.54 , inference loss is 4.94, weight deacy loss is 6.60, training accuracy is 0.593750, time 221.743 samples/sec
+epoch 11, total_step 171840, total loss is 12.91 , inference loss is 6.30, weight deacy loss is 6.60, training accuracy is 0.593750, time 216.751 samples/sec
+epoch 11, total_step 171860, total loss is 17.82 , inference loss is 11.22, weight deacy loss is 6.60, training accuracy is 0.437500, time 237.610 samples/sec
+epoch 11, total_step 171880, total loss is 12.46 , inference loss is 5.86, weight deacy loss is 6.60, training accuracy is 0.593750, time 203.021 samples/sec
+epoch 11, total_step 171900, total loss is 12.96 , inference loss is 6.35, weight deacy loss is 6.60, training accuracy is 0.468750, time 12.484 samples/sec
+epoch 11, total_step 171920, total loss is 15.48 , inference loss is 8.88, weight deacy loss is 6.60, training accuracy is 0.468750, time 226.407 samples/sec
+epoch 11, total_step 171940, total loss is 14.27 , inference loss is 7.67, weight deacy loss is 6.60, training accuracy is 0.531250, time 204.644 samples/sec
+epoch 11, total_step 171960, total loss is 17.07 , inference loss is 10.46, weight deacy loss is 6.60, training accuracy is 0.437500, time 227.279 samples/sec
+epoch 11, total_step 171980, total loss is 14.52 , inference loss is 7.91, weight deacy loss is 6.60, training accuracy is 0.375000, time 221.507 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.403120999999967
+best_threshold_index 120 0.9888888888888889
+best_threshold_index 119 0.9881481481481481
+best_threshold_index 120 0.9885185185185185
+best_threshold_index 120 0.9894444444444445
+best_threshold_index 120 0.9888888888888889
+best_threshold_index 120 0.9879629629629629
+best_threshold_index 120 0.9888888888888889
+best_threshold_index 120 0.9877777777777778
+best_threshold_index 120 0.9874074074074074
+best_threshold_index 119 0.9874074074074074
+[lfw][172000]XNorm: 21.122836
+[lfw][172000]Accuracy-Flip: 0.98800+-0.00572
+test accuracy is: 0.9879999999999999
+epoch 11, total_step 172000, total loss is 13.81 , inference loss is 7.20, weight deacy loss is 6.60, training accuracy is 0.437500, time 16.689 samples/sec
+epoch 11, total_step 172020, total loss is 13.83 , inference loss is 7.23, weight deacy loss is 6.60, training accuracy is 0.500000, time 218.113 samples/sec
+epoch 11, total_step 172040, total loss is 20.72 , inference loss is 14.12, weight deacy loss is 6.60, training accuracy is 0.375000, time 226.416 samples/sec
+epoch 11, total_step 172060, total loss is 14.08 , inference loss is 7.47, weight deacy loss is 6.60, training accuracy is 0.500000, time 216.769 samples/sec
+epoch 11, total_step 172080, total loss is 16.26 , inference loss is 9.66, weight deacy loss is 6.60, training accuracy is 0.468750, time 224.481 samples/sec
+epoch 11, total_step 172100, total loss is 18.53 , inference loss is 11.93, weight deacy loss is 6.60, training accuracy is 0.312500, time 237.653 samples/sec
+epoch 11, total_step 172120, total loss is 13.28 , inference loss is 6.68, weight deacy loss is 6.60, training accuracy is 0.531250, time 218.980 samples/sec
+epoch 11, total_step 172140, total loss is 13.56 , inference loss is 6.96, weight deacy loss is 6.60, training accuracy is 0.500000, time 219.809 samples/sec
+epoch 11, total_step 172160, total loss is 15.50 , inference loss is 8.90, weight deacy loss is 6.60, training accuracy is 0.375000, time 295.573 samples/sec
+epoch 11, total_step 172180, total loss is 15.97 , inference loss is 9.36, weight deacy loss is 6.60, training accuracy is 0.500000, time 230.255 samples/sec
+epoch 11, total_step 172200, total loss is 17.55 , inference loss is 10.95, weight deacy loss is 6.60, training accuracy is 0.500000, time 12.859 samples/sec
+epoch 11, total_step 172220, total loss is 15.33 , inference loss is 8.73, weight deacy loss is 6.60, training accuracy is 0.375000, time 219.388 samples/sec
+epoch 11, total_step 172240, total loss is 13.82 , inference loss is 7.22, weight deacy loss is 6.60, training accuracy is 0.468750, time 222.291 samples/sec
+epoch 11, total_step 172260, total loss is 9.55 , inference loss is 2.95, weight deacy loss is 6.60, training accuracy is 0.750000, time 230.436 samples/sec
+epoch 11, total_step 172280, total loss is 11.24 , inference loss is 4.63, weight deacy loss is 6.60, training accuracy is 0.593750, time 224.779 samples/sec
+epoch 11, total_step 172300, total loss is 14.47 , inference loss is 7.87, weight deacy loss is 6.60, training accuracy is 0.500000, time 239.568 samples/sec
+epoch 11, total_step 172320, total loss is 14.73 , inference loss is 8.13, weight deacy loss is 6.60, training accuracy is 0.343750, time 225.237 samples/sec
+epoch 11, total_step 172340, total loss is 14.17 , inference loss is 7.57, weight deacy loss is 6.60, training accuracy is 0.593750, time 224.545 samples/sec
+epoch 11, total_step 172360, total loss is 16.40 , inference loss is 9.80, weight deacy loss is 6.60, training accuracy is 0.375000, time 226.558 samples/sec
+epoch 11, total_step 172380, total loss is 16.44 , inference loss is 9.84, weight deacy loss is 6.60, training accuracy is 0.437500, time 207.400 samples/sec
+epoch 11, total_step 172400, total loss is 15.44 , inference loss is 8.84, weight deacy loss is 6.60, training accuracy is 0.593750, time 225.578 samples/sec
+epoch 11, total_step 172420, total loss is 14.06 , inference loss is 7.45, weight deacy loss is 6.60, training accuracy is 0.437500, time 299.742 samples/sec
+epoch 11, total_step 172440, total loss is 12.24 , inference loss is 5.64, weight deacy loss is 6.60, training accuracy is 0.500000, time 224.689 samples/sec
+epoch 11, total_step 172460, total loss is 12.62 , inference loss is 6.02, weight deacy loss is 6.60, training accuracy is 0.500000, time 227.792 samples/sec
+epoch 11, total_step 172480, total loss is 13.67 , inference loss is 7.07, weight deacy loss is 6.60, training accuracy is 0.468750, time 224.599 samples/sec
+epoch 11, total_step 172500, total loss is 13.26 , inference loss is 6.66, weight deacy loss is 6.60, training accuracy is 0.375000, time 13.239 samples/sec
+epoch 11, total_step 172520, total loss is 11.45 , inference loss is 4.85, weight deacy loss is 6.60, training accuracy is 0.437500, time 224.366 samples/sec
+epoch 11, total_step 172540, total loss is 17.56 , inference loss is 10.96, weight deacy loss is 6.60, training accuracy is 0.375000, time 225.721 samples/sec
+epoch 11, total_step 172560, total loss is 14.92 , inference loss is 8.32, weight deacy loss is 6.60, training accuracy is 0.375000, time 222.160 samples/sec
+epoch 11, total_step 172580, total loss is 15.90 , inference loss is 9.30, weight deacy loss is 6.60, training accuracy is 0.406250, time 225.283 samples/sec
+epoch 11, total_step 172600, total loss is 12.78 , inference loss is 6.18, weight deacy loss is 6.60, training accuracy is 0.468750, time 201.821 samples/sec
+epoch 11, total_step 172620, total loss is 13.70 , inference loss is 7.10, weight deacy loss is 6.60, training accuracy is 0.406250, time 217.592 samples/sec
+epoch 11, total_step 172640, total loss is 13.14 , inference loss is 6.54, weight deacy loss is 6.60, training accuracy is 0.500000, time 222.441 samples/sec
+epoch 11, total_step 172660, total loss is 14.87 , inference loss is 8.27, weight deacy loss is 6.60, training accuracy is 0.500000, time 215.580 samples/sec
+epoch 11, total_step 172680, total loss is 12.83 , inference loss is 6.23, weight deacy loss is 6.60, training accuracy is 0.468750, time 296.953 samples/sec
+epoch 11, total_step 172700, total loss is 12.87 , inference loss is 6.27, weight deacy loss is 6.60, training accuracy is 0.375000, time 219.859 samples/sec
+epoch 11, total_step 172720, total loss is 12.47 , inference loss is 5.87, weight deacy loss is 6.60, training accuracy is 0.531250, time 224.680 samples/sec
+epoch 11, total_step 172740, total loss is 15.88 , inference loss is 9.28, weight deacy loss is 6.60, training accuracy is 0.375000, time 232.223 samples/sec
+epoch 11, total_step 172760, total loss is 11.98 , inference loss is 5.38, weight deacy loss is 6.60, training accuracy is 0.625000, time 221.070 samples/sec
+epoch 11, total_step 172780, total loss is 11.10 , inference loss is 4.50, weight deacy loss is 6.60, training accuracy is 0.562500, time 210.215 samples/sec
+epoch 11, total_step 172800, total loss is 14.97 , inference loss is 8.37, weight deacy loss is 6.60, training accuracy is 0.593750, time 12.547 samples/sec
+epoch 11, total_step 172820, total loss is 15.62 , inference loss is 9.02, weight deacy loss is 6.60, training accuracy is 0.437500, time 227.705 samples/sec
+epoch 11, total_step 172840, total loss is 17.47 , inference loss is 10.87, weight deacy loss is 6.60, training accuracy is 0.437500, time 221.480 samples/sec
+epoch 11, total_step 172860, total loss is 13.09 , inference loss is 6.49, weight deacy loss is 6.60, training accuracy is 0.625000, time 220.928 samples/sec
+epoch 11, total_step 172880, total loss is 13.13 , inference loss is 6.53, weight deacy loss is 6.60, training accuracy is 0.687500, time 224.871 samples/sec
+epoch 11, total_step 172900, total loss is 12.96 , inference loss is 6.36, weight deacy loss is 6.60, training accuracy is 0.437500, time 236.165 samples/sec
+epoch 11, total_step 172920, total loss is 14.08 , inference loss is 7.48, weight deacy loss is 6.60, training accuracy is 0.468750, time 231.622 samples/sec
+epoch 11, total_step 172940, total loss is 18.58 , inference loss is 11.98, weight deacy loss is 6.60, training accuracy is 0.218750, time 299.511 samples/sec
+epoch 11, total_step 172960, total loss is 14.15 , inference loss is 7.55, weight deacy loss is 6.60, training accuracy is 0.406250, time 222.873 samples/sec
+epoch 11, total_step 172980, total loss is 12.84 , inference loss is 6.24, weight deacy loss is 6.60, training accuracy is 0.562500, time 220.359 samples/sec
+epoch 11, total_step 173000, total loss is 14.31 , inference loss is 7.72, weight deacy loss is 6.60, training accuracy is 0.593750, time 222.752 samples/sec
+epoch 11, total_step 173020, total loss is 13.19 , inference loss is 6.59, weight deacy loss is 6.60, training accuracy is 0.437500, time 211.051 samples/sec
+epoch 11, total_step 173040, total loss is 13.21 , inference loss is 6.62, weight deacy loss is 6.60, training accuracy is 0.437500, time 223.985 samples/sec
+epoch 11, total_step 173060, total loss is 10.81 , inference loss is 4.21, weight deacy loss is 6.60, training accuracy is 0.593750, time 223.423 samples/sec
+epoch 11, total_step 173080, total loss is 17.48 , inference loss is 10.88, weight deacy loss is 6.60, training accuracy is 0.250000, time 225.152 samples/sec
+epoch 11, total_step 173100, total loss is 13.37 , inference loss is 6.77, weight deacy loss is 6.60, training accuracy is 0.437500, time 12.750 samples/sec
+epoch 11, total_step 173120, total loss is 15.98 , inference loss is 9.38, weight deacy loss is 6.60, training accuracy is 0.562500, time 219.339 samples/sec
+epoch 11, total_step 173140, total loss is 15.04 , inference loss is 8.44, weight deacy loss is 6.60, training accuracy is 0.375000, time 223.992 samples/sec
+epoch 11, total_step 173160, total loss is 18.10 , inference loss is 11.50, weight deacy loss is 6.60, training accuracy is 0.406250, time 226.195 samples/sec
+epoch 11, total_step 173180, total loss is 13.54 , inference loss is 6.95, weight deacy loss is 6.60, training accuracy is 0.562500, time 216.429 samples/sec
+epoch 11, total_step 173200, total loss is 14.24 , inference loss is 7.64, weight deacy loss is 6.60, training accuracy is 0.500000, time 299.228 samples/sec
+epoch 11, total_step 173220, total loss is 13.82 , inference loss is 7.22, weight deacy loss is 6.60, training accuracy is 0.468750, time 211.941 samples/sec
+epoch 11, total_step 173240, total loss is 13.75 , inference loss is 7.16, weight deacy loss is 6.60, training accuracy is 0.437500, time 219.872 samples/sec
+epoch 11, total_step 173260, total loss is 13.57 , inference loss is 6.98, weight deacy loss is 6.60, training accuracy is 0.500000, time 228.412 samples/sec
+epoch 11, total_step 173280, total loss is 19.17 , inference loss is 12.58, weight deacy loss is 6.60, training accuracy is 0.343750, time 218.042 samples/sec
+epoch 11, total_step 173300, total loss is 13.68 , inference loss is 7.09, weight deacy loss is 6.60, training accuracy is 0.500000, time 206.828 samples/sec
+epoch 11, total_step 173320, total loss is 13.69 , inference loss is 7.09, weight deacy loss is 6.60, training accuracy is 0.406250, time 222.869 samples/sec
+epoch 11, total_step 173340, total loss is 14.34 , inference loss is 7.75, weight deacy loss is 6.60, training accuracy is 0.437500, time 222.087 samples/sec
+epoch 11, total_step 173360, total loss is 10.23 , inference loss is 3.64, weight deacy loss is 6.60, training accuracy is 0.562500, time 213.037 samples/sec
+epoch 11, total_step 173380, total loss is 14.06 , inference loss is 7.47, weight deacy loss is 6.60, training accuracy is 0.437500, time 215.190 samples/sec
+epoch 11, total_step 173400, total loss is 12.26 , inference loss is 5.66, weight deacy loss is 6.60, training accuracy is 0.656250, time 12.520 samples/sec
+epoch 11, total_step 173420, total loss is 13.49 , inference loss is 6.89, weight deacy loss is 6.60, training accuracy is 0.562500, time 214.691 samples/sec
+epoch 11, total_step 173440, total loss is 12.85 , inference loss is 6.25, weight deacy loss is 6.59, training accuracy is 0.531250, time 218.563 samples/sec
+epoch 11, total_step 173460, total loss is 14.16 , inference loss is 7.57, weight deacy loss is 6.59, training accuracy is 0.500000, time 215.667 samples/sec
+epoch 11, total_step 173480, total loss is 12.78 , inference loss is 6.18, weight deacy loss is 6.59, training accuracy is 0.468750, time 206.373 samples/sec
+epoch 11, total_step 173500, total loss is 13.54 , inference loss is 6.94, weight deacy loss is 6.59, training accuracy is 0.437500, time 217.677 samples/sec
+epoch 11, total_step 173520, total loss is 10.46 , inference loss is 3.87, weight deacy loss is 6.59, training accuracy is 0.562500, time 297.198 samples/sec
+epoch 11, total_step 173540, total loss is 12.38 , inference loss is 5.78, weight deacy loss is 6.59, training accuracy is 0.625000, time 297.837 samples/sec
+epoch 11, total_step 173560, total loss is 14.14 , inference loss is 7.55, weight deacy loss is 6.59, training accuracy is 0.531250, time 177.540 samples/sec
+epoch 11, total_step 173580, total loss is 13.98 , inference loss is 7.38, weight deacy loss is 6.59, training accuracy is 0.343750, time 226.131 samples/sec
+epoch 11, total_step 173600, total loss is 11.74 , inference loss is 5.14, weight deacy loss is 6.59, training accuracy is 0.500000, time 225.909 samples/sec
+epoch 11, total_step 173620, total loss is 13.21 , inference loss is 6.61, weight deacy loss is 6.59, training accuracy is 0.406250, time 226.865 samples/sec
+epoch 11, total_step 173640, total loss is 11.69 , inference loss is 5.09, weight deacy loss is 6.59, training accuracy is 0.468750, time 216.948 samples/sec
+epoch 11, total_step 173660, total loss is 12.48 , inference loss is 5.89, weight deacy loss is 6.59, training accuracy is 0.531250, time 300.073 samples/sec
+epoch 11, total_step 173680, total loss is 13.75 , inference loss is 7.16, weight deacy loss is 6.59, training accuracy is 0.437500, time 224.775 samples/sec
+epoch 11, total_step 173700, total loss is 16.77 , inference loss is 10.17, weight deacy loss is 6.59, training accuracy is 0.406250, time 13.054 samples/sec
+epoch 11, total_step 173720, total loss is 14.55 , inference loss is 7.96, weight deacy loss is 6.59, training accuracy is 0.375000, time 200.781 samples/sec
+epoch 11, total_step 173740, total loss is 13.77 , inference loss is 7.18, weight deacy loss is 6.59, training accuracy is 0.500000, time 220.796 samples/sec
+epoch 11, total_step 173760, total loss is 14.95 , inference loss is 8.36, weight deacy loss is 6.59, training accuracy is 0.593750, time 209.326 samples/sec
+epoch 11, total_step 173780, total loss is 10.21 , inference loss is 3.62, weight deacy loss is 6.59, training accuracy is 0.593750, time 226.815 samples/sec
+epoch 11, total_step 173800, total loss is 16.08 , inference loss is 9.49, weight deacy loss is 6.59, training accuracy is 0.468750, time 174.757 samples/sec
+epoch 11, total_step 173820, total loss is 14.97 , inference loss is 8.38, weight deacy loss is 6.59, training accuracy is 0.531250, time 220.039 samples/sec
+epoch 11, total_step 173840, total loss is 10.68 , inference loss is 4.09, weight deacy loss is 6.59, training accuracy is 0.562500, time 225.543 samples/sec
+epoch 11, total_step 173860, total loss is 12.63 , inference loss is 6.04, weight deacy loss is 6.59, training accuracy is 0.500000, time 227.870 samples/sec
+epoch 11, total_step 173880, total loss is 9.97 , inference loss is 3.38, weight deacy loss is 6.59, training accuracy is 0.593750, time 224.662 samples/sec
+epoch 11, total_step 173900, total loss is 11.98 , inference loss is 5.39, weight deacy loss is 6.59, training accuracy is 0.437500, time 210.614 samples/sec
+epoch 11, total_step 173920, total loss is 13.48 , inference loss is 6.88, weight deacy loss is 6.59, training accuracy is 0.531250, time 298.980 samples/sec
+epoch 11, total_step 173940, total loss is 14.75 , inference loss is 8.15, weight deacy loss is 6.59, training accuracy is 0.375000, time 227.607 samples/sec
+epoch 11, total_step 173960, total loss is 12.35 , inference loss is 5.75, weight deacy loss is 6.59, training accuracy is 0.468750, time 226.779 samples/sec
+epoch 11, total_step 173980, total loss is 16.49 , inference loss is 9.90, weight deacy loss is 6.59, training accuracy is 0.375000, time 223.632 samples/sec
+testing verification..
+(12000, 512)
+infer time 26.010637000000006
+best_threshold_index 117 0.9879629629629629
+best_threshold_index 112 0.9877777777777778
+best_threshold_index 117 0.9883333333333333
+best_threshold_index 117 0.9888888888888889
+best_threshold_index 112 0.9881481481481481
+best_threshold_index 117 0.9872222222222222
+best_threshold_index 119 0.9875925925925926
+best_threshold_index 117 0.987037037037037
+best_threshold_index 117 0.9868518518518519
+best_threshold_index 117 0.987037037037037
+[lfw][174000]XNorm: 21.142689
+[lfw][174000]Accuracy-Flip: 0.98650+-0.00608
+test accuracy is: 0.9865
+epoch 11, total_step 174000, total loss is 14.07 , inference loss is 7.48, weight deacy loss is 6.59, training accuracy is 0.562500, time 16.154 samples/sec
+epoch 11, total_step 174020, total loss is 14.77 , inference loss is 8.17, weight deacy loss is 6.59, training accuracy is 0.406250, time 224.382 samples/sec
+epoch 11, total_step 174040, total loss is 16.68 , inference loss is 10.09, weight deacy loss is 6.59, training accuracy is 0.375000, time 226.122 samples/sec
+epoch 11, total_step 174060, total loss is 10.77 , inference loss is 4.18, weight deacy loss is 6.59, training accuracy is 0.531250, time 225.415 samples/sec
+epoch 11, total_step 174080, total loss is 16.68 , inference loss is 10.09, weight deacy loss is 6.59, training accuracy is 0.343750, time 199.096 samples/sec
+epoch 11, total_step 174100, total loss is 14.14 , inference loss is 7.55, weight deacy loss is 6.59, training accuracy is 0.437500, time 220.799 samples/sec
+epoch 11, total_step 174120, total loss is 12.17 , inference loss is 5.58, weight deacy loss is 6.59, training accuracy is 0.593750, time 228.740 samples/sec
+epoch 11, total_step 174140, total loss is 14.44 , inference loss is 7.85, weight deacy loss is 6.59, training accuracy is 0.562500, time 226.260 samples/sec
+epoch 11, total_step 174160, total loss is 12.50 , inference loss is 5.91, weight deacy loss is 6.59, training accuracy is 0.531250, time 228.485 samples/sec
+epoch 11, total_step 174180, total loss is 12.83 , inference loss is 6.25, weight deacy loss is 6.59, training accuracy is 0.437500, time 203.617 samples/sec
+epoch 11, total_step 174200, total loss is 15.06 , inference loss is 8.47, weight deacy loss is 6.59, training accuracy is 0.375000, time 223.036 samples/sec
+epoch 11, total_step 174220, total loss is 12.37 , inference loss is 5.78, weight deacy loss is 6.59, training accuracy is 0.437500, time 226.550 samples/sec
+epoch 11, total_step 174240, total loss is 15.21 , inference loss is 8.62, weight deacy loss is 6.59, training accuracy is 0.437500, time 225.829 samples/sec
+epoch 11, total_step 174260, total loss is 18.24 , inference loss is 11.65, weight deacy loss is 6.59, training accuracy is 0.312500, time 225.429 samples/sec
+epoch 11, total_step 174280, total loss is 13.46 , inference loss is 6.88, weight deacy loss is 6.59, training accuracy is 0.500000, time 299.432 samples/sec
+epoch 11, total_step 174300, total loss is 13.62 , inference loss is 7.03, weight deacy loss is 6.59, training accuracy is 0.468750, time 12.482 samples/sec
+epoch 11, total_step 174320, total loss is 17.74 , inference loss is 11.16, weight deacy loss is 6.59, training accuracy is 0.531250, time 221.001 samples/sec
+epoch 11, total_step 174340, total loss is 14.93 , inference loss is 8.34, weight deacy loss is 6.59, training accuracy is 0.437500, time 220.277 samples/sec
+epoch 11, total_step 174360, total loss is 11.35 , inference loss is 4.76, weight deacy loss is 6.59, training accuracy is 0.593750, time 214.308 samples/sec
+epoch 11, total_step 174380, total loss is 15.30 , inference loss is 8.72, weight deacy loss is 6.59, training accuracy is 0.375000, time 220.936 samples/sec
+epoch 11, total_step 174400, total loss is 14.34 , inference loss is 7.75, weight deacy loss is 6.59, training accuracy is 0.437500, time 226.728 samples/sec
+epoch 11, total_step 174420, total loss is 14.34 , inference loss is 7.75, weight deacy loss is 6.59, training accuracy is 0.531250, time 225.648 samples/sec
+epoch 11, total_step 174440, total loss is 17.07 , inference loss is 10.48, weight deacy loss is 6.59, training accuracy is 0.468750, time 227.669 samples/sec
+epoch 11, total_step 174460, total loss is 15.72 , inference loss is 9.13, weight deacy loss is 6.59, training accuracy is 0.406250, time 225.506 samples/sec
+epoch 11, total_step 174480, total loss is 10.66 , inference loss is 4.07, weight deacy loss is 6.59, training accuracy is 0.562500, time 223.171 samples/sec
+epoch 11, total_step 174500, total loss is 14.56 , inference loss is 7.97, weight deacy loss is 6.59, training accuracy is 0.468750, time 222.580 samples/sec
+epoch 11, total_step 174520, total loss is 13.42 , inference loss is 6.83, weight deacy loss is 6.59, training accuracy is 0.500000, time 225.316 samples/sec
+epoch 11, total_step 174540, total loss is 17.49 , inference loss is 10.91, weight deacy loss is 6.59, training accuracy is 0.343750, time 298.069 samples/sec
+epoch 11, total_step 174560, total loss is 12.22 , inference loss is 5.63, weight deacy loss is 6.59, training accuracy is 0.531250, time 212.377 samples/sec
+epoch 11, total_step 174580, total loss is 13.24 , inference loss is 6.66, weight deacy loss is 6.59, training accuracy is 0.500000, time 223.478 samples/sec
+epoch 11, total_step 174600, total loss is 13.87 , inference loss is 7.29, weight deacy loss is 6.59, training accuracy is 0.500000, time 12.422 samples/sec
+epoch 11, total_step 174620, total loss is 17.95 , inference loss is 11.37, weight deacy loss is 6.59, training accuracy is 0.312500, time 223.118 samples/sec
+epoch 11, total_step 174640, total loss is 11.97 , inference loss is 5.39, weight deacy loss is 6.59, training accuracy is 0.375000, time 225.531 samples/sec
+epoch 11, total_step 174660, total loss is 11.92 , inference loss is 5.34, weight deacy loss is 6.59, training accuracy is 0.531250, time 225.013 samples/sec
+epoch 11, total_step 174680, total loss is 13.70 , inference loss is 7.12, weight deacy loss is 6.59, training accuracy is 0.500000, time 222.745 samples/sec
+epoch 11, total_step 174700, total loss is 14.96 , inference loss is 8.38, weight deacy loss is 6.59, training accuracy is 0.468750, time 241.791 samples/sec
+epoch 11, total_step 174720, total loss is 16.78 , inference loss is 10.20, weight deacy loss is 6.58, training accuracy is 0.406250, time 221.931 samples/sec
+epoch 11, total_step 174740, total loss is 14.98 , inference loss is 8.40, weight deacy loss is 6.58, training accuracy is 0.375000, time 222.172 samples/sec
+epoch 11, total_step 174760, total loss is 15.52 , inference loss is 8.93, weight deacy loss is 6.58, training accuracy is 0.468750, time 228.407 samples/sec
+epoch 11, total_step 174780, total loss is 11.67 , inference loss is 5.08, weight deacy loss is 6.58, training accuracy is 0.593750, time 219.760 samples/sec
+epoch 11, total_step 174800, total loss is 15.19 , inference loss is 8.61, weight deacy loss is 6.58, training accuracy is 0.437500, time 296.065 samples/sec
+epoch 11, total_step 174820, total loss is 14.49 , inference loss is 7.91, weight deacy loss is 6.58, training accuracy is 0.437500, time 224.445 samples/sec
+epoch 11, total_step 174840, total loss is 14.27 , inference loss is 7.68, weight deacy loss is 6.58, training accuracy is 0.375000, time 225.114 samples/sec
+epoch 11, total_step 174860, total loss is 11.42 , inference loss is 4.84, weight deacy loss is 6.58, training accuracy is 0.468750, time 230.950 samples/sec
+epoch 11, total_step 174880, total loss is 14.27 , inference loss is 7.69, weight deacy loss is 6.58, training accuracy is 0.406250, time 224.655 samples/sec
+epoch 11, total_step 174900, total loss is 12.83 , inference loss is 6.25, weight deacy loss is 6.58, training accuracy is 0.562500, time 12.581 samples/sec
+epoch 11, total_step 174920, total loss is 14.53 , inference loss is 7.94, weight deacy loss is 6.58, training accuracy is 0.468750, time 226.484 samples/sec
+epoch 11, total_step 174940, total loss is 15.81 , inference loss is 9.23, weight deacy loss is 6.58, training accuracy is 0.468750, time 227.098 samples/sec
+epoch 11, total_step 174960, total loss is 10.85 , inference loss is 4.27, weight deacy loss is 6.58, training accuracy is 0.593750, time 224.426 samples/sec
+epoch 11, total_step 174980, total loss is 12.99 , inference loss is 6.41, weight deacy loss is 6.58, training accuracy is 0.531250, time 219.260 samples/sec
+epoch 11, total_step 175000, total loss is 15.19 , inference loss is 8.60, weight deacy loss is 6.58, training accuracy is 0.437500, time 222.247 samples/sec
+epoch 11, total_step 175020, total loss is 14.30 , inference loss is 7.72, weight deacy loss is 6.58, training accuracy is 0.343750, time 226.272 samples/sec
+epoch 11, total_step 175040, total loss is 13.65 , inference loss is 7.06, weight deacy loss is 6.58, training accuracy is 0.375000, time 227.872 samples/sec
+epoch 11, total_step 175060, total loss is 12.18 , inference loss is 5.60, weight deacy loss is 6.58, training accuracy is 0.437500, time 296.248 samples/sec
+epoch 11, total_step 175080, total loss is 18.12 , inference loss is 11.54, weight deacy loss is 6.58, training accuracy is 0.312500, time 238.887 samples/sec
+epoch 11, total_step 175100, total loss is 12.23 , inference loss is 5.65, weight deacy loss is 6.58, training accuracy is 0.531250, time 216.371 samples/sec
+epoch 11, total_step 175120, total loss is 10.77 , inference loss is 4.19, weight deacy loss is 6.58, training accuracy is 0.656250, time 224.716 samples/sec
+epoch 11, total_step 175140, total loss is 12.57 , inference loss is 5.99, weight deacy loss is 6.58, training accuracy is 0.406250, time 211.696 samples/sec
+epoch 11, total_step 175160, total loss is 13.73 , inference loss is 7.15, weight deacy loss is 6.58, training accuracy is 0.656250, time 219.095 samples/sec
+epoch 11, total_step 175180, total loss is 12.65 , inference loss is 6.07, weight deacy loss is 6.58, training accuracy is 0.531250, time 216.206 samples/sec
+epoch 11, total_step 175200, total loss is 19.30 , inference loss is 12.72, weight deacy loss is 6.58, training accuracy is 0.281250, time 12.947 samples/sec
+epoch 11, total_step 175220, total loss is 11.66 , inference loss is 5.08, weight deacy loss is 6.58, training accuracy is 0.468750, time 213.508 samples/sec
+epoch 11, total_step 175240, total loss is 15.34 , inference loss is 8.76, weight deacy loss is 6.58, training accuracy is 0.343750, time 216.042 samples/sec
+epoch 11, total_step 175260, total loss is 14.78 , inference loss is 8.20, weight deacy loss is 6.58, training accuracy is 0.312500, time 226.053 samples/sec
+epoch 11, total_step 175280, total loss is 16.04 , inference loss is 9.46, weight deacy loss is 6.58, training accuracy is 0.437500, time 297.271 samples/sec
+epoch 11, total_step 175300, total loss is 13.43 , inference loss is 6.86, weight deacy loss is 6.58, training accuracy is 0.375000, time 212.480 samples/sec
+epoch 11, total_step 175320, total loss is 17.34 , inference loss is 10.76, weight deacy loss is 6.58, training accuracy is 0.500000, time 221.867 samples/sec
+epoch 11, total_step 175340, total loss is 16.92 , inference loss is 10.34, weight deacy loss is 6.58, training accuracy is 0.375000, time 219.993 samples/sec
+epoch 11, total_step 175360, total loss is 12.79 , inference loss is 6.21, weight deacy loss is 6.58, training accuracy is 0.468750, time 226.492 samples/sec
+epoch 11, total_step 175380, total loss is 12.44 , inference loss is 5.86, weight deacy loss is 6.58, training accuracy is 0.562500, time 232.301 samples/sec
+epoch 11, total_step 175400, total loss is 17.87 , inference loss is 11.30, weight deacy loss is 6.58, training accuracy is 0.406250, time 221.855 samples/sec
+epoch 11, total_step 175420, total loss is 12.05 , inference loss is 5.47, weight deacy loss is 6.58, training accuracy is 0.500000, time 226.552 samples/sec
+epoch 11, total_step 175440, total loss is 13.02 , inference loss is 6.44, weight deacy loss is 6.58, training accuracy is 0.437500, time 226.786 samples/sec
+epoch 11, total_step 175460, total loss is 16.75 , inference loss is 10.18, weight deacy loss is 6.58, training accuracy is 0.406250, time 227.717 samples/sec
+epoch 11, total_step 175480, total loss is 14.94 , inference loss is 8.37, weight deacy loss is 6.58, training accuracy is 0.500000, time 177.621 samples/sec
+epoch 11, total_step 175500, total loss is 20.48 , inference loss is 13.90, weight deacy loss is 6.58, training accuracy is 0.343750, time 11.497 samples/sec
+epoch 11, total_step 175520, total loss is 15.35 , inference loss is 8.77, weight deacy loss is 6.58, training accuracy is 0.406250, time 216.685 samples/sec
+epoch 11, total_step 175540, total loss is 18.08 , inference loss is 11.50, weight deacy loss is 6.58, training accuracy is 0.156250, time 220.347 samples/sec
+epoch 11, total_step 175560, total loss is 13.09 , inference loss is 6.51, weight deacy loss is 6.58, training accuracy is 0.562500, time 227.517 samples/sec
+epoch 11, total_step 175580, total loss is 13.89 , inference loss is 7.31, weight deacy loss is 6.58, training accuracy is 0.437500, time 222.068 samples/sec
+epoch 11, total_step 175600, total loss is 15.77 , inference loss is 9.19, weight deacy loss is 6.58, training accuracy is 0.406250, time 219.361 samples/sec
+epoch 11, total_step 175620, total loss is 13.37 , inference loss is 6.79, weight deacy loss is 6.58, training accuracy is 0.468750, time 220.190 samples/sec
+epoch 11, total_step 175640, total loss is 12.68 , inference loss is 6.11, weight deacy loss is 6.58, training accuracy is 0.468750, time 217.522 samples/sec
+epoch 11, total_step 175660, total loss is 13.72 , inference loss is 7.15, weight deacy loss is 6.58, training accuracy is 0.312500, time 217.617 samples/sec
+epoch 11, total_step 175680, total loss is 15.01 , inference loss is 8.43, weight deacy loss is 6.58, training accuracy is 0.500000, time 222.660 samples/sec
+epoch 11, total_step 175700, total loss is 16.35 , inference loss is 9.78, weight deacy loss is 6.58, training accuracy is 0.406250, time 213.024 samples/sec
+epoch 11, total_step 175720, total loss is 14.54 , inference loss is 7.96, weight deacy loss is 6.58, training accuracy is 0.437500, time 225.511 samples/sec
+epoch 11, total_step 175740, total loss is 15.38 , inference loss is 8.80, weight deacy loss is 6.58, training accuracy is 0.375000, time 226.382 samples/sec
+epoch 11, total_step 175760, total loss is 14.40 , inference loss is 7.83, weight deacy loss is 6.58, training accuracy is 0.468750, time 299.313 samples/sec
+epoch 11, total_step 175780, total loss is 19.35 , inference loss is 12.77, weight deacy loss is 6.58, training accuracy is 0.312500, time 224.256 samples/sec
+epoch 11, total_step 175800, total loss is 16.98 , inference loss is 10.41, weight deacy loss is 6.58, training accuracy is 0.250000, time 12.210 samples/sec
+epoch 11, total_step 175820, total loss is 12.43 , inference loss is 5.86, weight deacy loss is 6.58, training accuracy is 0.500000, time 224.587 samples/sec
+epoch 11, total_step 175840, total loss is 9.97 , inference loss is 3.39, weight deacy loss is 6.58, training accuracy is 0.625000, time 241.639 samples/sec
+epoch 11, total_step 175860, total loss is 17.12 , inference loss is 10.55, weight deacy loss is 6.58, training accuracy is 0.406250, time 224.772 samples/sec
+epoch 11, total_step 175880, total loss is 14.17 , inference loss is 7.60, weight deacy loss is 6.57, training accuracy is 0.437500, time 223.158 samples/sec
+epoch 11, total_step 175900, total loss is 11.07 , inference loss is 4.49, weight deacy loss is 6.57, training accuracy is 0.562500, time 218.146 samples/sec
+epoch 11, total_step 175920, total loss is 9.63 , inference loss is 3.05, weight deacy loss is 6.57, training accuracy is 0.656250, time 225.131 samples/sec
+epoch 11, total_step 175940, total loss is 13.46 , inference loss is 6.89, weight deacy loss is 6.57, training accuracy is 0.437500, time 224.428 samples/sec
+epoch 11, total_step 175960, total loss is 16.61 , inference loss is 10.03, weight deacy loss is 6.57, training accuracy is 0.406250, time 223.033 samples/sec
+epoch 11, total_step 175980, total loss is 18.64 , inference loss is 12.07, weight deacy loss is 6.57, training accuracy is 0.375000, time 223.277 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.279884000000003
+best_threshold_index 115 0.9896296296296296
+best_threshold_index 123 0.9892592592592593
+best_threshold_index 123 0.9888888888888889
+best_threshold_index 115 0.9892592592592593
+best_threshold_index 116 0.9898148148148148
+best_threshold_index 119 0.9881481481481481
+best_threshold_index 119 0.9892592592592593
+best_threshold_index 115 0.9887037037037038
+best_threshold_index 115 0.9883333333333333
+best_threshold_index 123 0.9885185185185185
+[lfw][176000]XNorm: 22.814379
+[lfw][176000]Accuracy-Flip: 0.98750+-0.00473
+test accuracy is: 0.9875000000000002
+epoch 11, total_step 176000, total loss is 13.79 , inference loss is 7.22, weight deacy loss is 6.57, training accuracy is 0.468750, time 13.457 samples/sec
+epoch 11, total_step 176020, total loss is 13.15 , inference loss is 6.58, weight deacy loss is 6.57, training accuracy is 0.593750, time 296.891 samples/sec
+epoch 11, total_step 176040, total loss is 16.16 , inference loss is 9.58, weight deacy loss is 6.57, training accuracy is 0.437500, time 223.999 samples/sec
+epoch 11, total_step 176060, total loss is 15.60 , inference loss is 9.03, weight deacy loss is 6.57, training accuracy is 0.406250, time 208.156 samples/sec
+epoch 11, total_step 176080, total loss is 17.84 , inference loss is 11.27, weight deacy loss is 6.57, training accuracy is 0.312500, time 214.218 samples/sec
+epoch 11, total_step 176100, total loss is 17.72 , inference loss is 11.14, weight deacy loss is 6.57, training accuracy is 0.406250, time 12.652 samples/sec
+epoch 11, total_step 176120, total loss is 16.33 , inference loss is 9.76, weight deacy loss is 6.57, training accuracy is 0.437500, time 225.209 samples/sec
+epoch 11, total_step 176140, total loss is 17.12 , inference loss is 10.55, weight deacy loss is 6.57, training accuracy is 0.375000, time 206.155 samples/sec
+epoch 11, total_step 176160, total loss is 14.10 , inference loss is 7.52, weight deacy loss is 6.57, training accuracy is 0.437500, time 217.449 samples/sec
+epoch 11, total_step 176180, total loss is 15.68 , inference loss is 9.10, weight deacy loss is 6.57, training accuracy is 0.375000, time 234.357 samples/sec
+epoch 11, total_step 176200, total loss is 11.17 , inference loss is 4.59, weight deacy loss is 6.57, training accuracy is 0.468750, time 210.212 samples/sec
+epoch 11, total_step 176220, total loss is 9.86 , inference loss is 3.28, weight deacy loss is 6.57, training accuracy is 0.593750, time 226.550 samples/sec
+epoch 11, total_step 176240, total loss is 13.34 , inference loss is 6.76, weight deacy loss is 6.57, training accuracy is 0.593750, time 222.449 samples/sec
+epoch 11, total_step 176260, total loss is 13.69 , inference loss is 7.12, weight deacy loss is 6.57, training accuracy is 0.562500, time 224.884 samples/sec
+epoch 11, total_step 176280, total loss is 15.22 , inference loss is 8.65, weight deacy loss is 6.57, training accuracy is 0.312500, time 298.998 samples/sec
+epoch 11, total_step 176300, total loss is 14.07 , inference loss is 7.50, weight deacy loss is 6.57, training accuracy is 0.312500, time 222.950 samples/sec
+epoch 11, total_step 176320, total loss is 12.02 , inference loss is 5.45, weight deacy loss is 6.57, training accuracy is 0.562500, time 225.916 samples/sec
+epoch 11, total_step 176340, total loss is 11.99 , inference loss is 5.42, weight deacy loss is 6.57, training accuracy is 0.437500, time 214.896 samples/sec
+epoch 11, total_step 176360, total loss is 14.32 , inference loss is 7.75, weight deacy loss is 6.57, training accuracy is 0.531250, time 216.114 samples/sec
+epoch 11, total_step 176380, total loss is 13.36 , inference loss is 6.79, weight deacy loss is 6.57, training accuracy is 0.531250, time 220.270 samples/sec
+epoch 11, total_step 176400, total loss is 9.96 , inference loss is 3.39, weight deacy loss is 6.57, training accuracy is 0.625000, time 12.327 samples/sec
+epoch 11, total_step 176420, total loss is 12.01 , inference loss is 5.44, weight deacy loss is 6.57, training accuracy is 0.593750, time 218.529 samples/sec
+epoch 11, total_step 176440, total loss is 13.93 , inference loss is 7.36, weight deacy loss is 6.57, training accuracy is 0.500000, time 220.309 samples/sec
+epoch 11, total_step 176460, total loss is 13.90 , inference loss is 7.33, weight deacy loss is 6.57, training accuracy is 0.531250, time 220.018 samples/sec
+epoch 11, total_step 176480, total loss is 11.47 , inference loss is 4.90, weight deacy loss is 6.57, training accuracy is 0.531250, time 219.143 samples/sec
+epoch 11, total_step 176500, total loss is 10.19 , inference loss is 3.63, weight deacy loss is 6.57, training accuracy is 0.562500, time 225.641 samples/sec
+epoch 11, total_step 176520, total loss is 15.38 , inference loss is 8.81, weight deacy loss is 6.57, training accuracy is 0.562500, time 228.536 samples/sec
+epoch 11, total_step 176540, total loss is 18.22 , inference loss is 11.65, weight deacy loss is 6.57, training accuracy is 0.250000, time 296.669 samples/sec
+epoch 11, total_step 176560, total loss is 11.39 , inference loss is 4.82, weight deacy loss is 6.57, training accuracy is 0.593750, time 222.793 samples/sec
+epoch 11, total_step 176580, total loss is 18.49 , inference loss is 11.92, weight deacy loss is 6.57, training accuracy is 0.406250, time 220.240 samples/sec
+epoch 11, total_step 176600, total loss is 12.80 , inference loss is 6.23, weight deacy loss is 6.57, training accuracy is 0.406250, time 221.555 samples/sec
+epoch 11, total_step 176620, total loss is 18.17 , inference loss is 11.61, weight deacy loss is 6.57, training accuracy is 0.312500, time 240.709 samples/sec
+epoch 11, total_step 176640, total loss is 14.88 , inference loss is 8.32, weight deacy loss is 6.57, training accuracy is 0.375000, time 218.937 samples/sec
+epoch 11, total_step 176660, total loss is 14.95 , inference loss is 8.38, weight deacy loss is 6.57, training accuracy is 0.343750, time 229.687 samples/sec
+epoch 11, total_step 176680, total loss is 14.65 , inference loss is 8.08, weight deacy loss is 6.57, training accuracy is 0.531250, time 222.832 samples/sec
+epoch 11, total_step 176700, total loss is 15.80 , inference loss is 9.23, weight deacy loss is 6.57, training accuracy is 0.468750, time 12.471 samples/sec
+epoch 11, total_step 176720, total loss is 12.68 , inference loss is 6.11, weight deacy loss is 6.57, training accuracy is 0.406250, time 227.451 samples/sec
+epoch 11, total_step 176740, total loss is 19.05 , inference loss is 12.48, weight deacy loss is 6.57, training accuracy is 0.281250, time 223.778 samples/sec
+epoch 11, total_step 176760, total loss is 13.99 , inference loss is 7.42, weight deacy loss is 6.57, training accuracy is 0.406250, time 230.940 samples/sec
+epoch 11, total_step 176780, total loss is 15.47 , inference loss is 8.91, weight deacy loss is 6.57, training accuracy is 0.375000, time 228.462 samples/sec
+epoch 11, total_step 176800, total loss is 13.33 , inference loss is 6.77, weight deacy loss is 6.57, training accuracy is 0.437500, time 300.561 samples/sec
+epoch 11, total_step 176820, total loss is 15.11 , inference loss is 8.54, weight deacy loss is 6.57, training accuracy is 0.437500, time 215.136 samples/sec
+epoch 11, total_step 176840, total loss is 14.44 , inference loss is 7.87, weight deacy loss is 6.57, training accuracy is 0.500000, time 215.410 samples/sec
+epoch 11, total_step 176860, total loss is 16.11 , inference loss is 9.54, weight deacy loss is 6.57, training accuracy is 0.406250, time 216.970 samples/sec
+epoch 11, total_step 176880, total loss is 14.59 , inference loss is 8.03, weight deacy loss is 6.56, training accuracy is 0.375000, time 214.420 samples/sec
+epoch 11, total_step 176900, total loss is 15.93 , inference loss is 9.36, weight deacy loss is 6.56, training accuracy is 0.406250, time 214.733 samples/sec
+epoch 11, total_step 176920, total loss is 15.94 , inference loss is 9.38, weight deacy loss is 6.56, training accuracy is 0.406250, time 215.476 samples/sec
+epoch 11, total_step 176940, total loss is 11.69 , inference loss is 5.13, weight deacy loss is 6.56, training accuracy is 0.468750, time 215.375 samples/sec
+epoch 11, total_step 176960, total loss is 14.53 , inference loss is 7.97, weight deacy loss is 6.56, training accuracy is 0.375000, time 299.480 samples/sec
+epoch 11, total_step 176980, total loss is 16.82 , inference loss is 10.26, weight deacy loss is 6.56, training accuracy is 0.375000, time 297.749 samples/sec
+epoch 11, total_step 177000, total loss is 14.46 , inference loss is 7.90, weight deacy loss is 6.56, training accuracy is 0.437500, time 12.561 samples/sec
+epoch 11, total_step 177020, total loss is 11.70 , inference loss is 5.14, weight deacy loss is 6.56, training accuracy is 0.406250, time 226.762 samples/sec
+epoch 11, total_step 177040, total loss is 18.10 , inference loss is 11.54, weight deacy loss is 6.56, training accuracy is 0.343750, time 225.798 samples/sec
+epoch 11, total_step 177060, total loss is 17.05 , inference loss is 10.48, weight deacy loss is 6.56, training accuracy is 0.312500, time 218.689 samples/sec
+epoch 11, total_step 177080, total loss is 15.52 , inference loss is 8.95, weight deacy loss is 6.56, training accuracy is 0.343750, time 217.903 samples/sec
+epoch 11, total_step 177100, total loss is 15.50 , inference loss is 8.94, weight deacy loss is 6.56, training accuracy is 0.437500, time 224.641 samples/sec
+epoch 11, total_step 177120, total loss is 15.09 , inference loss is 8.53, weight deacy loss is 6.56, training accuracy is 0.468750, time 220.121 samples/sec
+epoch 11, total_step 177140, total loss is 13.58 , inference loss is 7.02, weight deacy loss is 6.56, training accuracy is 0.437500, time 217.986 samples/sec
+epoch 11, total_step 177160, total loss is 13.00 , inference loss is 6.44, weight deacy loss is 6.56, training accuracy is 0.562500, time 214.327 samples/sec
+epoch 11, total_step 177180, total loss is 13.65 , inference loss is 7.09, weight deacy loss is 6.56, training accuracy is 0.343750, time 222.334 samples/sec
+epoch 11, total_step 177200, total loss is 14.58 , inference loss is 8.02, weight deacy loss is 6.56, training accuracy is 0.531250, time 207.109 samples/sec
+epoch 11, total_step 177220, total loss is 15.22 , inference loss is 8.65, weight deacy loss is 6.56, training accuracy is 0.406250, time 221.961 samples/sec
+epoch 11, total_step 177240, total loss is 10.25 , inference loss is 3.69, weight deacy loss is 6.56, training accuracy is 0.562500, time 298.588 samples/sec
+epoch 11, total_step 177260, total loss is 16.00 , inference loss is 9.44, weight deacy loss is 6.56, training accuracy is 0.468750, time 227.805 samples/sec
+epoch 11, total_step 177280, total loss is 13.24 , inference loss is 6.68, weight deacy loss is 6.56, training accuracy is 0.406250, time 235.194 samples/sec
+epoch 11, total_step 177300, total loss is 20.42 , inference loss is 13.86, weight deacy loss is 6.56, training accuracy is 0.218750, time 13.895 samples/sec
+epoch 11, total_step 177320, total loss is 11.81 , inference loss is 5.25, weight deacy loss is 6.56, training accuracy is 0.500000, time 219.131 samples/sec
+epoch 11, total_step 177340, total loss is 16.26 , inference loss is 9.70, weight deacy loss is 6.56, training accuracy is 0.375000, time 226.508 samples/sec
+epoch 11, total_step 177360, total loss is 9.73 , inference loss is 3.17, weight deacy loss is 6.56, training accuracy is 0.500000, time 225.477 samples/sec
+epoch 11, total_step 177380, total loss is 12.61 , inference loss is 6.05, weight deacy loss is 6.56, training accuracy is 0.375000, time 242.163 samples/sec
+epoch 11, total_step 177400, total loss is 16.59 , inference loss is 10.03, weight deacy loss is 6.56, training accuracy is 0.468750, time 219.819 samples/sec
+epoch 11, total_step 177420, total loss is 13.85 , inference loss is 7.29, weight deacy loss is 6.56, training accuracy is 0.343750, time 221.548 samples/sec
+epoch 11, total_step 177440, total loss is 11.16 , inference loss is 4.60, weight deacy loss is 6.56, training accuracy is 0.562500, time 239.361 samples/sec
+epoch 11, total_step 177460, total loss is 14.07 , inference loss is 7.51, weight deacy loss is 6.56, training accuracy is 0.562500, time 206.284 samples/sec
+epoch 11, total_step 177480, total loss is 13.30 , inference loss is 6.74, weight deacy loss is 6.56, training accuracy is 0.562500, time 215.828 samples/sec
+epoch 11, total_step 177500, total loss is 17.75 , inference loss is 11.19, weight deacy loss is 6.56, training accuracy is 0.312500, time 297.579 samples/sec
+epoch 11, total_step 177520, total loss is 13.31 , inference loss is 6.75, weight deacy loss is 6.56, training accuracy is 0.656250, time 206.591 samples/sec
+epoch 11, total_step 177540, total loss is 13.50 , inference loss is 6.94, weight deacy loss is 6.56, training accuracy is 0.437500, time 216.521 samples/sec
+epoch 11, total_step 177560, total loss is 14.14 , inference loss is 7.58, weight deacy loss is 6.56, training accuracy is 0.468750, time 212.927 samples/sec
+epoch 11, total_step 177580, total loss is 12.21 , inference loss is 5.65, weight deacy loss is 6.56, training accuracy is 0.468750, time 237.062 samples/sec
+epoch 11, total_step 177600, total loss is 14.19 , inference loss is 7.63, weight deacy loss is 6.56, training accuracy is 0.500000, time 12.530 samples/sec
+epoch 11, total_step 177620, total loss is 13.49 , inference loss is 6.93, weight deacy loss is 6.56, training accuracy is 0.468750, time 217.859 samples/sec
+epoch 11, total_step 177640, total loss is 11.69 , inference loss is 5.13, weight deacy loss is 6.56, training accuracy is 0.562500, time 221.396 samples/sec
+epoch 11, total_step 177660, total loss is 17.56 , inference loss is 11.00, weight deacy loss is 6.56, training accuracy is 0.468750, time 206.653 samples/sec
+epoch 11, total_step 177680, total loss is 12.41 , inference loss is 5.85, weight deacy loss is 6.56, training accuracy is 0.468750, time 224.910 samples/sec
+epoch 11, total_step 177700, total loss is 13.63 , inference loss is 7.07, weight deacy loss is 6.56, training accuracy is 0.437500, time 223.138 samples/sec
+epoch 11, total_step 177720, total loss is 16.13 , inference loss is 9.57, weight deacy loss is 6.56, training accuracy is 0.375000, time 227.228 samples/sec
+epoch 11, total_step 177740, total loss is 16.42 , inference loss is 9.87, weight deacy loss is 6.56, training accuracy is 0.406250, time 224.897 samples/sec
+epoch 11, total_step 177760, total loss is 13.53 , inference loss is 6.97, weight deacy loss is 6.56, training accuracy is 0.312500, time 301.831 samples/sec
+epoch 11, total_step 177780, total loss is 16.69 , inference loss is 10.13, weight deacy loss is 6.56, training accuracy is 0.468750, time 221.511 samples/sec
+epoch 11, total_step 177800, total loss is 16.34 , inference loss is 9.78, weight deacy loss is 6.56, training accuracy is 0.468750, time 220.016 samples/sec
+epoch 11, total_step 177820, total loss is 12.19 , inference loss is 5.64, weight deacy loss is 6.56, training accuracy is 0.625000, time 240.044 samples/sec
+epoch 11, total_step 177840, total loss is 13.47 , inference loss is 6.91, weight deacy loss is 6.56, training accuracy is 0.437500, time 227.392 samples/sec
+epoch 11, total_step 177860, total loss is 14.65 , inference loss is 8.09, weight deacy loss is 6.56, training accuracy is 0.437500, time 223.007 samples/sec
+epoch 11, total_step 177880, total loss is 13.40 , inference loss is 6.84, weight deacy loss is 6.56, training accuracy is 0.531250, time 225.065 samples/sec
+epoch 11, total_step 177900, total loss is 13.74 , inference loss is 7.18, weight deacy loss is 6.56, training accuracy is 0.562500, time 12.624 samples/sec
+epoch 11, total_step 177920, total loss is 11.61 , inference loss is 5.05, weight deacy loss is 6.56, training accuracy is 0.562500, time 217.270 samples/sec
+epoch 11, total_step 177940, total loss is 13.31 , inference loss is 6.75, weight deacy loss is 6.56, training accuracy is 0.468750, time 238.553 samples/sec
+epoch 11, total_step 177960, total loss is 13.15 , inference loss is 6.60, weight deacy loss is 6.56, training accuracy is 0.500000, time 226.112 samples/sec
+epoch 11, total_step 177980, total loss is 16.76 , inference loss is 10.20, weight deacy loss is 6.56, training accuracy is 0.437500, time 226.766 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.323046999999956
+best_threshold_index 116 0.9887037037037038
+best_threshold_index 116 0.9877777777777778
+best_threshold_index 114 0.9881481481481481
+best_threshold_index 116 0.9887037037037038
+best_threshold_index 116 0.9879629629629629
+best_threshold_index 118 0.9875925925925926
+best_threshold_index 114 0.9883333333333333
+best_threshold_index 114 0.9875925925925926
+best_threshold_index 116 0.9866666666666667
+best_threshold_index 116 0.9872222222222222
+[lfw][178000]XNorm: 22.369717
+[lfw][178000]Accuracy-Flip: 0.98683+-0.00560
+test accuracy is: 0.9868333333333335
+epoch 11, total_step 178000, total loss is 13.87 , inference loss is 7.31, weight deacy loss is 6.56, training accuracy is 0.468750, time 18.833 samples/sec
+epoch 11, total_step 178020, total loss is 11.00 , inference loss is 4.44, weight deacy loss is 6.56, training accuracy is 0.656250, time 197.560 samples/sec
+epoch 11, total_step 178040, total loss is 12.08 , inference loss is 5.52, weight deacy loss is 6.56, training accuracy is 0.625000, time 239.491 samples/sec
+epoch 11, total_step 178060, total loss is 13.86 , inference loss is 7.30, weight deacy loss is 6.56, training accuracy is 0.562500, time 224.403 samples/sec
+epoch 11, total_step 178080, total loss is 16.60 , inference loss is 10.04, weight deacy loss is 6.56, training accuracy is 0.406250, time 200.002 samples/sec
+epoch 11, total_step 178100, total loss is 11.24 , inference loss is 4.69, weight deacy loss is 6.56, training accuracy is 0.593750, time 212.982 samples/sec
+epoch 11, total_step 178120, total loss is 12.46 , inference loss is 5.90, weight deacy loss is 6.56, training accuracy is 0.437500, time 294.743 samples/sec
+epoch 11, total_step 178140, total loss is 13.26 , inference loss is 6.70, weight deacy loss is 6.56, training accuracy is 0.562500, time 235.425 samples/sec
+epoch 11, total_step 178160, total loss is 15.39 , inference loss is 8.84, weight deacy loss is 6.56, training accuracy is 0.500000, time 225.711 samples/sec
+epoch 11, total_step 178180, total loss is 13.44 , inference loss is 6.88, weight deacy loss is 6.56, training accuracy is 0.656250, time 223.697 samples/sec
+epoch 11, total_step 178200, total loss is 15.17 , inference loss is 8.61, weight deacy loss is 6.56, training accuracy is 0.406250, time 12.545 samples/sec
+epoch 11, total_step 178220, total loss is 16.89 , inference loss is 10.34, weight deacy loss is 6.56, training accuracy is 0.281250, time 220.130 samples/sec
+epoch 11, total_step 178240, total loss is 16.41 , inference loss is 9.86, weight deacy loss is 6.56, training accuracy is 0.375000, time 223.589 samples/sec
+epoch 11, total_step 178260, total loss is 13.68 , inference loss is 7.13, weight deacy loss is 6.56, training accuracy is 0.500000, time 224.684 samples/sec
+epoch 11, total_step 178280, total loss is 14.07 , inference loss is 7.51, weight deacy loss is 6.56, training accuracy is 0.562500, time 221.794 samples/sec
+epoch 11, total_step 178300, total loss is 12.28 , inference loss is 5.73, weight deacy loss is 6.56, training accuracy is 0.468750, time 213.169 samples/sec
+epoch 11, total_step 178320, total loss is 19.13 , inference loss is 12.57, weight deacy loss is 6.55, training accuracy is 0.468750, time 223.460 samples/sec
+epoch 11, total_step 178340, total loss is 14.68 , inference loss is 8.12, weight deacy loss is 6.55, training accuracy is 0.375000, time 232.582 samples/sec
+epoch 11, total_step 178360, total loss is 15.93 , inference loss is 9.38, weight deacy loss is 6.55, training accuracy is 0.500000, time 214.414 samples/sec
+epoch 11, total_step 178380, total loss is 13.25 , inference loss is 6.69, weight deacy loss is 6.55, training accuracy is 0.562500, time 299.210 samples/sec
+epoch 11, total_step 178400, total loss is 10.78 , inference loss is 4.23, weight deacy loss is 6.55, training accuracy is 0.437500, time 222.689 samples/sec
+epoch 11, total_step 178420, total loss is 15.27 , inference loss is 8.71, weight deacy loss is 6.55, training accuracy is 0.531250, time 221.792 samples/sec
+epoch 11, total_step 178440, total loss is 14.80 , inference loss is 8.25, weight deacy loss is 6.55, training accuracy is 0.500000, time 222.205 samples/sec
+epoch 11, total_step 178460, total loss is 15.03 , inference loss is 8.48, weight deacy loss is 6.55, training accuracy is 0.375000, time 225.216 samples/sec
+epoch 11, total_step 178480, total loss is 13.51 , inference loss is 6.96, weight deacy loss is 6.55, training accuracy is 0.500000, time 221.426 samples/sec
+epoch 11, total_step 178500, total loss is 13.66 , inference loss is 7.11, weight deacy loss is 6.55, training accuracy is 0.406250, time 12.465 samples/sec
+epoch 11, total_step 178520, total loss is 12.23 , inference loss is 5.68, weight deacy loss is 6.55, training accuracy is 0.468750, time 234.711 samples/sec
+epoch 11, total_step 178540, total loss is 12.50 , inference loss is 5.94, weight deacy loss is 6.55, training accuracy is 0.593750, time 211.800 samples/sec
+epoch 11, total_step 178560, total loss is 19.47 , inference loss is 12.91, weight deacy loss is 6.55, training accuracy is 0.218750, time 213.800 samples/sec
+epoch 11, total_step 178580, total loss is 14.01 , inference loss is 7.46, weight deacy loss is 6.55, training accuracy is 0.406250, time 211.743 samples/sec
+epoch 11, total_step 178600, total loss is 13.68 , inference loss is 7.12, weight deacy loss is 6.55, training accuracy is 0.437500, time 222.937 samples/sec
+epoch 11, total_step 178620, total loss is 12.76 , inference loss is 6.21, weight deacy loss is 6.55, training accuracy is 0.562500, time 212.032 samples/sec
+epoch 11, total_step 178640, total loss is 15.51 , inference loss is 8.96, weight deacy loss is 6.55, training accuracy is 0.406250, time 213.701 samples/sec
+epoch 11, total_step 178660, total loss is 14.34 , inference loss is 7.79, weight deacy loss is 6.55, training accuracy is 0.468750, time 230.785 samples/sec
+epoch 11, total_step 178680, total loss is 12.31 , inference loss is 5.76, weight deacy loss is 6.55, training accuracy is 0.593750, time 299.081 samples/sec
+epoch 11, total_step 178700, total loss is 14.27 , inference loss is 7.72, weight deacy loss is 6.55, training accuracy is 0.531250, time 294.968 samples/sec
+epoch 11, total_step 178720, total loss is 14.33 , inference loss is 7.78, weight deacy loss is 6.55, training accuracy is 0.531250, time 220.093 samples/sec
+epoch 11, total_step 178740, total loss is 12.00 , inference loss is 5.45, weight deacy loss is 6.55, training accuracy is 0.531250, time 221.478 samples/sec
+epoch 11, total_step 178760, total loss is 10.75 , inference loss is 4.20, weight deacy loss is 6.55, training accuracy is 0.625000, time 226.830 samples/sec
+epoch 11, total_step 178780, total loss is 14.15 , inference loss is 7.60, weight deacy loss is 6.55, training accuracy is 0.437500, time 220.425 samples/sec
+epoch 11, total_step 178800, total loss is 14.86 , inference loss is 8.31, weight deacy loss is 6.55, training accuracy is 0.468750, time 12.445 samples/sec
+epoch 11, total_step 178820, total loss is 14.85 , inference loss is 8.30, weight deacy loss is 6.55, training accuracy is 0.500000, time 220.549 samples/sec
+epoch 11, total_step 178840, total loss is 14.93 , inference loss is 8.38, weight deacy loss is 6.55, training accuracy is 0.437500, time 228.436 samples/sec
+epoch 11, total_step 178860, total loss is 15.17 , inference loss is 8.62, weight deacy loss is 6.55, training accuracy is 0.531250, time 218.209 samples/sec
+epoch 11, total_step 178880, total loss is 14.56 , inference loss is 8.01, weight deacy loss is 6.55, training accuracy is 0.500000, time 226.294 samples/sec
+epoch 11, total_step 178900, total loss is 15.74 , inference loss is 9.19, weight deacy loss is 6.55, training accuracy is 0.437500, time 226.378 samples/sec
+epoch 11, total_step 178920, total loss is 13.56 , inference loss is 7.01, weight deacy loss is 6.55, training accuracy is 0.437500, time 234.676 samples/sec
+epoch 11, total_step 178940, total loss is 15.32 , inference loss is 8.77, weight deacy loss is 6.55, training accuracy is 0.375000, time 215.565 samples/sec
+epoch 11, total_step 178960, total loss is 15.90 , inference loss is 9.35, weight deacy loss is 6.55, training accuracy is 0.375000, time 237.253 samples/sec
+epoch 11, total_step 178980, total loss is 16.17 , inference loss is 9.62, weight deacy loss is 6.55, training accuracy is 0.312500, time 212.934 samples/sec
+epoch 11, total_step 179000, total loss is 12.81 , inference loss is 6.26, weight deacy loss is 6.55, training accuracy is 0.593750, time 221.137 samples/sec
+epoch 11, total_step 179020, total loss is 10.76 , inference loss is 4.21, weight deacy loss is 6.55, training accuracy is 0.593750, time 237.396 samples/sec
+epoch 11, total_step 179040, total loss is 13.67 , inference loss is 7.12, weight deacy loss is 6.55, training accuracy is 0.312500, time 224.900 samples/sec
+epoch 11, total_step 179060, total loss is 14.05 , inference loss is 7.50, weight deacy loss is 6.55, training accuracy is 0.406250, time 217.501 samples/sec
+epoch 11, total_step 179080, total loss is 12.58 , inference loss is 6.03, weight deacy loss is 6.55, training accuracy is 0.406250, time 179.566 samples/sec
+epoch 11, total_step 179100, total loss is 12.97 , inference loss is 6.43, weight deacy loss is 6.55, training accuracy is 0.531250, time 12.420 samples/sec
+epoch 11, total_step 179120, total loss is 13.82 , inference loss is 7.28, weight deacy loss is 6.55, training accuracy is 0.468750, time 221.786 samples/sec
+epoch 11, total_step 179140, total loss is 14.61 , inference loss is 8.06, weight deacy loss is 6.55, training accuracy is 0.468750, time 219.141 samples/sec
+epoch 11, total_step 179160, total loss is 17.86 , inference loss is 11.31, weight deacy loss is 6.55, training accuracy is 0.343750, time 222.880 samples/sec
+epoch 11, total_step 179180, total loss is 11.94 , inference loss is 5.39, weight deacy loss is 6.55, training accuracy is 0.625000, time 222.612 samples/sec
+epoch 11, total_step 179200, total loss is 11.65 , inference loss is 5.10, weight deacy loss is 6.55, training accuracy is 0.468750, time 225.806 samples/sec
+epoch 11, total_step 179220, total loss is 18.94 , inference loss is 12.40, weight deacy loss is 6.55, training accuracy is 0.281250, time 224.240 samples/sec
+epoch 11, total_step 179240, total loss is 13.04 , inference loss is 6.49, weight deacy loss is 6.55, training accuracy is 0.500000, time 220.203 samples/sec
+epoch 11, total_step 179260, total loss is 14.49 , inference loss is 7.94, weight deacy loss is 6.55, training accuracy is 0.468750, time 227.342 samples/sec
+epoch 11, total_step 179280, total loss is 13.03 , inference loss is 6.48, weight deacy loss is 6.55, training accuracy is 0.500000, time 223.540 samples/sec
+epoch 11, total_step 179300, total loss is 13.45 , inference loss is 6.91, weight deacy loss is 6.55, training accuracy is 0.562500, time 245.297 samples/sec
+epoch 11, total_step 179320, total loss is 10.47 , inference loss is 3.92, weight deacy loss is 6.54, training accuracy is 0.656250, time 225.230 samples/sec
+epoch 11, total_step 179340, total loss is 12.17 , inference loss is 5.63, weight deacy loss is 6.54, training accuracy is 0.531250, time 187.813 samples/sec
+epoch 11, total_step 179360, total loss is 18.51 , inference loss is 11.96, weight deacy loss is 6.54, training accuracy is 0.343750, time 209.923 samples/sec
+epoch 11, total_step 179380, total loss is 14.98 , inference loss is 8.43, weight deacy loss is 6.54, training accuracy is 0.406250, time 220.844 samples/sec
+epoch 11, total_step 179400, total loss is 14.41 , inference loss is 7.87, weight deacy loss is 6.54, training accuracy is 0.437500, time 12.921 samples/sec
+epoch 11, total_step 179420, total loss is 9.89 , inference loss is 3.34, weight deacy loss is 6.54, training accuracy is 0.562500, time 224.306 samples/sec
+epoch 11, total_step 179440, total loss is 17.45 , inference loss is 10.91, weight deacy loss is 6.54, training accuracy is 0.406250, time 230.560 samples/sec
+epoch 11, total_step 179460, total loss is 11.26 , inference loss is 4.72, weight deacy loss is 6.54, training accuracy is 0.531250, time 238.307 samples/sec
+epoch 11, total_step 179480, total loss is 12.62 , inference loss is 6.07, weight deacy loss is 6.54, training accuracy is 0.468750, time 224.751 samples/sec
+epoch 11, total_step 179500, total loss is 13.54 , inference loss is 7.00, weight deacy loss is 6.54, training accuracy is 0.468750, time 219.038 samples/sec
+epoch 11, total_step 179520, total loss is 16.79 , inference loss is 10.24, weight deacy loss is 6.54, training accuracy is 0.500000, time 216.085 samples/sec
+epoch 11, total_step 179540, total loss is 12.44 , inference loss is 5.90, weight deacy loss is 6.54, training accuracy is 0.468750, time 215.094 samples/sec
+epoch 11, total_step 179560, total loss is 13.80 , inference loss is 7.26, weight deacy loss is 6.54, training accuracy is 0.437500, time 216.021 samples/sec
+epoch 11, total_step 179580, total loss is 13.07 , inference loss is 6.53, weight deacy loss is 6.54, training accuracy is 0.468750, time 217.744 samples/sec
+epoch 11, total_step 179600, total loss is 11.62 , inference loss is 5.08, weight deacy loss is 6.54, training accuracy is 0.531250, time 213.011 samples/sec
+epoch 11, total_step 179620, total loss is 11.98 , inference loss is 5.44, weight deacy loss is 6.54, training accuracy is 0.562500, time 299.276 samples/sec
+epoch 11, total_step 179640, total loss is 12.75 , inference loss is 6.21, weight deacy loss is 6.54, training accuracy is 0.437500, time 236.871 samples/sec
+epoch 11, total_step 179660, total loss is 15.72 , inference loss is 9.18, weight deacy loss is 6.54, training accuracy is 0.468750, time 222.263 samples/sec
+epoch 11, total_step 179680, total loss is 10.31 , inference loss is 3.77, weight deacy loss is 6.54, training accuracy is 0.593750, time 234.718 samples/sec
+epoch 11, total_step 179700, total loss is 13.00 , inference loss is 6.46, weight deacy loss is 6.54, training accuracy is 0.500000, time 12.443 samples/sec
+epoch 11, total_step 179720, total loss is 15.29 , inference loss is 8.75, weight deacy loss is 6.54, training accuracy is 0.437500, time 226.226 samples/sec
+epoch 11, total_step 179740, total loss is 15.55 , inference loss is 9.01, weight deacy loss is 6.54, training accuracy is 0.468750, time 226.419 samples/sec
+epoch 11, total_step 179760, total loss is 13.90 , inference loss is 7.35, weight deacy loss is 6.54, training accuracy is 0.250000, time 225.887 samples/sec
+epoch 11, total_step 179780, total loss is 13.08 , inference loss is 6.54, weight deacy loss is 6.54, training accuracy is 0.437500, time 226.257 samples/sec
+epoch 11, total_step 179800, total loss is 11.15 , inference loss is 4.61, weight deacy loss is 6.54, training accuracy is 0.531250, time 224.411 samples/sec
+epoch 11, total_step 179820, total loss is 13.37 , inference loss is 6.83, weight deacy loss is 6.54, training accuracy is 0.343750, time 203.873 samples/sec
+epoch 11, total_step 179840, total loss is 18.00 , inference loss is 11.46, weight deacy loss is 6.54, training accuracy is 0.375000, time 211.447 samples/sec
+epoch 11, total_step 179860, total loss is 15.20 , inference loss is 8.65, weight deacy loss is 6.54, training accuracy is 0.281250, time 209.165 samples/sec
+epoch 11, total_step 179880, total loss is 15.32 , inference loss is 8.78, weight deacy loss is 6.54, training accuracy is 0.406250, time 188.635 samples/sec
+epoch 11, total_step 179900, total loss is 13.06 , inference loss is 6.52, weight deacy loss is 6.54, training accuracy is 0.500000, time 225.510 samples/sec
+epoch 11, total_step 179920, total loss is 13.93 , inference loss is 7.39, weight deacy loss is 6.54, training accuracy is 0.312500, time 229.683 samples/sec
+epoch 11, total_step 179940, total loss is 12.33 , inference loss is 5.79, weight deacy loss is 6.54, training accuracy is 0.375000, time 224.249 samples/sec
+epoch 11, total_step 179960, total loss is 14.28 , inference loss is 7.74, weight deacy loss is 6.54, training accuracy is 0.468750, time 224.688 samples/sec
+epoch 11, total_step 179980, total loss is 11.45 , inference loss is 4.91, weight deacy loss is 6.54, training accuracy is 0.406250, time 222.443 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.669421000000003
+best_threshold_index 112 0.9848148148148148
+best_threshold_index 112 0.9835185185185186
+best_threshold_index 110 0.9835185185185186
+best_threshold_index 109 0.9851851851851852
+best_threshold_index 112 0.9846296296296296
+best_threshold_index 112 0.9831481481481481
+best_threshold_index 112 0.9844444444444445
+best_threshold_index 105 0.9833333333333333
+best_threshold_index 110 0.9833333333333333
+best_threshold_index 111 0.9829629629629629
+[lfw][180000]XNorm: 23.515961
+[lfw][180000]Accuracy-Flip: 0.98200+-0.00678
+test accuracy is: 0.982
+epoch 11, total_step 180000, total loss is 14.91 , inference loss is 8.37, weight deacy loss is 6.54, training accuracy is 0.406250, time 15.466 samples/sec
+epoch 11, total_step 180020, total loss is 13.38 , inference loss is 6.84, weight deacy loss is 6.54, training accuracy is 0.437500, time 227.632 samples/sec
+epoch 11, total_step 180040, total loss is 16.04 , inference loss is 9.50, weight deacy loss is 6.54, training accuracy is 0.468750, time 206.371 samples/sec
+epoch 11, total_step 180060, total loss is 14.31 , inference loss is 7.77, weight deacy loss is 6.54, training accuracy is 0.406250, time 221.497 samples/sec
+epoch 11, total_step 180080, total loss is 10.98 , inference loss is 4.44, weight deacy loss is 6.54, training accuracy is 0.531250, time 229.832 samples/sec
+epoch 11, total_step 180100, total loss is 17.43 , inference loss is 10.90, weight deacy loss is 6.54, training accuracy is 0.312500, time 227.871 samples/sec
+epoch 11, total_step 180120, total loss is 11.93 , inference loss is 5.39, weight deacy loss is 6.54, training accuracy is 0.468750, time 206.114 samples/sec
+epoch 11, total_step 180140, total loss is 11.88 , inference loss is 5.34, weight deacy loss is 6.54, training accuracy is 0.468750, time 214.972 samples/sec
+epoch 11, total_step 180160, total loss is 15.45 , inference loss is 8.91, weight deacy loss is 6.54, training accuracy is 0.187500, time 225.603 samples/sec
+epoch 11, total_step 180180, total loss is 10.52 , inference loss is 3.99, weight deacy loss is 6.54, training accuracy is 0.593750, time 237.218 samples/sec
+epoch 11, total_step 180200, total loss is 13.86 , inference loss is 7.32, weight deacy loss is 6.54, training accuracy is 0.500000, time 212.216 samples/sec
+epoch 11, total_step 180220, total loss is 14.74 , inference loss is 8.20, weight deacy loss is 6.54, training accuracy is 0.343750, time 223.850 samples/sec
+epoch 11, total_step 180240, total loss is 16.59 , inference loss is 10.06, weight deacy loss is 6.54, training accuracy is 0.375000, time 227.096 samples/sec
+epoch 11, total_step 180260, total loss is 15.32 , inference loss is 8.78, weight deacy loss is 6.54, training accuracy is 0.406250, time 224.049 samples/sec
+epoch 11, total_step 180280, total loss is 21.01 , inference loss is 14.47, weight deacy loss is 6.54, training accuracy is 0.343750, time 225.555 samples/sec
+epoch 11, total_step 180300, total loss is 19.09 , inference loss is 12.56, weight deacy loss is 6.54, training accuracy is 0.312500, time 13.144 samples/sec
+epoch 11, total_step 180320, total loss is 15.85 , inference loss is 9.32, weight deacy loss is 6.54, training accuracy is 0.406250, time 217.880 samples/sec
+epoch 11, total_step 180340, total loss is 15.79 , inference loss is 9.26, weight deacy loss is 6.54, training accuracy is 0.406250, time 220.527 samples/sec
+epoch 11, total_step 180360, total loss is 12.84 , inference loss is 6.31, weight deacy loss is 6.54, training accuracy is 0.468750, time 212.019 samples/sec
+epoch 11, total_step 180380, total loss is 16.14 , inference loss is 9.60, weight deacy loss is 6.54, training accuracy is 0.343750, time 223.788 samples/sec
+epoch 11, total_step 180400, total loss is 16.12 , inference loss is 9.58, weight deacy loss is 6.54, training accuracy is 0.406250, time 300.141 samples/sec
+epoch 11, total_step 180420, total loss is 11.81 , inference loss is 5.27, weight deacy loss is 6.54, training accuracy is 0.437500, time 298.659 samples/sec
+epoch 11, total_step 180440, total loss is 11.98 , inference loss is 5.45, weight deacy loss is 6.54, training accuracy is 0.468750, time 222.069 samples/sec
+epoch 11, total_step 180460, total loss is 15.46 , inference loss is 8.93, weight deacy loss is 6.54, training accuracy is 0.406250, time 236.233 samples/sec
+epoch 11, total_step 180480, total loss is 13.11 , inference loss is 6.58, weight deacy loss is 6.54, training accuracy is 0.593750, time 215.521 samples/sec
+epoch 11, total_step 180500, total loss is 15.03 , inference loss is 8.50, weight deacy loss is 6.53, training accuracy is 0.312500, time 213.393 samples/sec
+epoch 11, total_step 180520, total loss is 15.12 , inference loss is 8.58, weight deacy loss is 6.53, training accuracy is 0.218750, time 216.063 samples/sec
+epoch 11, total_step 180540, total loss is 13.56 , inference loss is 7.02, weight deacy loss is 6.53, training accuracy is 0.625000, time 222.027 samples/sec
+epoch 11, total_step 180560, total loss is 16.34 , inference loss is 9.80, weight deacy loss is 6.53, training accuracy is 0.375000, time 217.432 samples/sec
+epoch 11, total_step 180580, total loss is 13.47 , inference loss is 6.94, weight deacy loss is 6.53, training accuracy is 0.343750, time 234.780 samples/sec
+epoch 11, total_step 180600, total loss is 15.13 , inference loss is 8.59, weight deacy loss is 6.53, training accuracy is 0.343750, time 12.946 samples/sec
+epoch 11, total_step 180620, total loss is 15.16 , inference loss is 8.63, weight deacy loss is 6.53, training accuracy is 0.437500, time 221.374 samples/sec
+epoch 11, total_step 180640, total loss is 12.90 , inference loss is 6.37, weight deacy loss is 6.53, training accuracy is 0.343750, time 217.746 samples/sec
+epoch 11, total_step 180660, total loss is 18.47 , inference loss is 11.93, weight deacy loss is 6.53, training accuracy is 0.250000, time 225.873 samples/sec
+epoch 11, total_step 180680, total loss is 16.74 , inference loss is 10.20, weight deacy loss is 6.53, training accuracy is 0.250000, time 218.867 samples/sec
+epoch 11, total_step 180700, total loss is 13.72 , inference loss is 7.19, weight deacy loss is 6.53, training accuracy is 0.437500, time 237.995 samples/sec
+epoch 11, total_step 180720, total loss is 18.28 , inference loss is 11.74, weight deacy loss is 6.53, training accuracy is 0.281250, time 223.529 samples/sec
+epoch 11, total_step 180740, total loss is 14.38 , inference loss is 7.84, weight deacy loss is 6.53, training accuracy is 0.281250, time 226.225 samples/sec
+epoch 11, total_step 180760, total loss is 11.80 , inference loss is 5.27, weight deacy loss is 6.53, training accuracy is 0.406250, time 220.565 samples/sec
+epoch 11, total_step 180780, total loss is 14.04 , inference loss is 7.51, weight deacy loss is 6.53, training accuracy is 0.437500, time 220.858 samples/sec
+epoch 11, total_step 180800, total loss is 15.06 , inference loss is 8.52, weight deacy loss is 6.53, training accuracy is 0.343750, time 224.012 samples/sec
+epoch 11, total_step 180820, total loss is 15.80 , inference loss is 9.27, weight deacy loss is 6.53, training accuracy is 0.312500, time 228.721 samples/sec
+epoch 11, total_step 180840, total loss is 16.37 , inference loss is 9.84, weight deacy loss is 6.53, training accuracy is 0.250000, time 221.923 samples/sec
+epoch 11, total_step 180860, total loss is 19.12 , inference loss is 12.58, weight deacy loss is 6.53, training accuracy is 0.406250, time 225.410 samples/sec
+epoch 11, total_step 180880, total loss is 11.14 , inference loss is 4.61, weight deacy loss is 6.53, training accuracy is 0.562500, time 223.112 samples/sec
+epoch 11, total_step 180900, total loss is 17.41 , inference loss is 10.88, weight deacy loss is 6.53, training accuracy is 0.375000, time 13.166 samples/sec
+epoch 11, total_step 180920, total loss is 15.69 , inference loss is 9.16, weight deacy loss is 6.53, training accuracy is 0.281250, time 228.828 samples/sec
+epoch 11, total_step 180940, total loss is 13.17 , inference loss is 6.64, weight deacy loss is 6.53, training accuracy is 0.531250, time 227.505 samples/sec
+epoch 11, total_step 180960, total loss is 20.95 , inference loss is 14.42, weight deacy loss is 6.53, training accuracy is 0.187500, time 220.128 samples/sec
+epoch 11, total_step 180980, total loss is 17.31 , inference loss is 10.78, weight deacy loss is 6.53, training accuracy is 0.187500, time 209.101 samples/sec
+epoch 11, total_step 181000, total loss is 14.36 , inference loss is 7.83, weight deacy loss is 6.53, training accuracy is 0.343750, time 237.460 samples/sec
+epoch 11, total_step 181020, total loss is 11.44 , inference loss is 4.91, weight deacy loss is 6.53, training accuracy is 0.468750, time 225.711 samples/sec
+epoch 11, total_step 181040, total loss is 14.80 , inference loss is 8.27, weight deacy loss is 6.53, training accuracy is 0.281250, time 226.126 samples/sec
+epoch 11, total_step 181060, total loss is 12.03 , inference loss is 5.50, weight deacy loss is 6.53, training accuracy is 0.562500, time 222.615 samples/sec
+epoch 11, total_step 181080, total loss is 16.19 , inference loss is 9.66, weight deacy loss is 6.53, training accuracy is 0.468750, time 225.904 samples/sec
+epoch 11, total_step 181100, total loss is 13.86 , inference loss is 7.33, weight deacy loss is 6.53, training accuracy is 0.437500, time 220.746 samples/sec
+epoch 11, total_step 181120, total loss is 11.27 , inference loss is 4.74, weight deacy loss is 6.53, training accuracy is 0.468750, time 180.092 samples/sec
+epoch 11, total_step 181140, total loss is 14.80 , inference loss is 8.27, weight deacy loss is 6.53, training accuracy is 0.406250, time 234.615 samples/sec
+epoch 11, total_step 181160, total loss is 16.12 , inference loss is 9.59, weight deacy loss is 6.53, training accuracy is 0.343750, time 297.076 samples/sec
+epoch 11, total_step 181180, total loss is 12.06 , inference loss is 5.53, weight deacy loss is 6.53, training accuracy is 0.468750, time 232.950 samples/sec
+epoch 11, total_step 181200, total loss is 12.60 , inference loss is 6.07, weight deacy loss is 6.53, training accuracy is 0.437500, time 14.010 samples/sec
+epoch 11, total_step 181220, total loss is 14.13 , inference loss is 7.60, weight deacy loss is 6.53, training accuracy is 0.343750, time 221.208 samples/sec
+epoch 11, total_step 181240, total loss is 11.60 , inference loss is 5.07, weight deacy loss is 6.53, training accuracy is 0.406250, time 200.589 samples/sec
+epoch 11, total_step 181260, total loss is 12.56 , inference loss is 6.04, weight deacy loss is 6.53, training accuracy is 0.375000, time 216.778 samples/sec
+epoch 11, total_step 181280, total loss is 16.00 , inference loss is 9.47, weight deacy loss is 6.53, training accuracy is 0.375000, time 224.354 samples/sec
+epoch 11, total_step 181300, total loss is 15.60 , inference loss is 9.07, weight deacy loss is 6.53, training accuracy is 0.281250, time 217.542 samples/sec
+epoch 11, total_step 181320, total loss is 16.54 , inference loss is 10.01, weight deacy loss is 6.53, training accuracy is 0.375000, time 226.779 samples/sec
+epoch 11, total_step 181340, total loss is 13.11 , inference loss is 6.59, weight deacy loss is 6.53, training accuracy is 0.406250, time 226.278 samples/sec
+epoch 11, total_step 181360, total loss is 13.35 , inference loss is 6.83, weight deacy loss is 6.53, training accuracy is 0.343750, time 222.281 samples/sec
+epoch 11, total_step 181380, total loss is 13.62 , inference loss is 7.09, weight deacy loss is 6.53, training accuracy is 0.406250, time 219.545 samples/sec
+epoch 11, total_step 181400, total loss is 19.64 , inference loss is 13.12, weight deacy loss is 6.53, training accuracy is 0.250000, time 222.339 samples/sec
+epoch 11, total_step 181420, total loss is 16.17 , inference loss is 9.64, weight deacy loss is 6.53, training accuracy is 0.406250, time 298.782 samples/sec
+epoch 11, total_step 181440, total loss is 13.64 , inference loss is 7.11, weight deacy loss is 6.53, training accuracy is 0.406250, time 230.855 samples/sec
+epoch 11, total_step 181460, total loss is 18.93 , inference loss is 12.41, weight deacy loss is 6.53, training accuracy is 0.218750, time 225.128 samples/sec
+epoch 11, total_step 181480, total loss is 15.43 , inference loss is 8.90, weight deacy loss is 6.53, training accuracy is 0.312500, time 221.583 samples/sec
+epoch 11, total_step 181500, total loss is 14.53 , inference loss is 8.00, weight deacy loss is 6.53, training accuracy is 0.281250, time 12.727 samples/sec
+epoch 11, total_step 181520, total loss is 15.33 , inference loss is 8.81, weight deacy loss is 6.53, training accuracy is 0.468750, time 227.052 samples/sec
+epoch 11, total_step 181540, total loss is 16.97 , inference loss is 10.44, weight deacy loss is 6.53, training accuracy is 0.281250, time 232.523 samples/sec
+epoch 11, total_step 181560, total loss is 18.88 , inference loss is 12.35, weight deacy loss is 6.52, training accuracy is 0.406250, time 230.733 samples/sec
+epoch 11, total_step 181580, total loss is 12.65 , inference loss is 6.12, weight deacy loss is 6.52, training accuracy is 0.375000, time 226.392 samples/sec
+epoch 11, total_step 181600, total loss is 14.02 , inference loss is 7.49, weight deacy loss is 6.52, training accuracy is 0.343750, time 233.755 samples/sec
+epoch 11, total_step 181620, total loss is 14.98 , inference loss is 8.46, weight deacy loss is 6.52, training accuracy is 0.250000, time 226.073 samples/sec
+epoch 11, total_step 181640, total loss is 17.57 , inference loss is 11.05, weight deacy loss is 6.52, training accuracy is 0.343750, time 207.193 samples/sec
+epoch 11, total_step 181660, total loss is 13.91 , inference loss is 7.39, weight deacy loss is 6.52, training accuracy is 0.406250, time 176.531 samples/sec
+epoch 11, total_step 181680, total loss is 16.46 , inference loss is 9.94, weight deacy loss is 6.52, training accuracy is 0.281250, time 297.462 samples/sec
+epoch 11, total_step 181700, total loss is 15.40 , inference loss is 8.88, weight deacy loss is 6.52, training accuracy is 0.437500, time 216.562 samples/sec
+epoch 11, total_step 181720, total loss is 13.13 , inference loss is 6.60, weight deacy loss is 6.52, training accuracy is 0.343750, time 219.506 samples/sec
+epoch 11, total_step 181740, total loss is 14.16 , inference loss is 7.63, weight deacy loss is 6.52, training accuracy is 0.343750, time 225.402 samples/sec
+epoch 11, total_step 181760, total loss is 14.87 , inference loss is 8.35, weight deacy loss is 6.52, training accuracy is 0.343750, time 227.787 samples/sec
+epoch 11, total_step 181780, total loss is 17.04 , inference loss is 10.51, weight deacy loss is 6.52, training accuracy is 0.250000, time 218.736 samples/sec
+epoch 11, total_step 181800, total loss is 18.23 , inference loss is 11.71, weight deacy loss is 6.52, training accuracy is 0.250000, time 13.191 samples/sec
+epoch 11, total_step 181820, total loss is 14.19 , inference loss is 7.67, weight deacy loss is 6.52, training accuracy is 0.343750, time 221.125 samples/sec
+epoch 11, total_step 181840, total loss is 10.82 , inference loss is 4.30, weight deacy loss is 6.52, training accuracy is 0.500000, time 225.426 samples/sec
+epoch 11, total_step 181860, total loss is 10.83 , inference loss is 4.31, weight deacy loss is 6.52, training accuracy is 0.468750, time 176.526 samples/sec
+epoch 11, total_step 181880, total loss is 16.28 , inference loss is 9.76, weight deacy loss is 6.52, training accuracy is 0.406250, time 217.449 samples/sec
+epoch 11, total_step 181900, total loss is 11.92 , inference loss is 5.40, weight deacy loss is 6.52, training accuracy is 0.562500, time 220.654 samples/sec
+epoch 11, total_step 181920, total loss is 17.23 , inference loss is 10.71, weight deacy loss is 6.52, training accuracy is 0.281250, time 222.751 samples/sec
+epoch 11, total_step 181940, total loss is 10.99 , inference loss is 4.47, weight deacy loss is 6.52, training accuracy is 0.375000, time 296.932 samples/sec
+epoch 11, total_step 181960, total loss is 15.83 , inference loss is 9.31, weight deacy loss is 6.52, training accuracy is 0.437500, time 246.535 samples/sec
+epoch 11, total_step 181980, total loss is 13.48 , inference loss is 6.96, weight deacy loss is 6.52, training accuracy is 0.343750, time 234.616 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.112255
+best_threshold_index 115 0.9857407407407407
+best_threshold_index 115 0.9827777777777778
+best_threshold_index 113 0.9838888888888889
+best_threshold_index 113 0.9846296296296296
+best_threshold_index 115 0.9846296296296296
+best_threshold_index 115 0.9837037037037037
+best_threshold_index 115 0.9842592592592593
+best_threshold_index 113 0.9831481481481481
+best_threshold_index 115 0.9825925925925926
+best_threshold_index 115 0.9831481481481481
+[lfw][182000]XNorm: 21.552021
+[lfw][182000]Accuracy-Flip: 0.98317+-0.00845
+test accuracy is: 0.9831666666666665
+epoch 11, total_step 182000, total loss is 18.06 , inference loss is 11.54, weight deacy loss is 6.52, training accuracy is 0.343750, time 18.205 samples/sec
+epoch 11, total_step 182020, total loss is 14.24 , inference loss is 7.72, weight deacy loss is 6.52, training accuracy is 0.375000, time 236.824 samples/sec
+epoch 11, total_step 182040, total loss is 12.95 , inference loss is 6.43, weight deacy loss is 6.52, training accuracy is 0.531250, time 242.691 samples/sec
+epoch 11, total_step 182060, total loss is 18.92 , inference loss is 12.40, weight deacy loss is 6.52, training accuracy is 0.156250, time 297.065 samples/sec
+epoch 11, total_step 182080, total loss is 17.86 , inference loss is 11.34, weight deacy loss is 6.52, training accuracy is 0.343750, time 296.252 samples/sec
+epoch 11, total_step 182100, total loss is 15.33 , inference loss is 8.81, weight deacy loss is 6.52, training accuracy is 0.281250, time 14.293 samples/sec
+epoch 11, total_step 182120, total loss is 15.16 , inference loss is 8.64, weight deacy loss is 6.52, training accuracy is 0.312500, time 215.715 samples/sec
+epoch 11, total_step 182140, total loss is 14.14 , inference loss is 7.63, weight deacy loss is 6.52, training accuracy is 0.406250, time 215.424 samples/sec
+epoch 11, total_step 182160, total loss is 12.40 , inference loss is 5.88, weight deacy loss is 6.52, training accuracy is 0.437500, time 223.294 samples/sec
+epoch 11, total_step 182180, total loss is 13.02 , inference loss is 6.51, weight deacy loss is 6.52, training accuracy is 0.437500, time 299.405 samples/sec
+epoch 11, total_step 182200, total loss is 12.54 , inference loss is 6.02, weight deacy loss is 6.52, training accuracy is 0.281250, time 299.596 samples/sec
+epoch 11, total_step 182220, total loss is 17.53 , inference loss is 11.01, weight deacy loss is 6.52, training accuracy is 0.312500, time 204.029 samples/sec
+epoch 11, total_step 182240, total loss is 11.35 , inference loss is 4.83, weight deacy loss is 6.52, training accuracy is 0.406250, time 227.581 samples/sec
+epoch 11, total_step 182260, total loss is 14.96 , inference loss is 8.44, weight deacy loss is 6.52, training accuracy is 0.406250, time 221.669 samples/sec
+epoch 11, total_step 182280, total loss is 12.58 , inference loss is 6.06, weight deacy loss is 6.52, training accuracy is 0.437500, time 232.487 samples/sec
+epoch 11, total_step 182300, total loss is 17.30 , inference loss is 10.79, weight deacy loss is 6.52, training accuracy is 0.343750, time 217.998 samples/sec
+epoch 11, total_step 182320, total loss is 14.30 , inference loss is 7.79, weight deacy loss is 6.52, training accuracy is 0.593750, time 225.903 samples/sec
+epoch 11, total_step 182340, total loss is 14.91 , inference loss is 8.40, weight deacy loss is 6.52, training accuracy is 0.375000, time 226.801 samples/sec
+epoch 11, total_step 182360, total loss is 12.57 , inference loss is 6.05, weight deacy loss is 6.52, training accuracy is 0.375000, time 227.156 samples/sec
+epoch 11, total_step 182380, total loss is 16.96 , inference loss is 10.44, weight deacy loss is 6.52, training accuracy is 0.437500, time 219.041 samples/sec
+epoch 11, total_step 182400, total loss is 15.54 , inference loss is 9.03, weight deacy loss is 6.52, training accuracy is 0.312500, time 12.740 samples/sec
+epoch 11, total_step 182420, total loss is 15.72 , inference loss is 9.21, weight deacy loss is 6.52, training accuracy is 0.312500, time 220.147 samples/sec
+epoch 11, total_step 182440, total loss is 13.40 , inference loss is 6.88, weight deacy loss is 6.52, training accuracy is 0.500000, time 221.961 samples/sec
+epoch 11, total_step 182460, total loss is 13.51 , inference loss is 6.99, weight deacy loss is 6.52, training accuracy is 0.375000, time 224.481 samples/sec
+epoch 11, total_step 182480, total loss is 13.70 , inference loss is 7.19, weight deacy loss is 6.52, training accuracy is 0.406250, time 184.603 samples/sec
+epoch 11, total_step 182500, total loss is 16.48 , inference loss is 9.97, weight deacy loss is 6.51, training accuracy is 0.281250, time 227.817 samples/sec
+epoch 11, total_step 182520, total loss is 13.82 , inference loss is 7.31, weight deacy loss is 6.51, training accuracy is 0.531250, time 199.357 samples/sec
+epoch 11, total_step 182540, total loss is 18.14 , inference loss is 11.63, weight deacy loss is 6.51, training accuracy is 0.281250, time 218.007 samples/sec
+epoch 11, total_step 182560, total loss is 14.25 , inference loss is 7.73, weight deacy loss is 6.51, training accuracy is 0.406250, time 222.902 samples/sec
+epoch 11, total_step 182580, total loss is 17.41 , inference loss is 10.90, weight deacy loss is 6.51, training accuracy is 0.312500, time 227.563 samples/sec
+epoch 11, total_step 182600, total loss is 13.11 , inference loss is 6.60, weight deacy loss is 6.51, training accuracy is 0.500000, time 211.631 samples/sec
+epoch 11, total_step 182620, total loss is 15.55 , inference loss is 9.04, weight deacy loss is 6.51, training accuracy is 0.218750, time 224.621 samples/sec
+epoch 11, total_step 182640, total loss is 17.33 , inference loss is 10.81, weight deacy loss is 6.51, training accuracy is 0.156250, time 237.516 samples/sec
+epoch 11, total_step 182660, total loss is 15.93 , inference loss is 9.41, weight deacy loss is 6.51, training accuracy is 0.312500, time 224.705 samples/sec
+epoch 11, total_step 182680, total loss is 15.88 , inference loss is 9.37, weight deacy loss is 6.51, training accuracy is 0.375000, time 226.840 samples/sec
+epoch 11, total_step 182700, total loss is 13.44 , inference loss is 6.93, weight deacy loss is 6.51, training accuracy is 0.500000, time 12.750 samples/sec
+epoch 11, total_step 182720, total loss is 16.18 , inference loss is 9.66, weight deacy loss is 6.51, training accuracy is 0.281250, time 205.453 samples/sec
+epoch 11, total_step 182740, total loss is 15.53 , inference loss is 9.01, weight deacy loss is 6.51, training accuracy is 0.437500, time 206.769 samples/sec
+epoch 11, total_step 182760, total loss is 16.57 , inference loss is 10.06, weight deacy loss is 6.51, training accuracy is 0.312500, time 219.968 samples/sec
+epoch 11, total_step 182780, total loss is 16.38 , inference loss is 9.86, weight deacy loss is 6.51, training accuracy is 0.312500, time 224.462 samples/sec
+epoch 11, total_step 182800, total loss is 12.29 , inference loss is 5.78, weight deacy loss is 6.51, training accuracy is 0.437500, time 225.022 samples/sec
+epoch 11, total_step 182820, total loss is 15.37 , inference loss is 8.86, weight deacy loss is 6.51, training accuracy is 0.375000, time 225.759 samples/sec
+epoch 11, total_step 182840, total loss is 15.13 , inference loss is 8.62, weight deacy loss is 6.51, training accuracy is 0.250000, time 217.208 samples/sec
+epoch 11, total_step 182860, total loss is 17.42 , inference loss is 10.91, weight deacy loss is 6.51, training accuracy is 0.437500, time 224.697 samples/sec
+epoch 11, total_step 182880, total loss is 21.52 , inference loss is 15.01, weight deacy loss is 6.51, training accuracy is 0.312500, time 236.525 samples/sec
+epoch 11, total_step 182900, total loss is 16.14 , inference loss is 9.63, weight deacy loss is 6.51, training accuracy is 0.406250, time 219.521 samples/sec
+epoch 11, total_step 182920, total loss is 14.82 , inference loss is 8.31, weight deacy loss is 6.51, training accuracy is 0.343750, time 223.430 samples/sec
+epoch 11, total_step 182940, total loss is 11.88 , inference loss is 5.37, weight deacy loss is 6.51, training accuracy is 0.437500, time 223.928 samples/sec
+epoch 11, total_step 182960, total loss is 18.51 , inference loss is 12.00, weight deacy loss is 6.51, training accuracy is 0.218750, time 216.438 samples/sec
+epoch 11, total_step 182980, total loss is 12.33 , inference loss is 5.82, weight deacy loss is 6.51, training accuracy is 0.312500, time 223.956 samples/sec
+epoch 11, total_step 183000, total loss is 13.04 , inference loss is 6.53, weight deacy loss is 6.51, training accuracy is 0.375000, time 13.239 samples/sec
+epoch 11, total_step 183020, total loss is 13.50 , inference loss is 6.99, weight deacy loss is 6.51, training accuracy is 0.312500, time 226.638 samples/sec
+epoch 11, total_step 183040, total loss is 14.04 , inference loss is 7.53, weight deacy loss is 6.51, training accuracy is 0.593750, time 224.859 samples/sec
+epoch 11, total_step 183060, total loss is 16.86 , inference loss is 10.35, weight deacy loss is 6.51, training accuracy is 0.375000, time 242.065 samples/sec
+epoch 11, total_step 183080, total loss is 11.86 , inference loss is 5.35, weight deacy loss is 6.51, training accuracy is 0.375000, time 222.469 samples/sec2022-12-01 12:30:33.409700: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7416 of 10000
+2022-12-01 12:30:36.875612: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 11, total_step 183100, total loss is 16.46 , inference loss is 9.95, weight deacy loss is 6.51, training accuracy is 0.437500, time 216.934 samples/sec
+epoch 11, total_step 183120, total loss is 13.78 , inference loss is 7.27, weight deacy loss is 6.51, training accuracy is 0.406250, time 218.228 samples/sec
+epoch 11, total_step 183140, total loss is 14.60 , inference loss is 8.09, weight deacy loss is 6.51, training accuracy is 0.343750, time 222.241 samples/sec
+epoch 11, total_step 183160, total loss is 16.55 , inference loss is 10.04, weight deacy loss is 6.51, training accuracy is 0.437500, time 224.116 samples/sec
+epoch 11, total_step 183180, total loss is 18.63 , inference loss is 12.12, weight deacy loss is 6.51, training accuracy is 0.312500, time 226.758 samples/sec
+epoch 11, total_step 183200, total loss is 12.96 , inference loss is 6.45, weight deacy loss is 6.51, training accuracy is 0.406250, time 213.769 samples/sec
+epoch 11, total_step 183220, total loss is 18.54 , inference loss is 12.03, weight deacy loss is 6.51, training accuracy is 0.281250, time 216.231 samples/sec
+epoch 11, total_step 183240, total loss is 15.64 , inference loss is 9.14, weight deacy loss is 6.51, training accuracy is 0.437500, time 218.073 samples/sec
+epoch 11, total_step 183260, total loss is 19.45 , inference loss is 12.94, weight deacy loss is 6.51, training accuracy is 0.218750, time 230.436 samples/sec
+epoch 11, total_step 183280, total loss is 15.21 , inference loss is 8.70, weight deacy loss is 6.51, training accuracy is 0.281250, time 230.871 samples/sec
+epoch 11, total_step 183300, total loss is 11.62 , inference loss is 5.11, weight deacy loss is 6.51, training accuracy is 0.468750, time 13.592 samples/sec
+epoch 11, total_step 183320, total loss is 15.92 , inference loss is 9.42, weight deacy loss is 6.51, training accuracy is 0.343750, time 226.253 samples/sec
+epoch 11, total_step 183340, total loss is 14.75 , inference loss is 8.25, weight deacy loss is 6.51, training accuracy is 0.312500, time 231.664 samples/sec
+epoch 11, total_step 183360, total loss is 13.96 , inference loss is 7.45, weight deacy loss is 6.51, training accuracy is 0.312500, time 225.498 samples/sec
+epoch 11, total_step 183380, total loss is 14.96 , inference loss is 8.45, weight deacy loss is 6.51, training accuracy is 0.343750, time 223.812 samples/sec
+epoch 11, total_step 183400, total loss is 13.00 , inference loss is 6.50, weight deacy loss is 6.51, training accuracy is 0.281250, time 226.775 samples/sec
+epoch 11, total_step 183420, total loss is 17.87 , inference loss is 11.37, weight deacy loss is 6.51, training accuracy is 0.312500, time 227.195 samples/sec
+epoch 11, total_step 183440, total loss is 14.30 , inference loss is 7.79, weight deacy loss is 6.51, training accuracy is 0.312500, time 224.712 samples/sec
+epoch 11, total_step 183460, total loss is 13.58 , inference loss is 7.08, weight deacy loss is 6.51, training accuracy is 0.375000, time 222.457 samples/sec
+epoch 11, total_step 183480, total loss is 14.68 , inference loss is 8.17, weight deacy loss is 6.50, training accuracy is 0.281250, time 210.355 samples/sec
+epoch 11, total_step 183500, total loss is 15.42 , inference loss is 8.91, weight deacy loss is 6.50, training accuracy is 0.437500, time 224.817 samples/sec
+epoch 11, total_step 183520, total loss is 12.88 , inference loss is 6.38, weight deacy loss is 6.50, training accuracy is 0.468750, time 226.142 samples/sec
+epoch 11, total_step 183540, total loss is 20.38 , inference loss is 13.88, weight deacy loss is 6.50, training accuracy is 0.281250, time 216.703 samples/sec
+epoch 11, total_step 183560, total loss is 14.31 , inference loss is 7.80, weight deacy loss is 6.50, training accuracy is 0.312500, time 220.946 samples/sec
+epoch 11, total_step 183580, total loss is 13.22 , inference loss is 6.71, weight deacy loss is 6.50, training accuracy is 0.312500, time 298.907 samples/sec
+epoch 11, total_step 183600, total loss is 13.99 , inference loss is 7.49, weight deacy loss is 6.50, training accuracy is 0.375000, time 12.624 samples/sec
+epoch 11, total_step 183620, total loss is 16.19 , inference loss is 9.69, weight deacy loss is 6.50, training accuracy is 0.281250, time 225.812 samples/sec
+epoch 11, total_step 183640, total loss is 14.61 , inference loss is 8.10, weight deacy loss is 6.50, training accuracy is 0.250000, time 221.799 samples/sec
+epoch 11, total_step 183660, total loss is 18.05 , inference loss is 11.55, weight deacy loss is 6.50, training accuracy is 0.281250, time 208.178 samples/sec
+epoch 11, total_step 183680, total loss is 16.88 , inference loss is 10.37, weight deacy loss is 6.50, training accuracy is 0.281250, time 213.346 samples/sec
+epoch 11, total_step 183700, total loss is 16.65 , inference loss is 10.15, weight deacy loss is 6.50, training accuracy is 0.343750, time 175.724 samples/sec
+epoch 11, total_step 183720, total loss is 17.34 , inference loss is 10.84, weight deacy loss is 6.50, training accuracy is 0.375000, time 203.620 samples/sec
+epoch 11, total_step 183740, total loss is 18.84 , inference loss is 12.34, weight deacy loss is 6.50, training accuracy is 0.187500, time 206.791 samples/sec
+epoch 11, total_step 183760, total loss is 14.06 , inference loss is 7.55, weight deacy loss is 6.50, training accuracy is 0.406250, time 204.731 samples/sec
+epoch 11, total_step 183780, total loss is 17.28 , inference loss is 10.78, weight deacy loss is 6.50, training accuracy is 0.375000, time 198.318 samples/sec
+epoch 11, total_step 183800, total loss is 14.77 , inference loss is 8.27, weight deacy loss is 6.50, training accuracy is 0.437500, time 226.324 samples/sec
+epoch 11, total_step 183820, total loss is 11.00 , inference loss is 4.50, weight deacy loss is 6.50, training accuracy is 0.468750, time 191.520 samples/sec
+epoch 11, total_step 183840, total loss is 15.25 , inference loss is 8.75, weight deacy loss is 6.50, training accuracy is 0.406250, time 198.039 samples/sec
+epoch 11, total_step 183860, total loss is 16.11 , inference loss is 9.61, weight deacy loss is 6.50, training accuracy is 0.406250, time 176.148 samples/sec
+epoch 11, total_step 183880, total loss is 12.78 , inference loss is 6.28, weight deacy loss is 6.50, training accuracy is 0.375000, time 184.104 samples/sec
+epoch 11, total_step 183900, total loss is 15.51 , inference loss is 9.01, weight deacy loss is 6.50, training accuracy is 0.375000, time 13.334 samples/sec
+epoch 11, total_step 183920, total loss is 13.77 , inference loss is 7.27, weight deacy loss is 6.50, training accuracy is 0.250000, time 227.426 samples/sec
+epoch 11, total_step 183940, total loss is 16.50 , inference loss is 10.00, weight deacy loss is 6.50, training accuracy is 0.343750, time 201.327 samples/sec
+epoch 11, total_step 183960, total loss is 18.75 , inference loss is 12.25, weight deacy loss is 6.50, training accuracy is 0.250000, time 215.562 samples/sec
+End of epoch 11
+epoch 12, total_step 183980, total loss is 21.01 , inference loss is 14.51, weight deacy loss is 6.50, training accuracy is 0.062500, time 213.428 samples/sec
+testing verification..
+(12000, 512)
+infer time 22.228492000000028
+best_threshold_index 89 0.9814814814814815
+best_threshold_index 89 0.980925925925926
+best_threshold_index 89 0.9805555555555555
+best_threshold_index 89 0.9831481481481481
+best_threshold_index 88 0.9807407407407407
+best_threshold_index 88 0.9798148148148148
+best_threshold_index 88 0.9814814814814815
+best_threshold_index 89 0.9803703703703703
+best_threshold_index 89 0.9801851851851852
+best_threshold_index 89 0.9798148148148148
+[lfw][184000]XNorm: 18.610883
+[lfw][184000]Accuracy-Flip: 0.98017+-0.00848
+test accuracy is: 0.9801666666666667
+epoch 12, total_step 184000, total loss is 12.75 , inference loss is 6.25, weight deacy loss is 6.50, training accuracy is 0.531250, time 17.171 samples/sec
+epoch 12, total_step 184020, total loss is 9.04 , inference loss is 2.54, weight deacy loss is 6.50, training accuracy is 0.750000, time 212.345 samples/sec
+epoch 12, total_step 184040, total loss is 11.09 , inference loss is 4.59, weight deacy loss is 6.50, training accuracy is 0.593750, time 216.887 samples/sec
+epoch 12, total_step 184060, total loss is 9.96 , inference loss is 3.46, weight deacy loss is 6.50, training accuracy is 0.750000, time 214.439 samples/sec
+epoch 12, total_step 184080, total loss is 11.17 , inference loss is 4.67, weight deacy loss is 6.50, training accuracy is 0.718750, time 298.032 samples/sec
+epoch 12, total_step 184100, total loss is 8.47 , inference loss is 1.97, weight deacy loss is 6.50, training accuracy is 0.750000, time 212.758 samples/sec
+epoch 12, total_step 184120, total loss is 10.45 , inference loss is 3.95, weight deacy loss is 6.50, training accuracy is 0.781250, time 207.182 samples/sec
+epoch 12, total_step 184140, total loss is 9.49 , inference loss is 2.99, weight deacy loss is 6.50, training accuracy is 0.750000, time 219.628 samples/sec
+epoch 12, total_step 184160, total loss is 11.27 , inference loss is 4.77, weight deacy loss is 6.50, training accuracy is 0.656250, time 213.926 samples/sec
+epoch 12, total_step 184180, total loss is 11.40 , inference loss is 4.90, weight deacy loss is 6.50, training accuracy is 0.687500, time 219.044 samples/sec
+epoch 12, total_step 184200, total loss is 9.90 , inference loss is 3.40, weight deacy loss is 6.50, training accuracy is 0.687500, time 12.969 samples/sec
+epoch 12, total_step 184220, total loss is 8.68 , inference loss is 2.18, weight deacy loss is 6.50, training accuracy is 0.812500, time 243.749 samples/sec
+epoch 12, total_step 184240, total loss is 12.20 , inference loss is 5.70, weight deacy loss is 6.50, training accuracy is 0.656250, time 211.880 samples/sec
+epoch 12, total_step 184260, total loss is 11.85 , inference loss is 5.35, weight deacy loss is 6.50, training accuracy is 0.687500, time 215.128 samples/sec
+epoch 12, total_step 184280, total loss is 12.26 , inference loss is 5.76, weight deacy loss is 6.50, training accuracy is 0.593750, time 230.791 samples/sec
+epoch 12, total_step 184300, total loss is 11.81 , inference loss is 5.31, weight deacy loss is 6.50, training accuracy is 0.593750, time 215.213 samples/sec
+epoch 12, total_step 184320, total loss is 9.88 , inference loss is 3.39, weight deacy loss is 6.50, training accuracy is 0.593750, time 221.724 samples/sec
+epoch 12, total_step 184340, total loss is 12.21 , inference loss is 5.72, weight deacy loss is 6.50, training accuracy is 0.625000, time 299.520 samples/sec
+epoch 12, total_step 184360, total loss is 11.27 , inference loss is 4.77, weight deacy loss is 6.50, training accuracy is 0.593750, time 210.518 samples/sec
+epoch 12, total_step 184380, total loss is 8.16 , inference loss is 1.66, weight deacy loss is 6.50, training accuracy is 0.718750, time 219.282 samples/sec
+epoch 12, total_step 184400, total loss is 11.12 , inference loss is 4.63, weight deacy loss is 6.50, training accuracy is 0.593750, time 209.016 samples/sec
+epoch 12, total_step 184420, total loss is 10.02 , inference loss is 3.52, weight deacy loss is 6.50, training accuracy is 0.656250, time 223.493 samples/sec
+epoch 12, total_step 184440, total loss is 10.02 , inference loss is 3.52, weight deacy loss is 6.50, training accuracy is 0.625000, time 220.249 samples/sec
+epoch 12, total_step 184460, total loss is 9.41 , inference loss is 2.91, weight deacy loss is 6.50, training accuracy is 0.781250, time 222.719 samples/sec
+epoch 12, total_step 184480, total loss is 11.19 , inference loss is 4.69, weight deacy loss is 6.50, training accuracy is 0.687500, time 220.924 samples/sec
+epoch 12, total_step 184500, total loss is 14.73 , inference loss is 8.23, weight deacy loss is 6.50, training accuracy is 0.625000, time 12.601 samples/sec
+epoch 12, total_step 184520, total loss is 11.95 , inference loss is 5.45, weight deacy loss is 6.50, training accuracy is 0.593750, time 219.853 samples/sec
+epoch 12, total_step 184540, total loss is 14.32 , inference loss is 7.83, weight deacy loss is 6.50, training accuracy is 0.593750, time 215.768 samples/sec
+epoch 12, total_step 184560, total loss is 11.37 , inference loss is 4.88, weight deacy loss is 6.50, training accuracy is 0.718750, time 228.269 samples/sec
+epoch 12, total_step 184580, total loss is 13.76 , inference loss is 7.26, weight deacy loss is 6.50, training accuracy is 0.687500, time 212.050 samples/sec
+epoch 12, total_step 184600, total loss is 14.70 , inference loss is 8.20, weight deacy loss is 6.50, training accuracy is 0.531250, time 297.462 samples/sec
+epoch 12, total_step 184620, total loss is 17.84 , inference loss is 11.34, weight deacy loss is 6.50, training accuracy is 0.500000, time 220.962 samples/sec
+epoch 12, total_step 184640, total loss is 13.93 , inference loss is 7.44, weight deacy loss is 6.50, training accuracy is 0.625000, time 211.481 samples/sec
+epoch 12, total_step 184660, total loss is 12.28 , inference loss is 5.78, weight deacy loss is 6.50, training accuracy is 0.625000, time 224.583 samples/sec
+epoch 12, total_step 184680, total loss is 12.93 , inference loss is 6.43, weight deacy loss is 6.50, training accuracy is 0.531250, time 220.891 samples/sec
+epoch 12, total_step 184700, total loss is 10.48 , inference loss is 3.98, weight deacy loss is 6.50, training accuracy is 0.656250, time 215.170 samples/sec
+epoch 12, total_step 184720, total loss is 16.49 , inference loss is 9.99, weight deacy loss is 6.50, training accuracy is 0.468750, time 226.902 samples/sec
+epoch 12, total_step 184740, total loss is 10.78 , inference loss is 4.28, weight deacy loss is 6.50, training accuracy is 0.593750, time 208.614 samples/sec
+epoch 12, total_step 184760, total loss is 12.37 , inference loss is 5.88, weight deacy loss is 6.50, training accuracy is 0.562500, time 219.481 samples/sec
+epoch 12, total_step 184780, total loss is 16.63 , inference loss is 10.13, weight deacy loss is 6.50, training accuracy is 0.593750, time 212.659 samples/sec
+epoch 12, total_step 184800, total loss is 11.38 , inference loss is 4.88, weight deacy loss is 6.50, training accuracy is 0.593750, time 12.767 samples/sec
+epoch 12, total_step 184820, total loss is 9.26 , inference loss is 2.76, weight deacy loss is 6.50, training accuracy is 0.718750, time 218.634 samples/sec
+epoch 12, total_step 184840, total loss is 15.90 , inference loss is 9.40, weight deacy loss is 6.50, training accuracy is 0.562500, time 218.158 samples/sec
+epoch 12, total_step 184860, total loss is 11.67 , inference loss is 5.18, weight deacy loss is 6.50, training accuracy is 0.562500, time 298.689 samples/sec
+epoch 12, total_step 184880, total loss is 13.55 , inference loss is 7.05, weight deacy loss is 6.50, training accuracy is 0.593750, time 212.684 samples/sec
+epoch 12, total_step 184900, total loss is 16.38 , inference loss is 9.88, weight deacy loss is 6.50, training accuracy is 0.593750, time 226.486 samples/sec
+epoch 12, total_step 184920, total loss is 13.92 , inference loss is 7.42, weight deacy loss is 6.50, training accuracy is 0.500000, time 233.433 samples/sec
+epoch 12, total_step 184940, total loss is 10.21 , inference loss is 3.71, weight deacy loss is 6.50, training accuracy is 0.718750, time 217.462 samples/sec
+epoch 12, total_step 184960, total loss is 10.75 , inference loss is 4.25, weight deacy loss is 6.50, training accuracy is 0.593750, time 214.635 samples/sec
+epoch 12, total_step 184980, total loss is 14.98 , inference loss is 8.48, weight deacy loss is 6.50, training accuracy is 0.468750, time 218.168 samples/sec
+epoch 12, total_step 185000, total loss is 12.29 , inference loss is 5.79, weight deacy loss is 6.50, training accuracy is 0.687500, time 212.354 samples/sec
+epoch 12, total_step 185020, total loss is 14.53 , inference loss is 8.03, weight deacy loss is 6.50, training accuracy is 0.687500, time 207.017 samples/sec
+epoch 12, total_step 185040, total loss is 14.09 , inference loss is 7.59, weight deacy loss is 6.50, training accuracy is 0.500000, time 212.019 samples/sec
+epoch 12, total_step 185060, total loss is 14.02 , inference loss is 7.52, weight deacy loss is 6.50, training accuracy is 0.593750, time 214.455 samples/sec
+epoch 12, total_step 185080, total loss is 14.53 , inference loss is 8.02, weight deacy loss is 6.50, training accuracy is 0.468750, time 217.476 samples/sec
+epoch 12, total_step 185100, total loss is 13.66 , inference loss is 7.15, weight deacy loss is 6.50, training accuracy is 0.406250, time 12.708 samples/sec
+epoch 12, total_step 185120, total loss is 10.77 , inference loss is 4.27, weight deacy loss is 6.50, training accuracy is 0.562500, time 294.845 samples/sec
+epoch 12, total_step 185140, total loss is 11.77 , inference loss is 5.27, weight deacy loss is 6.50, training accuracy is 0.500000, time 216.952 samples/sec
+epoch 12, total_step 185160, total loss is 9.55 , inference loss is 3.05, weight deacy loss is 6.50, training accuracy is 0.562500, time 212.826 samples/sec
+epoch 12, total_step 185180, total loss is 14.73 , inference loss is 8.23, weight deacy loss is 6.50, training accuracy is 0.593750, time 210.758 samples/sec
+epoch 12, total_step 185200, total loss is 12.15 , inference loss is 5.65, weight deacy loss is 6.50, training accuracy is 0.687500, time 220.411 samples/sec
+epoch 12, total_step 185220, total loss is 12.22 , inference loss is 5.72, weight deacy loss is 6.50, training accuracy is 0.656250, time 224.414 samples/sec
+epoch 12, total_step 185240, total loss is 12.11 , inference loss is 5.61, weight deacy loss is 6.50, training accuracy is 0.562500, time 215.704 samples/sec
+epoch 12, total_step 185260, total loss is 11.45 , inference loss is 4.95, weight deacy loss is 6.50, training accuracy is 0.593750, time 224.593 samples/sec
+epoch 12, total_step 185280, total loss is 14.53 , inference loss is 8.03, weight deacy loss is 6.50, training accuracy is 0.500000, time 226.035 samples/sec
+epoch 12, total_step 185300, total loss is 15.22 , inference loss is 8.72, weight deacy loss is 6.50, training accuracy is 0.593750, time 216.998 samples/sec
+epoch 12, total_step 185320, total loss is 12.39 , inference loss is 5.88, weight deacy loss is 6.50, training accuracy is 0.625000, time 223.898 samples/sec
+epoch 12, total_step 185340, total loss is 12.85 , inference loss is 6.34, weight deacy loss is 6.50, training accuracy is 0.562500, time 218.942 samples/sec
+epoch 12, total_step 185360, total loss is 14.13 , inference loss is 7.63, weight deacy loss is 6.50, training accuracy is 0.531250, time 229.188 samples/sec
+epoch 12, total_step 185380, total loss is 8.76 , inference loss is 2.26, weight deacy loss is 6.50, training accuracy is 0.812500, time 212.270 samples/sec
+epoch 12, total_step 185400, total loss is 10.54 , inference loss is 4.03, weight deacy loss is 6.50, training accuracy is 0.750000, time 13.327 samples/sec
+epoch 12, total_step 185420, total loss is 14.27 , inference loss is 7.77, weight deacy loss is 6.50, training accuracy is 0.562500, time 221.879 samples/sec
+epoch 12, total_step 185440, total loss is 11.86 , inference loss is 5.36, weight deacy loss is 6.50, training accuracy is 0.500000, time 224.432 samples/sec
+epoch 12, total_step 185460, total loss is 11.86 , inference loss is 5.36, weight deacy loss is 6.50, training accuracy is 0.625000, time 216.973 samples/sec
+epoch 12, total_step 185480, total loss is 13.07 , inference loss is 6.57, weight deacy loss is 6.50, training accuracy is 0.656250, time 230.809 samples/sec
+epoch 12, total_step 185500, total loss is 14.05 , inference loss is 7.55, weight deacy loss is 6.50, training accuracy is 0.500000, time 218.026 samples/sec
+epoch 12, total_step 185520, total loss is 16.09 , inference loss is 9.58, weight deacy loss is 6.50, training accuracy is 0.500000, time 219.778 samples/sec
+epoch 12, total_step 185540, total loss is 13.10 , inference loss is 6.60, weight deacy loss is 6.51, training accuracy is 0.562500, time 232.160 samples/sec
+epoch 12, total_step 185560, total loss is 16.07 , inference loss is 9.56, weight deacy loss is 6.51, training accuracy is 0.531250, time 218.072 samples/sec
+epoch 12, total_step 185580, total loss is 15.34 , inference loss is 8.83, weight deacy loss is 6.51, training accuracy is 0.437500, time 215.530 samples/sec
+epoch 12, total_step 185600, total loss is 9.26 , inference loss is 2.76, weight deacy loss is 6.51, training accuracy is 0.656250, time 218.000 samples/sec
+epoch 12, total_step 185620, total loss is 13.46 , inference loss is 6.96, weight deacy loss is 6.51, training accuracy is 0.500000, time 206.683 samples/sec
+epoch 12, total_step 185640, total loss is 12.26 , inference loss is 5.75, weight deacy loss is 6.51, training accuracy is 0.656250, time 210.121 samples/sec
+epoch 12, total_step 185660, total loss is 13.97 , inference loss is 7.47, weight deacy loss is 6.51, training accuracy is 0.468750, time 216.867 samples/sec
+epoch 12, total_step 185680, total loss is 10.12 , inference loss is 3.62, weight deacy loss is 6.51, training accuracy is 0.656250, time 220.227 samples/sec
+epoch 12, total_step 185700, total loss is 16.76 , inference loss is 10.25, weight deacy loss is 6.51, training accuracy is 0.500000, time 12.594 samples/sec
+epoch 12, total_step 185720, total loss is 13.34 , inference loss is 6.84, weight deacy loss is 6.51, training accuracy is 0.531250, time 209.810 samples/sec
+epoch 12, total_step 185740, total loss is 11.57 , inference loss is 5.06, weight deacy loss is 6.51, training accuracy is 0.531250, time 214.989 samples/sec
+epoch 12, total_step 185760, total loss is 12.49 , inference loss is 5.98, weight deacy loss is 6.51, training accuracy is 0.500000, time 215.880 samples/sec
+epoch 12, total_step 185780, total loss is 9.90 , inference loss is 3.40, weight deacy loss is 6.51, training accuracy is 0.718750, time 222.611 samples/sec
+epoch 12, total_step 185800, total loss is 12.78 , inference loss is 6.27, weight deacy loss is 6.51, training accuracy is 0.593750, time 219.791 samples/sec
+epoch 12, total_step 185820, total loss is 11.48 , inference loss is 4.98, weight deacy loss is 6.51, training accuracy is 0.656250, time 224.250 samples/sec
+epoch 12, total_step 185840, total loss is 12.75 , inference loss is 6.25, weight deacy loss is 6.51, training accuracy is 0.531250, time 214.290 samples/sec
+epoch 12, total_step 185860, total loss is 10.75 , inference loss is 4.24, weight deacy loss is 6.51, training accuracy is 0.562500, time 214.861 samples/sec
+epoch 12, total_step 185880, total loss is 12.92 , inference loss is 6.41, weight deacy loss is 6.51, training accuracy is 0.718750, time 218.358 samples/sec
+epoch 12, total_step 185900, total loss is 11.20 , inference loss is 4.70, weight deacy loss is 6.51, training accuracy is 0.500000, time 216.856 samples/sec
+epoch 12, total_step 185920, total loss is 13.55 , inference loss is 7.05, weight deacy loss is 6.51, training accuracy is 0.625000, time 297.737 samples/sec
+epoch 12, total_step 185940, total loss is 14.98 , inference loss is 8.47, weight deacy loss is 6.51, training accuracy is 0.468750, time 299.128 samples/sec
+epoch 12, total_step 185960, total loss is 12.96 , inference loss is 6.45, weight deacy loss is 6.51, training accuracy is 0.656250, time 215.111 samples/sec
+epoch 12, total_step 185980, total loss is 11.76 , inference loss is 5.26, weight deacy loss is 6.51, training accuracy is 0.500000, time 217.643 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.169746999999983
+best_threshold_index 120 0.9875925925925926
+best_threshold_index 124 0.9875925925925926
+best_threshold_index 120 0.987037037037037
+best_threshold_index 117 0.9879629629629629
+best_threshold_index 120 0.9875925925925926
+best_threshold_index 124 0.9872222222222222
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 117 0.9868518518518519
+best_threshold_index 124 0.9862962962962963
+best_threshold_index 120 0.987037037037037
+[lfw][186000]XNorm: 22.228304
+[lfw][186000]Accuracy-Flip: 0.98567+-0.00442
+test accuracy is: 0.9856666666666666
+epoch 12, total_step 186000, total loss is 13.65 , inference loss is 7.15, weight deacy loss is 6.51, training accuracy is 0.531250, time 16.927 samples/sec
+epoch 12, total_step 186020, total loss is 12.02 , inference loss is 5.51, weight deacy loss is 6.51, training accuracy is 0.718750, time 226.677 samples/sec
+epoch 12, total_step 186040, total loss is 14.27 , inference loss is 7.76, weight deacy loss is 6.51, training accuracy is 0.437500, time 213.839 samples/sec
+epoch 12, total_step 186060, total loss is 12.84 , inference loss is 6.33, weight deacy loss is 6.51, training accuracy is 0.562500, time 210.329 samples/sec
+epoch 12, total_step 186080, total loss is 16.10 , inference loss is 9.60, weight deacy loss is 6.51, training accuracy is 0.343750, time 216.447 samples/sec
+epoch 12, total_step 186100, total loss is 13.96 , inference loss is 7.45, weight deacy loss is 6.51, training accuracy is 0.437500, time 210.546 samples/sec
+epoch 12, total_step 186120, total loss is 8.02 , inference loss is 1.51, weight deacy loss is 6.51, training accuracy is 0.781250, time 208.838 samples/sec
+epoch 12, total_step 186140, total loss is 10.87 , inference loss is 4.37, weight deacy loss is 6.50, training accuracy is 0.531250, time 222.584 samples/sec
+epoch 12, total_step 186160, total loss is 11.47 , inference loss is 4.96, weight deacy loss is 6.50, training accuracy is 0.562500, time 207.670 samples/sec
+epoch 12, total_step 186180, total loss is 14.03 , inference loss is 7.53, weight deacy loss is 6.50, training accuracy is 0.562500, time 216.494 samples/sec
+epoch 12, total_step 186200, total loss is 12.64 , inference loss is 6.14, weight deacy loss is 6.50, training accuracy is 0.437500, time 230.678 samples/sec
+epoch 12, total_step 186220, total loss is 11.92 , inference loss is 5.41, weight deacy loss is 6.50, training accuracy is 0.468750, time 298.526 samples/sec
+epoch 12, total_step 186240, total loss is 12.80 , inference loss is 6.29, weight deacy loss is 6.50, training accuracy is 0.625000, time 213.611 samples/sec
+epoch 12, total_step 186260, total loss is 11.31 , inference loss is 4.81, weight deacy loss is 6.50, training accuracy is 0.687500, time 209.125 samples/sec
+epoch 12, total_step 186280, total loss is 13.59 , inference loss is 7.09, weight deacy loss is 6.50, training accuracy is 0.500000, time 211.649 samples/sec
+epoch 12, total_step 186300, total loss is 12.30 , inference loss is 5.79, weight deacy loss is 6.50, training accuracy is 0.531250, time 12.524 samples/sec
+epoch 12, total_step 186320, total loss is 9.33 , inference loss is 2.82, weight deacy loss is 6.50, training accuracy is 0.781250, time 217.818 samples/sec
+epoch 12, total_step 186340, total loss is 9.73 , inference loss is 3.22, weight deacy loss is 6.50, training accuracy is 0.718750, time 209.388 samples/sec
+epoch 12, total_step 186360, total loss is 12.96 , inference loss is 6.46, weight deacy loss is 6.50, training accuracy is 0.500000, time 216.967 samples/sec
+epoch 12, total_step 186380, total loss is 11.24 , inference loss is 4.73, weight deacy loss is 6.50, training accuracy is 0.687500, time 210.370 samples/sec
+epoch 12, total_step 186400, total loss is 8.13 , inference loss is 1.63, weight deacy loss is 6.50, training accuracy is 0.812500, time 231.358 samples/sec
+epoch 12, total_step 186420, total loss is 11.98 , inference loss is 5.48, weight deacy loss is 6.50, training accuracy is 0.562500, time 215.835 samples/sec
+epoch 12, total_step 186440, total loss is 9.39 , inference loss is 2.89, weight deacy loss is 6.50, training accuracy is 0.687500, time 224.492 samples/sec
+epoch 12, total_step 186460, total loss is 11.45 , inference loss is 4.95, weight deacy loss is 6.50, training accuracy is 0.593750, time 212.459 samples/sec
+epoch 12, total_step 186480, total loss is 11.08 , inference loss is 4.58, weight deacy loss is 6.50, training accuracy is 0.656250, time 188.332 samples/sec
+epoch 12, total_step 186500, total loss is 12.02 , inference loss is 5.51, weight deacy loss is 6.50, training accuracy is 0.593750, time 206.035 samples/sec
+epoch 12, total_step 186520, total loss is 11.63 , inference loss is 5.13, weight deacy loss is 6.50, training accuracy is 0.562500, time 230.560 samples/sec
+epoch 12, total_step 186540, total loss is 8.41 , inference loss is 1.91, weight deacy loss is 6.50, training accuracy is 0.687500, time 220.337 samples/sec
+epoch 12, total_step 186560, total loss is 10.81 , inference loss is 4.31, weight deacy loss is 6.50, training accuracy is 0.656250, time 233.785 samples/sec
+epoch 12, total_step 186580, total loss is 9.65 , inference loss is 3.15, weight deacy loss is 6.50, training accuracy is 0.687500, time 215.099 samples/sec
+epoch 12, total_step 186600, total loss is 12.99 , inference loss is 6.48, weight deacy loss is 6.50, training accuracy is 0.625000, time 12.686 samples/sec
+epoch 12, total_step 186620, total loss is 10.98 , inference loss is 4.48, weight deacy loss is 6.50, training accuracy is 0.656250, time 225.890 samples/sec
+epoch 12, total_step 186640, total loss is 14.92 , inference loss is 8.41, weight deacy loss is 6.50, training accuracy is 0.437500, time 240.454 samples/sec
+epoch 12, total_step 186660, total loss is 15.19 , inference loss is 8.69, weight deacy loss is 6.50, training accuracy is 0.531250, time 219.637 samples/sec
+epoch 12, total_step 186680, total loss is 11.79 , inference loss is 5.29, weight deacy loss is 6.50, training accuracy is 0.718750, time 218.952 samples/sec
+epoch 12, total_step 186700, total loss is 11.46 , inference loss is 4.96, weight deacy loss is 6.50, training accuracy is 0.562500, time 235.507 samples/sec
+epoch 12, total_step 186720, total loss is 11.08 , inference loss is 4.58, weight deacy loss is 6.50, training accuracy is 0.687500, time 215.355 samples/sec
+epoch 12, total_step 186740, total loss is 12.22 , inference loss is 5.72, weight deacy loss is 6.50, training accuracy is 0.656250, time 209.974 samples/sec
+epoch 12, total_step 186760, total loss is 11.15 , inference loss is 4.65, weight deacy loss is 6.50, training accuracy is 0.625000, time 296.559 samples/sec
+epoch 12, total_step 186780, total loss is 11.05 , inference loss is 4.55, weight deacy loss is 6.50, training accuracy is 0.656250, time 212.789 samples/sec
+epoch 12, total_step 186800, total loss is 10.85 , inference loss is 4.35, weight deacy loss is 6.50, training accuracy is 0.593750, time 217.273 samples/sec
+epoch 12, total_step 186820, total loss is 10.45 , inference loss is 3.95, weight deacy loss is 6.50, training accuracy is 0.687500, time 214.203 samples/sec
+epoch 12, total_step 186840, total loss is 11.40 , inference loss is 4.90, weight deacy loss is 6.50, training accuracy is 0.656250, time 216.879 samples/sec
+epoch 12, total_step 186860, total loss is 11.76 , inference loss is 5.26, weight deacy loss is 6.50, training accuracy is 0.718750, time 219.848 samples/sec
+epoch 12, total_step 186880, total loss is 13.84 , inference loss is 7.34, weight deacy loss is 6.50, training accuracy is 0.593750, time 236.742 samples/sec
+epoch 12, total_step 186900, total loss is 13.24 , inference loss is 6.74, weight deacy loss is 6.50, training accuracy is 0.562500, time 12.342 samples/sec
+epoch 12, total_step 186920, total loss is 8.06 , inference loss is 1.56, weight deacy loss is 6.50, training accuracy is 0.750000, time 219.584 samples/sec
+epoch 12, total_step 186940, total loss is 10.83 , inference loss is 4.33, weight deacy loss is 6.50, training accuracy is 0.625000, time 214.118 samples/sec
+epoch 12, total_step 186960, total loss is 8.92 , inference loss is 2.42, weight deacy loss is 6.50, training accuracy is 0.656250, time 219.948 samples/sec
+epoch 12, total_step 186980, total loss is 10.50 , inference loss is 4.00, weight deacy loss is 6.50, training accuracy is 0.531250, time 215.521 samples/sec
+epoch 12, total_step 187000, total loss is 8.07 , inference loss is 1.57, weight deacy loss is 6.50, training accuracy is 0.656250, time 208.001 samples/sec
+epoch 12, total_step 187020, total loss is 15.41 , inference loss is 8.91, weight deacy loss is 6.50, training accuracy is 0.531250, time 299.110 samples/sec
+epoch 12, total_step 187040, total loss is 12.31 , inference loss is 5.82, weight deacy loss is 6.50, training accuracy is 0.687500, time 227.134 samples/sec
+epoch 12, total_step 187060, total loss is 12.31 , inference loss is 5.81, weight deacy loss is 6.50, training accuracy is 0.562500, time 215.896 samples/sec
+epoch 12, total_step 187080, total loss is 13.57 , inference loss is 7.07, weight deacy loss is 6.50, training accuracy is 0.625000, time 210.751 samples/sec
+epoch 12, total_step 187100, total loss is 12.14 , inference loss is 5.64, weight deacy loss is 6.50, training accuracy is 0.593750, time 210.808 samples/sec
+epoch 12, total_step 187120, total loss is 12.72 , inference loss is 6.22, weight deacy loss is 6.50, training accuracy is 0.625000, time 224.486 samples/sec
+epoch 12, total_step 187140, total loss is 12.58 , inference loss is 6.08, weight deacy loss is 6.50, training accuracy is 0.593750, time 222.584 samples/sec
+epoch 12, total_step 187160, total loss is 13.82 , inference loss is 7.32, weight deacy loss is 6.50, training accuracy is 0.500000, time 237.834 samples/sec
+epoch 12, total_step 187180, total loss is 13.68 , inference loss is 7.18, weight deacy loss is 6.50, training accuracy is 0.593750, time 202.109 samples/sec
+epoch 12, total_step 187200, total loss is 12.99 , inference loss is 6.49, weight deacy loss is 6.50, training accuracy is 0.500000, time 14.141 samples/sec
+epoch 12, total_step 187220, total loss is 12.74 , inference loss is 6.24, weight deacy loss is 6.50, training accuracy is 0.562500, time 234.236 samples/sec
+epoch 12, total_step 187240, total loss is 15.47 , inference loss is 8.98, weight deacy loss is 6.50, training accuracy is 0.625000, time 210.173 samples/sec
+epoch 12, total_step 187260, total loss is 13.41 , inference loss is 6.91, weight deacy loss is 6.50, training accuracy is 0.687500, time 211.941 samples/sec
+epoch 12, total_step 187280, total loss is 14.07 , inference loss is 7.57, weight deacy loss is 6.50, training accuracy is 0.562500, time 299.582 samples/sec
+epoch 12, total_step 187300, total loss is 13.27 , inference loss is 6.77, weight deacy loss is 6.50, training accuracy is 0.562500, time 228.218 samples/sec
+epoch 12, total_step 187320, total loss is 16.39 , inference loss is 9.90, weight deacy loss is 6.50, training accuracy is 0.375000, time 229.156 samples/sec
+epoch 12, total_step 187340, total loss is 13.68 , inference loss is 7.18, weight deacy loss is 6.50, training accuracy is 0.531250, time 214.561 samples/sec
+epoch 12, total_step 187360, total loss is 14.99 , inference loss is 8.49, weight deacy loss is 6.50, training accuracy is 0.625000, time 224.867 samples/sec
+epoch 12, total_step 187380, total loss is 11.59 , inference loss is 5.09, weight deacy loss is 6.50, training accuracy is 0.531250, time 216.827 samples/sec
+epoch 12, total_step 187400, total loss is 12.45 , inference loss is 5.95, weight deacy loss is 6.50, training accuracy is 0.468750, time 207.803 samples/sec
+epoch 12, total_step 187420, total loss is 15.08 , inference loss is 8.58, weight deacy loss is 6.50, training accuracy is 0.468750, time 212.958 samples/sec
+epoch 12, total_step 187440, total loss is 16.19 , inference loss is 9.69, weight deacy loss is 6.50, training accuracy is 0.500000, time 212.132 samples/sec
+epoch 12, total_step 187460, total loss is 14.92 , inference loss is 8.43, weight deacy loss is 6.50, training accuracy is 0.531250, time 210.943 samples/sec
+epoch 12, total_step 187480, total loss is 10.64 , inference loss is 4.14, weight deacy loss is 6.50, training accuracy is 0.625000, time 226.866 samples/sec
+epoch 12, total_step 187500, total loss is 10.49 , inference loss is 3.99, weight deacy loss is 6.50, training accuracy is 0.750000, time 12.766 samples/sec
+epoch 12, total_step 187520, total loss is 17.86 , inference loss is 11.37, weight deacy loss is 6.50, training accuracy is 0.375000, time 214.689 samples/sec
+epoch 12, total_step 187540, total loss is 13.79 , inference loss is 7.29, weight deacy loss is 6.50, training accuracy is 0.406250, time 300.227 samples/sec
+epoch 12, total_step 187560, total loss is 14.18 , inference loss is 7.68, weight deacy loss is 6.50, training accuracy is 0.437500, time 216.477 samples/sec
+epoch 12, total_step 187580, total loss is 16.24 , inference loss is 9.74, weight deacy loss is 6.50, training accuracy is 0.468750, time 215.204 samples/sec
+epoch 12, total_step 187600, total loss is 15.86 , inference loss is 9.36, weight deacy loss is 6.50, training accuracy is 0.281250, time 211.492 samples/sec
+epoch 12, total_step 187620, total loss is 10.69 , inference loss is 4.19, weight deacy loss is 6.50, training accuracy is 0.625000, time 228.173 samples/sec
+epoch 12, total_step 187640, total loss is 12.87 , inference loss is 6.37, weight deacy loss is 6.50, training accuracy is 0.500000, time 216.060 samples/sec
+epoch 12, total_step 187660, total loss is 12.90 , inference loss is 6.41, weight deacy loss is 6.50, training accuracy is 0.468750, time 213.792 samples/sec
+epoch 12, total_step 187680, total loss is 15.18 , inference loss is 8.69, weight deacy loss is 6.50, training accuracy is 0.531250, time 214.594 samples/sec
+epoch 12, total_step 187700, total loss is 11.33 , inference loss is 4.83, weight deacy loss is 6.50, training accuracy is 0.656250, time 228.443 samples/sec
+epoch 12, total_step 187720, total loss is 14.76 , inference loss is 8.26, weight deacy loss is 6.50, training accuracy is 0.468750, time 296.051 samples/sec
+epoch 12, total_step 187740, total loss is 16.07 , inference loss is 9.58, weight deacy loss is 6.50, training accuracy is 0.406250, time 239.254 samples/sec
+epoch 12, total_step 187760, total loss is 14.02 , inference loss is 7.52, weight deacy loss is 6.50, training accuracy is 0.406250, time 219.326 samples/sec
+epoch 12, total_step 187780, total loss is 14.55 , inference loss is 8.05, weight deacy loss is 6.50, training accuracy is 0.562500, time 220.905 samples/sec
+epoch 12, total_step 187800, total loss is 14.20 , inference loss is 7.70, weight deacy loss is 6.50, training accuracy is 0.375000, time 13.752 samples/sec
+epoch 12, total_step 187820, total loss is 17.15 , inference loss is 10.65, weight deacy loss is 6.50, training accuracy is 0.437500, time 212.884 samples/sec
+epoch 12, total_step 187840, total loss is 10.96 , inference loss is 4.46, weight deacy loss is 6.50, training accuracy is 0.468750, time 228.287 samples/sec
+epoch 12, total_step 187860, total loss is 14.78 , inference loss is 8.28, weight deacy loss is 6.50, training accuracy is 0.468750, time 220.622 samples/sec
+epoch 12, total_step 187880, total loss is 13.65 , inference loss is 7.15, weight deacy loss is 6.50, training accuracy is 0.468750, time 213.435 samples/sec
+epoch 12, total_step 187900, total loss is 16.75 , inference loss is 10.25, weight deacy loss is 6.50, training accuracy is 0.437500, time 213.352 samples/sec
+epoch 12, total_step 187920, total loss is 16.45 , inference loss is 9.95, weight deacy loss is 6.50, training accuracy is 0.468750, time 215.704 samples/sec
+epoch 12, total_step 187940, total loss is 13.87 , inference loss is 7.37, weight deacy loss is 6.50, training accuracy is 0.406250, time 217.805 samples/sec
+epoch 12, total_step 187960, total loss is 15.43 , inference loss is 8.93, weight deacy loss is 6.50, training accuracy is 0.500000, time 220.670 samples/sec
+epoch 12, total_step 187980, total loss is 12.38 , inference loss is 5.89, weight deacy loss is 6.50, training accuracy is 0.500000, time 215.763 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.977706999999988
+best_threshold_index 120 0.9892592592592593
+best_threshold_index 120 0.9892592592592593
+best_threshold_index 120 0.9890740740740741
+best_threshold_index 118 0.9898148148148148
+best_threshold_index 122 0.9898148148148148
+best_threshold_index 120 0.9887037037037038
+best_threshold_index 120 0.9896296296296296
+best_threshold_index 120 0.9883333333333333
+best_threshold_index 122 0.9881481481481481
+best_threshold_index 120 0.9883333333333333
+[lfw][188000]XNorm: 19.732608
+[lfw][188000]Accuracy-Flip: 0.98833+-0.00587
+test accuracy is: 0.9883333333333333
+epoch 12, total_step 188000, total loss is 12.22 , inference loss is 5.72, weight deacy loss is 6.50, training accuracy is 0.593750, time 16.567 samples/sec
+epoch 12, total_step 188020, total loss is 16.72 , inference loss is 10.22, weight deacy loss is 6.50, training accuracy is 0.437500, time 214.436 samples/sec
+epoch 12, total_step 188040, total loss is 15.73 , inference loss is 9.24, weight deacy loss is 6.50, training accuracy is 0.406250, time 220.973 samples/sec
+epoch 12, total_step 188060, total loss is 15.30 , inference loss is 8.81, weight deacy loss is 6.50, training accuracy is 0.468750, time 215.286 samples/sec
+epoch 12, total_step 188080, total loss is 11.37 , inference loss is 4.88, weight deacy loss is 6.50, training accuracy is 0.593750, time 212.910 samples/sec
+epoch 12, total_step 188100, total loss is 13.90 , inference loss is 7.40, weight deacy loss is 6.50, training accuracy is 0.468750, time 13.605 samples/sec
+epoch 12, total_step 188120, total loss is 14.48 , inference loss is 7.98, weight deacy loss is 6.50, training accuracy is 0.500000, time 217.114 samples/sec
+epoch 12, total_step 188140, total loss is 12.19 , inference loss is 5.69, weight deacy loss is 6.50, training accuracy is 0.593750, time 221.328 samples/sec
+epoch 12, total_step 188160, total loss is 9.75 , inference loss is 3.26, weight deacy loss is 6.50, training accuracy is 0.531250, time 216.530 samples/sec
+epoch 12, total_step 188180, total loss is 14.98 , inference loss is 8.48, weight deacy loss is 6.50, training accuracy is 0.375000, time 229.514 samples/sec
+epoch 12, total_step 188200, total loss is 11.38 , inference loss is 4.88, weight deacy loss is 6.50, training accuracy is 0.500000, time 210.903 samples/sec
+epoch 12, total_step 188220, total loss is 15.32 , inference loss is 8.82, weight deacy loss is 6.50, training accuracy is 0.437500, time 226.180 samples/sec
+epoch 12, total_step 188240, total loss is 10.99 , inference loss is 4.50, weight deacy loss is 6.50, training accuracy is 0.500000, time 211.113 samples/sec
+epoch 12, total_step 188260, total loss is 10.50 , inference loss is 4.00, weight deacy loss is 6.50, training accuracy is 0.625000, time 220.021 samples/sec
+epoch 12, total_step 188280, total loss is 14.24 , inference loss is 7.74, weight deacy loss is 6.50, training accuracy is 0.468750, time 224.751 samples/sec
+epoch 12, total_step 188300, total loss is 13.36 , inference loss is 6.86, weight deacy loss is 6.50, training accuracy is 0.625000, time 208.010 samples/sec
+epoch 12, total_step 188320, total loss is 16.09 , inference loss is 9.59, weight deacy loss is 6.50, training accuracy is 0.437500, time 234.077 samples/sec
+epoch 12, total_step 188340, total loss is 16.91 , inference loss is 10.42, weight deacy loss is 6.50, training accuracy is 0.437500, time 210.546 samples/sec
+epoch 12, total_step 188360, total loss is 10.34 , inference loss is 3.84, weight deacy loss is 6.50, training accuracy is 0.531250, time 297.223 samples/sec
+epoch 12, total_step 188380, total loss is 16.05 , inference loss is 9.55, weight deacy loss is 6.50, training accuracy is 0.531250, time 222.959 samples/sec
+epoch 12, total_step 188400, total loss is 13.16 , inference loss is 6.66, weight deacy loss is 6.50, training accuracy is 0.531250, time 12.821 samples/sec
+epoch 12, total_step 188420, total loss is 16.92 , inference loss is 10.42, weight deacy loss is 6.50, training accuracy is 0.343750, time 228.054 samples/sec
+epoch 12, total_step 188440, total loss is 10.93 , inference loss is 4.44, weight deacy loss is 6.50, training accuracy is 0.656250, time 216.848 samples/sec
+epoch 12, total_step 188460, total loss is 14.57 , inference loss is 8.07, weight deacy loss is 6.50, training accuracy is 0.500000, time 230.479 samples/sec
+epoch 12, total_step 188480, total loss is 12.56 , inference loss is 6.06, weight deacy loss is 6.50, training accuracy is 0.531250, time 225.973 samples/sec
+epoch 12, total_step 188500, total loss is 15.05 , inference loss is 8.55, weight deacy loss is 6.50, training accuracy is 0.375000, time 227.281 samples/sec
+epoch 12, total_step 188520, total loss is 12.45 , inference loss is 5.96, weight deacy loss is 6.50, training accuracy is 0.500000, time 219.034 samples/sec
+epoch 12, total_step 188540, total loss is 14.23 , inference loss is 7.74, weight deacy loss is 6.49, training accuracy is 0.500000, time 216.796 samples/sec
+epoch 12, total_step 188560, total loss is 14.74 , inference loss is 8.25, weight deacy loss is 6.49, training accuracy is 0.656250, time 214.383 samples/sec
+epoch 12, total_step 188580, total loss is 12.46 , inference loss is 5.97, weight deacy loss is 6.49, training accuracy is 0.406250, time 217.008 samples/sec
+epoch 12, total_step 188600, total loss is 8.87 , inference loss is 2.38, weight deacy loss is 6.49, training accuracy is 0.687500, time 219.111 samples/sec
+epoch 12, total_step 188620, total loss is 13.30 , inference loss is 6.81, weight deacy loss is 6.49, training accuracy is 0.562500, time 296.264 samples/sec
+epoch 12, total_step 188640, total loss is 10.34 , inference loss is 3.84, weight deacy loss is 6.49, training accuracy is 0.531250, time 217.415 samples/sec
+epoch 12, total_step 188660, total loss is 16.30 , inference loss is 9.80, weight deacy loss is 6.49, training accuracy is 0.468750, time 224.559 samples/sec
+epoch 12, total_step 188680, total loss is 15.41 , inference loss is 8.91, weight deacy loss is 6.49, training accuracy is 0.500000, time 209.198 samples/sec
+epoch 12, total_step 188700, total loss is 17.73 , inference loss is 11.23, weight deacy loss is 6.49, training accuracy is 0.562500, time 12.612 samples/sec
+epoch 12, total_step 188720, total loss is 14.78 , inference loss is 8.29, weight deacy loss is 6.49, training accuracy is 0.500000, time 215.307 samples/sec
+epoch 12, total_step 188740, total loss is 13.16 , inference loss is 6.66, weight deacy loss is 6.49, training accuracy is 0.562500, time 222.525 samples/sec
+epoch 12, total_step 188760, total loss is 15.68 , inference loss is 9.19, weight deacy loss is 6.49, training accuracy is 0.312500, time 216.180 samples/sec
+epoch 12, total_step 188780, total loss is 9.57 , inference loss is 3.08, weight deacy loss is 6.49, training accuracy is 0.562500, time 222.869 samples/sec
+epoch 12, total_step 188800, total loss is 11.41 , inference loss is 4.92, weight deacy loss is 6.49, training accuracy is 0.593750, time 213.012 samples/sec
+epoch 12, total_step 188820, total loss is 14.26 , inference loss is 7.77, weight deacy loss is 6.49, training accuracy is 0.468750, time 216.534 samples/sec
+epoch 12, total_step 188840, total loss is 16.16 , inference loss is 9.67, weight deacy loss is 6.49, training accuracy is 0.437500, time 206.659 samples/sec
+epoch 12, total_step 188860, total loss is 18.11 , inference loss is 11.62, weight deacy loss is 6.49, training accuracy is 0.312500, time 211.867 samples/sec
+epoch 12, total_step 188880, total loss is 10.18 , inference loss is 3.69, weight deacy loss is 6.49, training accuracy is 0.718750, time 298.845 samples/sec
+epoch 12, total_step 188900, total loss is 12.85 , inference loss is 6.36, weight deacy loss is 6.49, training accuracy is 0.468750, time 226.051 samples/sec
+epoch 12, total_step 188920, total loss is 13.68 , inference loss is 7.19, weight deacy loss is 6.49, training accuracy is 0.343750, time 212.239 samples/sec
+epoch 12, total_step 188940, total loss is 12.30 , inference loss is 5.80, weight deacy loss is 6.49, training accuracy is 0.531250, time 214.567 samples/sec
+epoch 12, total_step 188960, total loss is 10.99 , inference loss is 4.50, weight deacy loss is 6.49, training accuracy is 0.656250, time 220.961 samples/sec
+epoch 12, total_step 188980, total loss is 15.53 , inference loss is 9.03, weight deacy loss is 6.49, training accuracy is 0.468750, time 220.781 samples/sec
+epoch 12, total_step 189000, total loss is 15.41 , inference loss is 8.92, weight deacy loss is 6.49, training accuracy is 0.468750, time 12.468 samples/sec
+epoch 12, total_step 189020, total loss is 13.50 , inference loss is 7.00, weight deacy loss is 6.49, training accuracy is 0.406250, time 231.424 samples/sec
+epoch 12, total_step 189040, total loss is 11.74 , inference loss is 5.24, weight deacy loss is 6.49, training accuracy is 0.562500, time 214.180 samples/sec
+epoch 12, total_step 189060, total loss is 12.41 , inference loss is 5.91, weight deacy loss is 6.49, training accuracy is 0.375000, time 220.152 samples/sec
+epoch 12, total_step 189080, total loss is 12.34 , inference loss is 5.84, weight deacy loss is 6.49, training accuracy is 0.500000, time 216.370 samples/sec
+epoch 12, total_step 189100, total loss is 12.03 , inference loss is 5.53, weight deacy loss is 6.49, training accuracy is 0.562500, time 227.572 samples/sec
+epoch 12, total_step 189120, total loss is 14.23 , inference loss is 7.73, weight deacy loss is 6.49, training accuracy is 0.468750, time 227.549 samples/sec
+epoch 12, total_step 189140, total loss is 11.43 , inference loss is 4.93, weight deacy loss is 6.49, training accuracy is 0.500000, time 295.018 samples/sec
+epoch 12, total_step 189160, total loss is 18.39 , inference loss is 11.90, weight deacy loss is 6.49, training accuracy is 0.343750, time 215.393 samples/sec
+epoch 12, total_step 189180, total loss is 13.91 , inference loss is 7.42, weight deacy loss is 6.49, training accuracy is 0.375000, time 220.377 samples/sec
+epoch 12, total_step 189200, total loss is 11.17 , inference loss is 4.68, weight deacy loss is 6.49, training accuracy is 0.625000, time 234.354 samples/sec
+epoch 12, total_step 189220, total loss is 14.11 , inference loss is 7.61, weight deacy loss is 6.49, training accuracy is 0.562500, time 222.404 samples/sec
+epoch 12, total_step 189240, total loss is 11.68 , inference loss is 5.19, weight deacy loss is 6.49, training accuracy is 0.656250, time 232.691 samples/sec
+epoch 12, total_step 189260, total loss is 14.46 , inference loss is 7.97, weight deacy loss is 6.49, training accuracy is 0.531250, time 213.189 samples/sec
+epoch 12, total_step 189280, total loss is 13.29 , inference loss is 6.80, weight deacy loss is 6.49, training accuracy is 0.593750, time 211.016 samples/sec
+epoch 12, total_step 189300, total loss is 12.34 , inference loss is 5.85, weight deacy loss is 6.49, training accuracy is 0.500000, time 12.432 samples/sec
+epoch 12, total_step 189320, total loss is 13.43 , inference loss is 6.94, weight deacy loss is 6.49, training accuracy is 0.593750, time 223.091 samples/sec
+epoch 12, total_step 189340, total loss is 10.37 , inference loss is 3.88, weight deacy loss is 6.49, training accuracy is 0.593750, time 213.330 samples/sec
+epoch 12, total_step 189360, total loss is 14.92 , inference loss is 8.43, weight deacy loss is 6.49, training accuracy is 0.562500, time 235.767 samples/sec
+epoch 12, total_step 189380, total loss is 15.35 , inference loss is 8.86, weight deacy loss is 6.49, training accuracy is 0.375000, time 216.197 samples/sec
+epoch 12, total_step 189400, total loss is 9.97 , inference loss is 3.48, weight deacy loss is 6.49, training accuracy is 0.656250, time 211.884 samples/sec
+epoch 12, total_step 189420, total loss is 12.41 , inference loss is 5.92, weight deacy loss is 6.49, training accuracy is 0.531250, time 217.757 samples/sec
+epoch 12, total_step 189440, total loss is 11.66 , inference loss is 5.17, weight deacy loss is 6.49, training accuracy is 0.687500, time 223.320 samples/sec
+epoch 12, total_step 189460, total loss is 14.85 , inference loss is 8.36, weight deacy loss is 6.49, training accuracy is 0.531250, time 298.532 samples/sec
+epoch 12, total_step 189480, total loss is 11.89 , inference loss is 5.39, weight deacy loss is 6.49, training accuracy is 0.500000, time 299.472 samples/sec
+epoch 12, total_step 189500, total loss is 10.89 , inference loss is 4.40, weight deacy loss is 6.49, training accuracy is 0.625000, time 217.085 samples/sec
+epoch 12, total_step 189520, total loss is 14.36 , inference loss is 7.87, weight deacy loss is 6.49, training accuracy is 0.500000, time 213.347 samples/sec
+epoch 12, total_step 189540, total loss is 12.93 , inference loss is 6.44, weight deacy loss is 6.49, training accuracy is 0.312500, time 213.192 samples/sec
+epoch 12, total_step 189560, total loss is 12.07 , inference loss is 5.58, weight deacy loss is 6.49, training accuracy is 0.562500, time 214.087 samples/sec
+epoch 12, total_step 189580, total loss is 11.58 , inference loss is 5.09, weight deacy loss is 6.49, training accuracy is 0.593750, time 220.040 samples/sec
+epoch 12, total_step 189600, total loss is 12.96 , inference loss is 6.47, weight deacy loss is 6.49, training accuracy is 0.312500, time 12.291 samples/sec
+epoch 12, total_step 189620, total loss is 11.76 , inference loss is 5.27, weight deacy loss is 6.49, training accuracy is 0.656250, time 216.313 samples/sec
+epoch 12, total_step 189640, total loss is 12.73 , inference loss is 6.24, weight deacy loss is 6.49, training accuracy is 0.656250, time 236.326 samples/sec
+epoch 12, total_step 189660, total loss is 14.54 , inference loss is 8.05, weight deacy loss is 6.49, training accuracy is 0.437500, time 218.912 samples/sec
+epoch 12, total_step 189680, total loss is 12.10 , inference loss is 5.61, weight deacy loss is 6.49, training accuracy is 0.500000, time 216.907 samples/sec
+epoch 12, total_step 189700, total loss is 17.45 , inference loss is 10.96, weight deacy loss is 6.49, training accuracy is 0.437500, time 214.484 samples/sec
+epoch 12, total_step 189720, total loss is 12.86 , inference loss is 6.37, weight deacy loss is 6.49, training accuracy is 0.406250, time 220.963 samples/sec
+epoch 12, total_step 189740, total loss is 13.97 , inference loss is 7.48, weight deacy loss is 6.49, training accuracy is 0.406250, time 218.187 samples/sec
+epoch 12, total_step 189760, total loss is 13.11 , inference loss is 6.62, weight deacy loss is 6.49, training accuracy is 0.312500, time 213.206 samples/sec
+epoch 12, total_step 189780, total loss is 14.15 , inference loss is 7.66, weight deacy loss is 6.49, training accuracy is 0.531250, time 228.013 samples/sec
+epoch 12, total_step 189800, total loss is 16.01 , inference loss is 9.52, weight deacy loss is 6.49, training accuracy is 0.500000, time 232.290 samples/sec
+epoch 12, total_step 189820, total loss is 15.60 , inference loss is 9.11, weight deacy loss is 6.49, training accuracy is 0.500000, time 219.984 samples/sec
+epoch 12, total_step 189840, total loss is 14.11 , inference loss is 7.62, weight deacy loss is 6.49, training accuracy is 0.625000, time 214.009 samples/sec
+epoch 12, total_step 189860, total loss is 12.00 , inference loss is 5.52, weight deacy loss is 6.49, training accuracy is 0.562500, time 299.665 samples/sec
+epoch 12, total_step 189880, total loss is 10.90 , inference loss is 4.42, weight deacy loss is 6.49, training accuracy is 0.406250, time 208.680 samples/sec
+epoch 12, total_step 189900, total loss is 12.65 , inference loss is 6.16, weight deacy loss is 6.49, training accuracy is 0.437500, time 13.071 samples/sec
+epoch 12, total_step 189920, total loss is 14.69 , inference loss is 8.20, weight deacy loss is 6.49, training accuracy is 0.531250, time 214.052 samples/sec
+epoch 12, total_step 189940, total loss is 14.13 , inference loss is 7.64, weight deacy loss is 6.49, training accuracy is 0.531250, time 210.199 samples/sec
+epoch 12, total_step 189960, total loss is 12.81 , inference loss is 6.32, weight deacy loss is 6.49, training accuracy is 0.500000, time 219.224 samples/sec
+epoch 12, total_step 189980, total loss is 14.40 , inference loss is 7.91, weight deacy loss is 6.49, training accuracy is 0.437500, time 216.243 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.381966999999996
+best_threshold_index 118 0.9890740740740741
+best_threshold_index 118 0.9885185185185185
+best_threshold_index 118 0.9890740740740741
+best_threshold_index 119 0.9896296296296296
+best_threshold_index 118 0.9888888888888889
+best_threshold_index 123 0.9877777777777778
+best_threshold_index 118 0.9885185185185185
+best_threshold_index 118 0.9879629629629629
+best_threshold_index 118 0.9875925925925926
+best_threshold_index 118 0.9883333333333333
+[lfw][190000]XNorm: 21.369249
+[lfw][190000]Accuracy-Flip: 0.98783+-0.00517
+test accuracy is: 0.9878333333333333
+epoch 12, total_step 190000, total loss is 16.05 , inference loss is 9.56, weight deacy loss is 6.49, training accuracy is 0.375000, time 15.933 samples/sec
+epoch 12, total_step 190020, total loss is 14.23 , inference loss is 7.74, weight deacy loss is 6.49, training accuracy is 0.437500, time 221.956 samples/sec
+epoch 12, total_step 190040, total loss is 11.50 , inference loss is 5.01, weight deacy loss is 6.49, training accuracy is 0.500000, time 227.162 samples/sec
+epoch 12, total_step 190060, total loss is 10.64 , inference loss is 4.16, weight deacy loss is 6.49, training accuracy is 0.593750, time 214.140 samples/sec
+epoch 12, total_step 190080, total loss is 12.24 , inference loss is 5.75, weight deacy loss is 6.49, training accuracy is 0.718750, time 217.156 samples/sec
+epoch 12, total_step 190100, total loss is 12.54 , inference loss is 6.05, weight deacy loss is 6.49, training accuracy is 0.562500, time 218.882 samples/sec
+epoch 12, total_step 190120, total loss is 12.24 , inference loss is 5.75, weight deacy loss is 6.49, training accuracy is 0.593750, time 211.117 samples/sec
+epoch 12, total_step 190140, total loss is 12.43 , inference loss is 5.95, weight deacy loss is 6.49, training accuracy is 0.437500, time 214.859 samples/sec
+epoch 12, total_step 190160, total loss is 13.81 , inference loss is 7.32, weight deacy loss is 6.49, training accuracy is 0.468750, time 229.820 samples/sec
+epoch 12, total_step 190180, total loss is 16.02 , inference loss is 9.53, weight deacy loss is 6.49, training accuracy is 0.468750, time 297.974 samples/sec
+epoch 12, total_step 190200, total loss is 9.41 , inference loss is 2.92, weight deacy loss is 6.49, training accuracy is 0.593750, time 13.325 samples/sec
+epoch 12, total_step 190220, total loss is 17.08 , inference loss is 10.59, weight deacy loss is 6.49, training accuracy is 0.468750, time 216.295 samples/sec
+epoch 12, total_step 190240, total loss is 19.21 , inference loss is 12.72, weight deacy loss is 6.49, training accuracy is 0.312500, time 214.352 samples/sec
+epoch 12, total_step 190260, total loss is 13.35 , inference loss is 6.86, weight deacy loss is 6.49, training accuracy is 0.562500, time 225.112 samples/sec
+epoch 12, total_step 190280, total loss is 16.16 , inference loss is 9.67, weight deacy loss is 6.49, training accuracy is 0.593750, time 214.874 samples/sec
+epoch 12, total_step 190300, total loss is 11.41 , inference loss is 4.92, weight deacy loss is 6.48, training accuracy is 0.625000, time 216.360 samples/sec
+epoch 12, total_step 190320, total loss is 15.33 , inference loss is 8.84, weight deacy loss is 6.48, training accuracy is 0.531250, time 207.707 samples/sec
+epoch 12, total_step 190340, total loss is 12.15 , inference loss is 5.67, weight deacy loss is 6.48, training accuracy is 0.531250, time 211.333 samples/sec
+epoch 12, total_step 190360, total loss is 12.42 , inference loss is 5.93, weight deacy loss is 6.48, training accuracy is 0.593750, time 224.593 samples/sec
+epoch 12, total_step 190380, total loss is 11.80 , inference loss is 5.31, weight deacy loss is 6.48, training accuracy is 0.625000, time 213.872 samples/sec
+epoch 12, total_step 190400, total loss is 17.15 , inference loss is 10.67, weight deacy loss is 6.48, training accuracy is 0.375000, time 212.022 samples/sec
+epoch 12, total_step 190420, total loss is 14.97 , inference loss is 8.48, weight deacy loss is 6.48, training accuracy is 0.437500, time 220.129 samples/sec
+epoch 12, total_step 190440, total loss is 15.83 , inference loss is 9.35, weight deacy loss is 6.48, training accuracy is 0.437500, time 298.997 samples/sec
+epoch 12, total_step 190460, total loss is 14.15 , inference loss is 7.66, weight deacy loss is 6.48, training accuracy is 0.531250, time 236.683 samples/sec
+epoch 12, total_step 190480, total loss is 9.60 , inference loss is 3.12, weight deacy loss is 6.48, training accuracy is 0.593750, time 218.736 samples/sec
+epoch 12, total_step 190500, total loss is 14.66 , inference loss is 8.17, weight deacy loss is 6.48, training accuracy is 0.500000, time 12.919 samples/sec
+epoch 12, total_step 190520, total loss is 16.40 , inference loss is 9.92, weight deacy loss is 6.48, training accuracy is 0.500000, time 213.021 samples/sec
+epoch 12, total_step 190540, total loss is 13.03 , inference loss is 6.54, weight deacy loss is 6.48, training accuracy is 0.531250, time 218.797 samples/sec
+epoch 12, total_step 190560, total loss is 15.54 , inference loss is 9.06, weight deacy loss is 6.48, training accuracy is 0.468750, time 210.002 samples/sec
+epoch 12, total_step 190580, total loss is 12.76 , inference loss is 6.28, weight deacy loss is 6.48, training accuracy is 0.593750, time 219.472 samples/sec
+epoch 12, total_step 190600, total loss is 10.79 , inference loss is 4.30, weight deacy loss is 6.48, training accuracy is 0.500000, time 236.597 samples/sec
+epoch 12, total_step 190620, total loss is 15.30 , inference loss is 8.81, weight deacy loss is 6.48, training accuracy is 0.437500, time 212.653 samples/sec
+epoch 12, total_step 190640, total loss is 11.49 , inference loss is 5.00, weight deacy loss is 6.48, training accuracy is 0.593750, time 216.506 samples/sec
+epoch 12, total_step 190660, total loss is 13.79 , inference loss is 7.31, weight deacy loss is 6.48, training accuracy is 0.562500, time 216.387 samples/sec
+epoch 12, total_step 190680, total loss is 14.91 , inference loss is 8.42, weight deacy loss is 6.48, training accuracy is 0.375000, time 225.078 samples/sec
+epoch 12, total_step 190700, total loss is 14.37 , inference loss is 7.89, weight deacy loss is 6.48, training accuracy is 0.406250, time 298.201 samples/sec
+epoch 12, total_step 190720, total loss is 14.51 , inference loss is 8.03, weight deacy loss is 6.48, training accuracy is 0.468750, time 207.605 samples/sec
+epoch 12, total_step 190740, total loss is 10.83 , inference loss is 4.35, weight deacy loss is 6.48, training accuracy is 0.656250, time 217.395 samples/sec
+epoch 12, total_step 190760, total loss is 16.46 , inference loss is 9.98, weight deacy loss is 6.48, training accuracy is 0.468750, time 213.601 samples/sec
+epoch 12, total_step 190780, total loss is 11.81 , inference loss is 5.33, weight deacy loss is 6.48, training accuracy is 0.625000, time 217.542 samples/sec
+epoch 12, total_step 190800, total loss is 14.17 , inference loss is 7.68, weight deacy loss is 6.48, training accuracy is 0.343750, time 12.840 samples/sec
+epoch 12, total_step 190820, total loss is 13.97 , inference loss is 7.49, weight deacy loss is 6.48, training accuracy is 0.531250, time 222.903 samples/sec
+epoch 12, total_step 190840, total loss is 15.38 , inference loss is 8.90, weight deacy loss is 6.48, training accuracy is 0.437500, time 229.984 samples/sec
+epoch 12, total_step 190860, total loss is 13.88 , inference loss is 7.40, weight deacy loss is 6.48, training accuracy is 0.406250, time 220.067 samples/sec
+epoch 12, total_step 190880, total loss is 11.51 , inference loss is 5.03, weight deacy loss is 6.48, training accuracy is 0.500000, time 225.870 samples/sec
+epoch 12, total_step 190900, total loss is 12.90 , inference loss is 6.42, weight deacy loss is 6.48, training accuracy is 0.500000, time 218.603 samples/sec
+epoch 12, total_step 190920, total loss is 10.71 , inference loss is 4.23, weight deacy loss is 6.48, training accuracy is 0.593750, time 224.383 samples/sec
+epoch 12, total_step 190940, total loss is 11.50 , inference loss is 5.01, weight deacy loss is 6.48, training accuracy is 0.406250, time 221.205 samples/sec
+epoch 12, total_step 190960, total loss is 12.47 , inference loss is 5.99, weight deacy loss is 6.48, training accuracy is 0.531250, time 296.683 samples/sec
+epoch 12, total_step 190980, total loss is 15.65 , inference loss is 9.17, weight deacy loss is 6.48, training accuracy is 0.406250, time 224.066 samples/sec
+epoch 12, total_step 191000, total loss is 10.91 , inference loss is 4.43, weight deacy loss is 6.48, training accuracy is 0.437500, time 222.170 samples/sec
+epoch 12, total_step 191020, total loss is 14.46 , inference loss is 7.98, weight deacy loss is 6.48, training accuracy is 0.531250, time 217.812 samples/sec
+epoch 12, total_step 191040, total loss is 12.85 , inference loss is 6.37, weight deacy loss is 6.48, training accuracy is 0.531250, time 215.300 samples/sec
+epoch 12, total_step 191060, total loss is 12.32 , inference loss is 5.84, weight deacy loss is 6.48, training accuracy is 0.500000, time 217.470 samples/sec
+epoch 12, total_step 191080, total loss is 13.57 , inference loss is 7.09, weight deacy loss is 6.48, training accuracy is 0.437500, time 295.331 samples/sec
+epoch 12, total_step 191100, total loss is 15.38 , inference loss is 8.90, weight deacy loss is 6.48, training accuracy is 0.437500, time 13.102 samples/sec
+epoch 12, total_step 191120, total loss is 12.46 , inference loss is 5.98, weight deacy loss is 6.48, training accuracy is 0.500000, time 224.887 samples/sec
+epoch 12, total_step 191140, total loss is 14.97 , inference loss is 8.49, weight deacy loss is 6.48, training accuracy is 0.343750, time 217.126 samples/sec
+epoch 12, total_step 191160, total loss is 14.70 , inference loss is 8.22, weight deacy loss is 6.48, training accuracy is 0.375000, time 217.313 samples/sec
+epoch 12, total_step 191180, total loss is 13.40 , inference loss is 6.92, weight deacy loss is 6.48, training accuracy is 0.531250, time 215.773 samples/sec
+epoch 12, total_step 191200, total loss is 16.36 , inference loss is 9.88, weight deacy loss is 6.48, training accuracy is 0.343750, time 298.224 samples/sec
+epoch 12, total_step 191220, total loss is 16.18 , inference loss is 9.70, weight deacy loss is 6.48, training accuracy is 0.375000, time 295.796 samples/sec
+epoch 12, total_step 191240, total loss is 12.74 , inference loss is 6.26, weight deacy loss is 6.48, training accuracy is 0.437500, time 226.108 samples/sec
+epoch 12, total_step 191260, total loss is 12.93 , inference loss is 6.45, weight deacy loss is 6.48, training accuracy is 0.562500, time 215.275 samples/sec
+epoch 12, total_step 191280, total loss is 15.56 , inference loss is 9.09, weight deacy loss is 6.48, training accuracy is 0.375000, time 212.533 samples/sec
+epoch 12, total_step 191300, total loss is 13.10 , inference loss is 6.62, weight deacy loss is 6.48, training accuracy is 0.531250, time 236.102 samples/sec
+epoch 12, total_step 191320, total loss is 9.59 , inference loss is 3.12, weight deacy loss is 6.48, training accuracy is 0.593750, time 220.338 samples/sec
+epoch 12, total_step 191340, total loss is 13.44 , inference loss is 6.96, weight deacy loss is 6.48, training accuracy is 0.500000, time 223.968 samples/sec
+epoch 12, total_step 191360, total loss is 11.06 , inference loss is 4.59, weight deacy loss is 6.48, training accuracy is 0.531250, time 217.982 samples/sec
+epoch 12, total_step 191380, total loss is 15.50 , inference loss is 9.03, weight deacy loss is 6.48, training accuracy is 0.375000, time 218.064 samples/sec
+epoch 12, total_step 191400, total loss is 16.60 , inference loss is 10.13, weight deacy loss is 6.48, training accuracy is 0.468750, time 13.921 samples/sec
+epoch 12, total_step 191420, total loss is 15.59 , inference loss is 9.11, weight deacy loss is 6.48, training accuracy is 0.375000, time 213.804 samples/sec
+epoch 12, total_step 191440, total loss is 11.97 , inference loss is 5.49, weight deacy loss is 6.48, training accuracy is 0.406250, time 221.564 samples/sec
+epoch 12, total_step 191460, total loss is 13.23 , inference loss is 6.75, weight deacy loss is 6.48, training accuracy is 0.437500, time 217.759 samples/sec
+epoch 12, total_step 191480, total loss is 12.70 , inference loss is 6.23, weight deacy loss is 6.48, training accuracy is 0.500000, time 218.095 samples/sec
+epoch 12, total_step 191500, total loss is 13.72 , inference loss is 7.24, weight deacy loss is 6.48, training accuracy is 0.406250, time 211.396 samples/sec
+epoch 12, total_step 191520, total loss is 14.20 , inference loss is 7.73, weight deacy loss is 6.48, training accuracy is 0.468750, time 220.856 samples/sec
+epoch 12, total_step 191540, total loss is 13.59 , inference loss is 7.12, weight deacy loss is 6.48, training accuracy is 0.531250, time 221.271 samples/sec
+epoch 12, total_step 191560, total loss is 15.71 , inference loss is 9.23, weight deacy loss is 6.48, training accuracy is 0.500000, time 226.744 samples/sec
+epoch 12, total_step 191580, total loss is 10.66 , inference loss is 4.19, weight deacy loss is 6.48, training accuracy is 0.562500, time 235.704 samples/sec
+epoch 12, total_step 191600, total loss is 9.98 , inference loss is 3.50, weight deacy loss is 6.48, training accuracy is 0.625000, time 221.164 samples/sec
+epoch 12, total_step 191620, total loss is 13.00 , inference loss is 6.52, weight deacy loss is 6.48, training accuracy is 0.593750, time 216.040 samples/sec
+epoch 12, total_step 191640, total loss is 13.56 , inference loss is 7.09, weight deacy loss is 6.48, training accuracy is 0.406250, time 210.500 samples/sec
+epoch 12, total_step 191660, total loss is 15.54 , inference loss is 9.07, weight deacy loss is 6.48, training accuracy is 0.500000, time 224.820 samples/sec
+epoch 12, total_step 191680, total loss is 12.58 , inference loss is 6.11, weight deacy loss is 6.48, training accuracy is 0.437500, time 210.641 samples/sec
+epoch 12, total_step 191700, total loss is 14.03 , inference loss is 7.55, weight deacy loss is 6.48, training accuracy is 0.593750, time 13.970 samples/sec
+epoch 12, total_step 191720, total loss is 16.85 , inference loss is 10.37, weight deacy loss is 6.48, training accuracy is 0.531250, time 225.149 samples/sec
+epoch 12, total_step 191740, total loss is 15.36 , inference loss is 8.89, weight deacy loss is 6.48, training accuracy is 0.625000, time 223.845 samples/sec
+epoch 12, total_step 191760, total loss is 9.14 , inference loss is 2.67, weight deacy loss is 6.47, training accuracy is 0.656250, time 215.967 samples/sec
+epoch 12, total_step 191780, total loss is 11.71 , inference loss is 5.24, weight deacy loss is 6.47, training accuracy is 0.625000, time 213.265 samples/sec
+epoch 12, total_step 191800, total loss is 13.16 , inference loss is 6.69, weight deacy loss is 6.47, training accuracy is 0.437500, time 223.354 samples/sec
+epoch 12, total_step 191820, total loss is 13.18 , inference loss is 6.71, weight deacy loss is 6.47, training accuracy is 0.625000, time 216.351 samples/sec
+epoch 12, total_step 191840, total loss is 15.08 , inference loss is 8.60, weight deacy loss is 6.47, training accuracy is 0.437500, time 235.263 samples/sec
+epoch 12, total_step 191860, total loss is 12.57 , inference loss is 6.10, weight deacy loss is 6.47, training accuracy is 0.437500, time 205.518 samples/sec
+epoch 12, total_step 191880, total loss is 16.85 , inference loss is 10.37, weight deacy loss is 6.47, training accuracy is 0.437500, time 229.726 samples/sec
+epoch 12, total_step 191900, total loss is 13.05 , inference loss is 6.57, weight deacy loss is 6.47, training accuracy is 0.500000, time 223.051 samples/sec
+epoch 12, total_step 191920, total loss is 17.91 , inference loss is 11.43, weight deacy loss is 6.47, training accuracy is 0.281250, time 221.400 samples/sec
+epoch 12, total_step 191940, total loss is 12.79 , inference loss is 6.32, weight deacy loss is 6.47, training accuracy is 0.406250, time 212.684 samples/sec
+epoch 12, total_step 191960, total loss is 14.54 , inference loss is 8.07, weight deacy loss is 6.47, training accuracy is 0.343750, time 209.068 samples/sec
+epoch 12, total_step 191980, total loss is 14.03 , inference loss is 7.56, weight deacy loss is 6.47, training accuracy is 0.593750, time 295.483 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.476279999999974
+best_threshold_index 121 0.9881481481481481
+best_threshold_index 121 0.9875925925925926
+best_threshold_index 121 0.9875925925925926
+best_threshold_index 121 0.9885185185185185
+best_threshold_index 121 0.9883333333333333
+best_threshold_index 123 0.9872222222222222
+best_threshold_index 121 0.9883333333333333
+best_threshold_index 121 0.987037037037037
+best_threshold_index 121 0.987037037037037
+best_threshold_index 123 0.9872222222222222
+[lfw][192000]XNorm: 21.710559
+[lfw][192000]Accuracy-Flip: 0.98733+-0.00496
+test accuracy is: 0.9873333333333335
+epoch 12, total_step 192000, total loss is 10.13 , inference loss is 3.66, weight deacy loss is 6.47, training accuracy is 0.656250, time 15.287 samples/sec
+epoch 12, total_step 192020, total loss is 12.90 , inference loss is 6.43, weight deacy loss is 6.47, training accuracy is 0.531250, time 222.232 samples/sec
+epoch 12, total_step 192040, total loss is 12.77 , inference loss is 6.30, weight deacy loss is 6.47, training accuracy is 0.500000, time 299.367 samples/sec
+epoch 12, total_step 192060, total loss is 12.98 , inference loss is 6.51, weight deacy loss is 6.47, training accuracy is 0.468750, time 228.694 samples/sec
+epoch 12, total_step 192080, total loss is 16.48 , inference loss is 10.01, weight deacy loss is 6.47, training accuracy is 0.531250, time 217.244 samples/sec
+epoch 12, total_step 192100, total loss is 13.42 , inference loss is 6.95, weight deacy loss is 6.47, training accuracy is 0.406250, time 216.627 samples/sec
+epoch 12, total_step 192120, total loss is 17.70 , inference loss is 11.22, weight deacy loss is 6.47, training accuracy is 0.312500, time 220.790 samples/sec
+epoch 12, total_step 192140, total loss is 12.46 , inference loss is 5.99, weight deacy loss is 6.47, training accuracy is 0.500000, time 224.480 samples/sec
+epoch 12, total_step 192160, total loss is 14.46 , inference loss is 7.98, weight deacy loss is 6.47, training accuracy is 0.375000, time 229.211 samples/sec
+epoch 12, total_step 192180, total loss is 10.44 , inference loss is 3.97, weight deacy loss is 6.47, training accuracy is 0.468750, time 216.912 samples/sec
+epoch 12, total_step 192200, total loss is 14.36 , inference loss is 7.89, weight deacy loss is 6.47, training accuracy is 0.437500, time 221.535 samples/sec
+epoch 12, total_step 192220, total loss is 14.45 , inference loss is 7.98, weight deacy loss is 6.47, training accuracy is 0.562500, time 221.814 samples/sec
+epoch 12, total_step 192240, total loss is 19.32 , inference loss is 12.85, weight deacy loss is 6.47, training accuracy is 0.250000, time 218.492 samples/sec
+epoch 12, total_step 192260, total loss is 12.80 , inference loss is 6.33, weight deacy loss is 6.47, training accuracy is 0.531250, time 212.875 samples/sec
+epoch 12, total_step 192280, total loss is 16.20 , inference loss is 9.73, weight deacy loss is 6.47, training accuracy is 0.468750, time 223.842 samples/sec
+epoch 12, total_step 192300, total loss is 18.92 , inference loss is 12.45, weight deacy loss is 6.47, training accuracy is 0.375000, time 12.914 samples/sec
+epoch 12, total_step 192320, total loss is 14.38 , inference loss is 7.91, weight deacy loss is 6.47, training accuracy is 0.406250, time 220.806 samples/sec
+epoch 12, total_step 192340, total loss is 16.90 , inference loss is 10.43, weight deacy loss is 6.47, training accuracy is 0.468750, time 215.144 samples/sec
+epoch 12, total_step 192360, total loss is 15.31 , inference loss is 8.84, weight deacy loss is 6.47, training accuracy is 0.468750, time 219.525 samples/sec
+epoch 12, total_step 192380, total loss is 10.67 , inference loss is 4.20, weight deacy loss is 6.47, training accuracy is 0.625000, time 224.166 samples/sec
+epoch 12, total_step 192400, total loss is 19.92 , inference loss is 13.45, weight deacy loss is 6.47, training accuracy is 0.437500, time 210.737 samples/sec
+epoch 12, total_step 192420, total loss is 14.14 , inference loss is 7.67, weight deacy loss is 6.47, training accuracy is 0.343750, time 210.890 samples/sec
+epoch 12, total_step 192440, total loss is 14.36 , inference loss is 7.89, weight deacy loss is 6.47, training accuracy is 0.406250, time 226.038 samples/sec
+epoch 12, total_step 192460, total loss is 11.08 , inference loss is 4.61, weight deacy loss is 6.47, training accuracy is 0.531250, time 218.926 samples/sec
+epoch 12, total_step 192480, total loss is 11.61 , inference loss is 5.14, weight deacy loss is 6.47, training accuracy is 0.500000, time 220.432 samples/sec
+epoch 12, total_step 192500, total loss is 15.72 , inference loss is 9.25, weight deacy loss is 6.47, training accuracy is 0.375000, time 220.756 samples/sec
+epoch 12, total_step 192520, total loss is 10.43 , inference loss is 3.96, weight deacy loss is 6.47, training accuracy is 0.562500, time 223.967 samples/sec
+epoch 12, total_step 192540, total loss is 14.79 , inference loss is 8.32, weight deacy loss is 6.47, training accuracy is 0.531250, time 224.330 samples/sec
+epoch 12, total_step 192560, total loss is 16.02 , inference loss is 9.55, weight deacy loss is 6.47, training accuracy is 0.437500, time 220.884 samples/sec
+epoch 12, total_step 192580, total loss is 12.96 , inference loss is 6.49, weight deacy loss is 6.47, training accuracy is 0.593750, time 218.210 samples/sec
+epoch 12, total_step 192600, total loss is 11.38 , inference loss is 4.91, weight deacy loss is 6.47, training accuracy is 0.468750, time 12.893 samples/sec
+epoch 12, total_step 192620, total loss is 13.26 , inference loss is 6.79, weight deacy loss is 6.47, training accuracy is 0.437500, time 218.497 samples/sec
+epoch 12, total_step 192640, total loss is 16.59 , inference loss is 10.12, weight deacy loss is 6.47, training accuracy is 0.437500, time 215.868 samples/sec
+epoch 12, total_step 192660, total loss is 10.01 , inference loss is 3.54, weight deacy loss is 6.47, training accuracy is 0.562500, time 223.211 samples/sec
+epoch 12, total_step 192680, total loss is 15.08 , inference loss is 8.61, weight deacy loss is 6.47, training accuracy is 0.468750, time 224.983 samples/sec
+epoch 12, total_step 192700, total loss is 14.14 , inference loss is 7.67, weight deacy loss is 6.47, training accuracy is 0.406250, time 215.359 samples/sec
+epoch 12, total_step 192720, total loss is 13.21 , inference loss is 6.74, weight deacy loss is 6.47, training accuracy is 0.437500, time 232.469 samples/sec
+epoch 12, total_step 192740, total loss is 13.71 , inference loss is 7.24, weight deacy loss is 6.47, training accuracy is 0.343750, time 232.270 samples/sec
+epoch 12, total_step 192760, total loss is 14.11 , inference loss is 7.64, weight deacy loss is 6.47, training accuracy is 0.406250, time 210.921 samples/sec
+epoch 12, total_step 192780, total loss is 14.09 , inference loss is 7.62, weight deacy loss is 6.47, training accuracy is 0.437500, time 219.177 samples/sec
+epoch 12, total_step 192800, total loss is 12.49 , inference loss is 6.02, weight deacy loss is 6.47, training accuracy is 0.406250, time 229.975 samples/sec
+epoch 12, total_step 192820, total loss is 14.09 , inference loss is 7.62, weight deacy loss is 6.47, training accuracy is 0.406250, time 214.435 samples/sec
+epoch 12, total_step 192840, total loss is 17.32 , inference loss is 10.85, weight deacy loss is 6.47, training accuracy is 0.375000, time 217.777 samples/sec
+epoch 12, total_step 192860, total loss is 13.25 , inference loss is 6.78, weight deacy loss is 6.47, training accuracy is 0.468750, time 214.540 samples/sec
+epoch 12, total_step 192880, total loss is 13.68 , inference loss is 7.21, weight deacy loss is 6.47, training accuracy is 0.468750, time 296.584 samples/sec
+epoch 12, total_step 192900, total loss is 12.97 , inference loss is 6.50, weight deacy loss is 6.47, training accuracy is 0.625000, time 12.781 samples/sec
+epoch 12, total_step 192920, total loss is 11.20 , inference loss is 4.73, weight deacy loss is 6.47, training accuracy is 0.531250, time 212.973 samples/sec
+epoch 12, total_step 192940, total loss is 17.32 , inference loss is 10.85, weight deacy loss is 6.47, training accuracy is 0.375000, time 215.134 samples/sec
+epoch 12, total_step 192960, total loss is 14.13 , inference loss is 7.66, weight deacy loss is 6.47, training accuracy is 0.500000, time 215.592 samples/sec
+epoch 12, total_step 192980, total loss is 14.54 , inference loss is 8.07, weight deacy loss is 6.47, training accuracy is 0.531250, time 220.219 samples/sec
+epoch 12, total_step 193000, total loss is 15.29 , inference loss is 8.82, weight deacy loss is 6.47, training accuracy is 0.500000, time 295.947 samples/sec
+epoch 12, total_step 193020, total loss is 15.30 , inference loss is 8.83, weight deacy loss is 6.47, training accuracy is 0.343750, time 297.674 samples/sec
+epoch 12, total_step 193040, total loss is 15.13 , inference loss is 8.66, weight deacy loss is 6.47, training accuracy is 0.437500, time 212.238 samples/sec
+epoch 12, total_step 193060, total loss is 11.46 , inference loss is 4.99, weight deacy loss is 6.47, training accuracy is 0.468750, time 211.381 samples/sec
+epoch 12, total_step 193080, total loss is 13.76 , inference loss is 7.29, weight deacy loss is 6.47, training accuracy is 0.468750, time 215.635 samples/sec
+epoch 12, total_step 193100, total loss is 12.56 , inference loss is 6.09, weight deacy loss is 6.47, training accuracy is 0.531250, time 217.894 samples/sec
+epoch 12, total_step 193120, total loss is 11.52 , inference loss is 5.05, weight deacy loss is 6.47, training accuracy is 0.562500, time 214.017 samples/sec
+epoch 12, total_step 193140, total loss is 14.39 , inference loss is 7.92, weight deacy loss is 6.47, training accuracy is 0.531250, time 219.855 samples/sec
+epoch 12, total_step 193160, total loss is 11.46 , inference loss is 4.99, weight deacy loss is 6.47, training accuracy is 0.531250, time 218.992 samples/sec
+epoch 12, total_step 193180, total loss is 10.45 , inference loss is 3.98, weight deacy loss is 6.47, training accuracy is 0.625000, time 212.577 samples/sec
+epoch 12, total_step 193200, total loss is 12.97 , inference loss is 6.50, weight deacy loss is 6.47, training accuracy is 0.531250, time 12.571 samples/sec
+epoch 12, total_step 193220, total loss is 13.17 , inference loss is 6.71, weight deacy loss is 6.47, training accuracy is 0.593750, time 221.521 samples/sec
+epoch 12, total_step 193240, total loss is 13.47 , inference loss is 7.00, weight deacy loss is 6.47, training accuracy is 0.468750, time 233.955 samples/sec
+epoch 12, total_step 193260, total loss is 16.14 , inference loss is 9.67, weight deacy loss is 6.47, training accuracy is 0.468750, time 211.489 samples/sec
+epoch 12, total_step 193280, total loss is 11.63 , inference loss is 5.16, weight deacy loss is 6.47, training accuracy is 0.531250, time 297.260 samples/sec
+epoch 12, total_step 193300, total loss is 11.70 , inference loss is 5.23, weight deacy loss is 6.47, training accuracy is 0.531250, time 216.035 samples/sec
+epoch 12, total_step 193320, total loss is 10.77 , inference loss is 4.30, weight deacy loss is 6.47, training accuracy is 0.656250, time 211.712 samples/sec
+epoch 12, total_step 193340, total loss is 13.92 , inference loss is 7.45, weight deacy loss is 6.47, training accuracy is 0.562500, time 227.588 samples/sec
+epoch 12, total_step 193360, total loss is 12.04 , inference loss is 5.58, weight deacy loss is 6.47, training accuracy is 0.531250, time 209.875 samples/sec
+epoch 12, total_step 193380, total loss is 15.95 , inference loss is 9.48, weight deacy loss is 6.47, training accuracy is 0.375000, time 223.884 samples/sec
+epoch 12, total_step 193400, total loss is 14.08 , inference loss is 7.61, weight deacy loss is 6.47, training accuracy is 0.406250, time 220.865 samples/sec
+epoch 12, total_step 193420, total loss is 15.55 , inference loss is 9.09, weight deacy loss is 6.47, training accuracy is 0.312500, time 218.974 samples/sec
+epoch 12, total_step 193440, total loss is 13.30 , inference loss is 6.84, weight deacy loss is 6.47, training accuracy is 0.250000, time 213.766 samples/sec
+epoch 12, total_step 193460, total loss is 14.96 , inference loss is 8.50, weight deacy loss is 6.47, training accuracy is 0.406250, time 221.913 samples/sec
+epoch 12, total_step 193480, total loss is 14.93 , inference loss is 8.46, weight deacy loss is 6.47, training accuracy is 0.406250, time 215.164 samples/sec
+epoch 12, total_step 193500, total loss is 14.95 , inference loss is 8.48, weight deacy loss is 6.47, training accuracy is 0.375000, time 13.207 samples/sec
+epoch 12, total_step 193520, total loss is 13.23 , inference loss is 6.76, weight deacy loss is 6.47, training accuracy is 0.468750, time 221.156 samples/sec
+epoch 12, total_step 193540, total loss is 17.51 , inference loss is 11.04, weight deacy loss is 6.47, training accuracy is 0.406250, time 297.540 samples/sec
+epoch 12, total_step 193560, total loss is 12.10 , inference loss is 5.63, weight deacy loss is 6.47, training accuracy is 0.500000, time 218.663 samples/sec
+epoch 12, total_step 193580, total loss is 14.67 , inference loss is 8.20, weight deacy loss is 6.47, training accuracy is 0.656250, time 212.132 samples/sec
+epoch 12, total_step 193600, total loss is 11.40 , inference loss is 4.94, weight deacy loss is 6.47, training accuracy is 0.625000, time 216.280 samples/sec
+epoch 12, total_step 193620, total loss is 15.00 , inference loss is 8.53, weight deacy loss is 6.47, training accuracy is 0.468750, time 214.633 samples/sec
+epoch 12, total_step 193640, total loss is 13.09 , inference loss is 6.62, weight deacy loss is 6.47, training accuracy is 0.468750, time 219.708 samples/sec
+epoch 12, total_step 193660, total loss is 12.51 , inference loss is 6.04, weight deacy loss is 6.47, training accuracy is 0.437500, time 221.304 samples/sec
+epoch 12, total_step 193680, total loss is 13.09 , inference loss is 6.62, weight deacy loss is 6.47, training accuracy is 0.468750, time 213.484 samples/sec
+epoch 12, total_step 193700, total loss is 10.15 , inference loss is 3.69, weight deacy loss is 6.47, training accuracy is 0.656250, time 223.462 samples/sec
+epoch 12, total_step 193720, total loss is 18.14 , inference loss is 11.67, weight deacy loss is 6.47, training accuracy is 0.312500, time 218.077 samples/sec
+epoch 12, total_step 193740, total loss is 14.06 , inference loss is 7.59, weight deacy loss is 6.47, training accuracy is 0.562500, time 220.917 samples/sec
+epoch 12, total_step 193760, total loss is 14.69 , inference loss is 8.23, weight deacy loss is 6.47, training accuracy is 0.312500, time 220.604 samples/sec
+epoch 12, total_step 193780, total loss is 14.02 , inference loss is 7.56, weight deacy loss is 6.46, training accuracy is 0.468750, time 220.072 samples/sec
+epoch 12, total_step 193800, total loss is 10.39 , inference loss is 3.92, weight deacy loss is 6.46, training accuracy is 0.718750, time 12.051 samples/sec
+epoch 12, total_step 193820, total loss is 11.36 , inference loss is 4.89, weight deacy loss is 6.46, training accuracy is 0.468750, time 218.348 samples/sec
+epoch 12, total_step 193840, total loss is 14.70 , inference loss is 8.24, weight deacy loss is 6.46, training accuracy is 0.468750, time 216.653 samples/sec
+epoch 12, total_step 193860, total loss is 12.23 , inference loss is 5.77, weight deacy loss is 6.46, training accuracy is 0.468750, time 216.601 samples/sec
+epoch 12, total_step 193880, total loss is 15.91 , inference loss is 9.45, weight deacy loss is 6.46, training accuracy is 0.437500, time 229.523 samples/sec
+epoch 12, total_step 193900, total loss is 15.68 , inference loss is 9.22, weight deacy loss is 6.46, training accuracy is 0.531250, time 214.854 samples/sec
+epoch 12, total_step 193920, total loss is 11.55 , inference loss is 5.09, weight deacy loss is 6.46, training accuracy is 0.437500, time 217.171 samples/sec
+epoch 12, total_step 193940, total loss is 14.56 , inference loss is 8.10, weight deacy loss is 6.46, training accuracy is 0.406250, time 218.226 samples/sec
+epoch 12, total_step 193960, total loss is 16.53 , inference loss is 10.07, weight deacy loss is 6.46, training accuracy is 0.437500, time 220.752 samples/sec
+epoch 12, total_step 193980, total loss is 13.27 , inference loss is 6.81, weight deacy loss is 6.46, training accuracy is 0.562500, time 212.662 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.779674999999997
+best_threshold_index 107 0.9866666666666667
+best_threshold_index 109 0.9859259259259259
+best_threshold_index 107 0.9855555555555555
+best_threshold_index 107 0.9864814814814815
+best_threshold_index 107 0.9855555555555555
+best_threshold_index 107 0.9851851851851852
+best_threshold_index 109 0.9859259259259259
+best_threshold_index 109 0.9853703703703703
+best_threshold_index 107 0.9846296296296296
+best_threshold_index 107 0.9844444444444445
+[lfw][194000]XNorm: 20.685266
+[lfw][194000]Accuracy-Flip: 0.98483+-0.00612
+test accuracy is: 0.9848333333333332
+epoch 12, total_step 194000, total loss is 12.87 , inference loss is 6.41, weight deacy loss is 6.46, training accuracy is 0.468750, time 16.191 samples/sec
+epoch 12, total_step 194020, total loss is 14.15 , inference loss is 7.69, weight deacy loss is 6.46, training accuracy is 0.500000, time 214.149 samples/sec
+epoch 12, total_step 194040, total loss is 12.82 , inference loss is 6.35, weight deacy loss is 6.46, training accuracy is 0.468750, time 227.040 samples/sec
+epoch 12, total_step 194060, total loss is 13.38 , inference loss is 6.92, weight deacy loss is 6.46, training accuracy is 0.468750, time 216.950 samples/sec
+epoch 12, total_step 194080, total loss is 16.45 , inference loss is 9.99, weight deacy loss is 6.46, training accuracy is 0.468750, time 220.497 samples/sec
+epoch 12, total_step 194100, total loss is 15.90 , inference loss is 9.43, weight deacy loss is 6.46, training accuracy is 0.437500, time 12.860 samples/sec
+epoch 12, total_step 194120, total loss is 12.96 , inference loss is 6.50, weight deacy loss is 6.46, training accuracy is 0.406250, time 211.040 samples/sec
+epoch 12, total_step 194140, total loss is 12.56 , inference loss is 6.10, weight deacy loss is 6.46, training accuracy is 0.406250, time 216.227 samples/sec
+epoch 12, total_step 194160, total loss is 14.64 , inference loss is 8.18, weight deacy loss is 6.46, training accuracy is 0.375000, time 297.203 samples/sec
+epoch 12, total_step 194180, total loss is 18.47 , inference loss is 12.01, weight deacy loss is 6.46, training accuracy is 0.312500, time 219.677 samples/sec
+epoch 12, total_step 194200, total loss is 13.17 , inference loss is 6.70, weight deacy loss is 6.46, training accuracy is 0.500000, time 216.046 samples/sec
+epoch 12, total_step 194220, total loss is 15.86 , inference loss is 9.40, weight deacy loss is 6.46, training accuracy is 0.437500, time 211.265 samples/sec
+epoch 12, total_step 194240, total loss is 16.28 , inference loss is 9.82, weight deacy loss is 6.46, training accuracy is 0.375000, time 238.617 samples/sec
+epoch 12, total_step 194260, total loss is 11.19 , inference loss is 4.73, weight deacy loss is 6.46, training accuracy is 0.625000, time 222.191 samples/sec
+epoch 12, total_step 194280, total loss is 14.26 , inference loss is 7.80, weight deacy loss is 6.46, training accuracy is 0.625000, time 222.588 samples/sec
+epoch 12, total_step 194300, total loss is 16.05 , inference loss is 9.59, weight deacy loss is 6.46, training accuracy is 0.468750, time 215.960 samples/sec
+epoch 12, total_step 194320, total loss is 12.22 , inference loss is 5.76, weight deacy loss is 6.46, training accuracy is 0.562500, time 221.177 samples/sec
+epoch 12, total_step 194340, total loss is 14.55 , inference loss is 8.09, weight deacy loss is 6.46, training accuracy is 0.625000, time 216.254 samples/sec
+epoch 12, total_step 194360, total loss is 13.29 , inference loss is 6.83, weight deacy loss is 6.46, training accuracy is 0.500000, time 223.071 samples/sec
+epoch 12, total_step 194380, total loss is 16.91 , inference loss is 10.45, weight deacy loss is 6.46, training accuracy is 0.437500, time 216.307 samples/sec
+epoch 12, total_step 194400, total loss is 12.61 , inference loss is 6.15, weight deacy loss is 6.46, training accuracy is 0.468750, time 12.626 samples/sec
+epoch 12, total_step 194420, total loss is 14.44 , inference loss is 7.98, weight deacy loss is 6.46, training accuracy is 0.593750, time 298.589 samples/sec
+epoch 12, total_step 194440, total loss is 11.04 , inference loss is 4.58, weight deacy loss is 6.46, training accuracy is 0.625000, time 214.637 samples/sec
+epoch 12, total_step 194460, total loss is 14.33 , inference loss is 7.87, weight deacy loss is 6.46, training accuracy is 0.375000, time 221.714 samples/sec
+epoch 12, total_step 194480, total loss is 12.77 , inference loss is 6.31, weight deacy loss is 6.46, training accuracy is 0.468750, time 217.635 samples/sec
+epoch 12, total_step 194500, total loss is 10.99 , inference loss is 4.53, weight deacy loss is 6.46, training accuracy is 0.531250, time 214.693 samples/sec
+epoch 12, total_step 194520, total loss is 11.52 , inference loss is 5.06, weight deacy loss is 6.46, training accuracy is 0.625000, time 239.463 samples/sec
+epoch 12, total_step 194540, total loss is 13.74 , inference loss is 7.28, weight deacy loss is 6.46, training accuracy is 0.562500, time 231.494 samples/sec
+epoch 12, total_step 194560, total loss is 14.16 , inference loss is 7.70, weight deacy loss is 6.46, training accuracy is 0.500000, time 213.425 samples/sec
+epoch 12, total_step 194580, total loss is 16.78 , inference loss is 10.33, weight deacy loss is 6.46, training accuracy is 0.343750, time 216.030 samples/sec
+epoch 12, total_step 194600, total loss is 14.12 , inference loss is 7.66, weight deacy loss is 6.46, training accuracy is 0.468750, time 221.310 samples/sec
+epoch 12, total_step 194620, total loss is 14.48 , inference loss is 8.02, weight deacy loss is 6.46, training accuracy is 0.437500, time 226.608 samples/sec
+epoch 12, total_step 194640, total loss is 16.41 , inference loss is 9.96, weight deacy loss is 6.46, training accuracy is 0.562500, time 215.881 samples/sec
+epoch 12, total_step 194660, total loss is 11.87 , inference loss is 5.41, weight deacy loss is 6.46, training accuracy is 0.437500, time 213.491 samples/sec
+epoch 12, total_step 194680, total loss is 12.85 , inference loss is 6.39, weight deacy loss is 6.46, training accuracy is 0.500000, time 211.924 samples/sec
+epoch 12, total_step 194700, total loss is 12.91 , inference loss is 6.45, weight deacy loss is 6.46, training accuracy is 0.531250, time 12.582 samples/sec
+epoch 12, total_step 194720, total loss is 15.40 , inference loss is 8.94, weight deacy loss is 6.46, training accuracy is 0.500000, time 222.584 samples/sec
+epoch 12, total_step 194740, total loss is 11.79 , inference loss is 5.33, weight deacy loss is 6.46, training accuracy is 0.562500, time 297.997 samples/sec
+epoch 12, total_step 194760, total loss is 13.17 , inference loss is 6.72, weight deacy loss is 6.46, training accuracy is 0.468750, time 199.418 samples/sec
+epoch 12, total_step 194780, total loss is 10.97 , inference loss is 4.51, weight deacy loss is 6.46, training accuracy is 0.593750, time 209.830 samples/sec
+epoch 12, total_step 194800, total loss is 12.79 , inference loss is 6.34, weight deacy loss is 6.46, training accuracy is 0.406250, time 224.917 samples/sec
+epoch 12, total_step 194820, total loss is 15.58 , inference loss is 9.12, weight deacy loss is 6.46, training accuracy is 0.437500, time 215.455 samples/sec
+epoch 12, total_step 194840, total loss is 13.39 , inference loss is 6.93, weight deacy loss is 6.46, training accuracy is 0.468750, time 221.435 samples/sec
+epoch 12, total_step 194860, total loss is 12.67 , inference loss is 6.21, weight deacy loss is 6.46, training accuracy is 0.375000, time 298.008 samples/sec
+epoch 12, total_step 194880, total loss is 10.39 , inference loss is 3.94, weight deacy loss is 6.46, training accuracy is 0.656250, time 211.229 samples/sec
+epoch 12, total_step 194900, total loss is 10.99 , inference loss is 4.53, weight deacy loss is 6.46, training accuracy is 0.437500, time 206.536 samples/sec
+epoch 12, total_step 194920, total loss is 13.92 , inference loss is 7.46, weight deacy loss is 6.46, training accuracy is 0.468750, time 217.819 samples/sec
+epoch 12, total_step 194940, total loss is 14.74 , inference loss is 8.28, weight deacy loss is 6.46, training accuracy is 0.406250, time 216.714 samples/sec
+epoch 12, total_step 194960, total loss is 14.39 , inference loss is 7.94, weight deacy loss is 6.46, training accuracy is 0.312500, time 225.227 samples/sec
+epoch 12, total_step 194980, total loss is 15.21 , inference loss is 8.75, weight deacy loss is 6.46, training accuracy is 0.375000, time 220.016 samples/sec
+epoch 12, total_step 195000, total loss is 16.63 , inference loss is 10.17, weight deacy loss is 6.46, training accuracy is 0.406250, time 13.274 samples/sec
+epoch 12, total_step 195020, total loss is 14.16 , inference loss is 7.70, weight deacy loss is 6.46, training accuracy is 0.531250, time 221.400 samples/sec
+epoch 12, total_step 195040, total loss is 12.29 , inference loss is 5.83, weight deacy loss is 6.46, training accuracy is 0.468750, time 219.257 samples/sec
+epoch 12, total_step 195060, total loss is 14.14 , inference loss is 7.69, weight deacy loss is 6.46, training accuracy is 0.406250, time 221.166 samples/sec
+epoch 12, total_step 195080, total loss is 13.29 , inference loss is 6.84, weight deacy loss is 6.46, training accuracy is 0.531250, time 220.488 samples/sec
+epoch 12, total_step 195100, total loss is 13.62 , inference loss is 7.16, weight deacy loss is 6.46, training accuracy is 0.468750, time 232.674 samples/sec
+epoch 12, total_step 195120, total loss is 11.57 , inference loss is 5.11, weight deacy loss is 6.46, training accuracy is 0.500000, time 186.483 samples/sec
+epoch 12, total_step 195140, total loss is 18.28 , inference loss is 11.83, weight deacy loss is 6.46, training accuracy is 0.406250, time 204.314 samples/sec
+epoch 12, total_step 195160, total loss is 14.58 , inference loss is 8.12, weight deacy loss is 6.46, training accuracy is 0.593750, time 213.168 samples/sec
+epoch 12, total_step 195180, total loss is 16.67 , inference loss is 10.22, weight deacy loss is 6.46, training accuracy is 0.312500, time 223.224 samples/sec
+epoch 12, total_step 195200, total loss is 13.23 , inference loss is 6.78, weight deacy loss is 6.46, training accuracy is 0.562500, time 213.417 samples/sec
+epoch 12, total_step 195220, total loss is 14.59 , inference loss is 8.14, weight deacy loss is 6.46, training accuracy is 0.468750, time 213.004 samples/sec
+epoch 12, total_step 195240, total loss is 13.54 , inference loss is 7.09, weight deacy loss is 6.46, training accuracy is 0.437500, time 223.043 samples/sec
+epoch 12, total_step 195260, total loss is 9.69 , inference loss is 3.24, weight deacy loss is 6.45, training accuracy is 0.593750, time 218.937 samples/sec
+epoch 12, total_step 195280, total loss is 13.90 , inference loss is 7.44, weight deacy loss is 6.45, training accuracy is 0.531250, time 218.562 samples/sec
+epoch 12, total_step 195300, total loss is 14.42 , inference loss is 7.97, weight deacy loss is 6.45, training accuracy is 0.437500, time 12.457 samples/sec
+epoch 12, total_step 195320, total loss is 13.48 , inference loss is 7.03, weight deacy loss is 6.45, training accuracy is 0.531250, time 219.205 samples/sec
+epoch 12, total_step 195340, total loss is 16.94 , inference loss is 10.48, weight deacy loss is 6.45, training accuracy is 0.468750, time 207.850 samples/sec
+epoch 12, total_step 195360, total loss is 10.67 , inference loss is 4.22, weight deacy loss is 6.45, training accuracy is 0.468750, time 216.038 samples/sec
+epoch 12, total_step 195380, total loss is 12.78 , inference loss is 6.32, weight deacy loss is 6.45, training accuracy is 0.562500, time 219.277 samples/sec
+epoch 12, total_step 195400, total loss is 10.11 , inference loss is 3.65, weight deacy loss is 6.45, training accuracy is 0.468750, time 298.362 samples/sec
+epoch 12, total_step 195420, total loss is 13.51 , inference loss is 7.06, weight deacy loss is 6.45, training accuracy is 0.406250, time 218.221 samples/sec
+epoch 12, total_step 195440, total loss is 16.62 , inference loss is 10.16, weight deacy loss is 6.45, training accuracy is 0.562500, time 224.672 samples/sec
+epoch 12, total_step 195460, total loss is 11.99 , inference loss is 5.54, weight deacy loss is 6.45, training accuracy is 0.437500, time 222.038 samples/sec
+epoch 12, total_step 195480, total loss is 11.49 , inference loss is 5.03, weight deacy loss is 6.45, training accuracy is 0.531250, time 221.612 samples/sec
+epoch 12, total_step 195500, total loss is 14.36 , inference loss is 7.90, weight deacy loss is 6.45, training accuracy is 0.437500, time 231.417 samples/sec
+epoch 12, total_step 195520, total loss is 11.03 , inference loss is 4.57, weight deacy loss is 6.45, training accuracy is 0.687500, time 220.378 samples/sec
+epoch 12, total_step 195540, total loss is 14.62 , inference loss is 8.16, weight deacy loss is 6.45, training accuracy is 0.437500, time 227.972 samples/sec
+epoch 12, total_step 195560, total loss is 20.33 , inference loss is 13.87, weight deacy loss is 6.45, training accuracy is 0.343750, time 238.961 samples/sec
+epoch 12, total_step 195580, total loss is 12.89 , inference loss is 6.44, weight deacy loss is 6.45, training accuracy is 0.562500, time 219.338 samples/sec
+epoch 12, total_step 195600, total loss is 15.29 , inference loss is 8.84, weight deacy loss is 6.45, training accuracy is 0.468750, time 13.365 samples/sec
+epoch 12, total_step 195620, total loss is 16.55 , inference loss is 10.10, weight deacy loss is 6.45, training accuracy is 0.406250, time 214.887 samples/sec
+epoch 12, total_step 195640, total loss is 10.31 , inference loss is 3.86, weight deacy loss is 6.45, training accuracy is 0.531250, time 241.745 samples/sec
+epoch 12, total_step 195660, total loss is 12.55 , inference loss is 6.10, weight deacy loss is 6.45, training accuracy is 0.406250, time 296.275 samples/sec
+epoch 12, total_step 195680, total loss is 14.79 , inference loss is 8.33, weight deacy loss is 6.45, training accuracy is 0.468750, time 214.770 samples/sec
+epoch 12, total_step 195700, total loss is 15.79 , inference loss is 9.34, weight deacy loss is 6.45, training accuracy is 0.437500, time 212.972 samples/sec
+epoch 12, total_step 195720, total loss is 13.53 , inference loss is 7.08, weight deacy loss is 6.45, training accuracy is 0.468750, time 216.969 samples/sec
+epoch 12, total_step 195740, total loss is 12.54 , inference loss is 6.09, weight deacy loss is 6.45, training accuracy is 0.468750, time 217.337 samples/sec
+epoch 12, total_step 195760, total loss is 15.48 , inference loss is 9.02, weight deacy loss is 6.45, training accuracy is 0.250000, time 215.398 samples/sec
+epoch 12, total_step 195780, total loss is 16.25 , inference loss is 9.80, weight deacy loss is 6.45, training accuracy is 0.531250, time 209.827 samples/sec
+epoch 12, total_step 195800, total loss is 12.88 , inference loss is 6.42, weight deacy loss is 6.45, training accuracy is 0.281250, time 215.328 samples/sec
+epoch 12, total_step 195820, total loss is 13.89 , inference loss is 7.44, weight deacy loss is 6.45, training accuracy is 0.281250, time 215.295 samples/sec
+epoch 12, total_step 195840, total loss is 16.82 , inference loss is 10.37, weight deacy loss is 6.45, training accuracy is 0.437500, time 212.781 samples/sec
+epoch 12, total_step 195860, total loss is 13.61 , inference loss is 7.16, weight deacy loss is 6.45, training accuracy is 0.437500, time 215.780 samples/sec
+epoch 12, total_step 195880, total loss is 16.46 , inference loss is 10.01, weight deacy loss is 6.45, training accuracy is 0.343750, time 219.706 samples/sec
+epoch 12, total_step 195900, total loss is 15.38 , inference loss is 8.93, weight deacy loss is 6.45, training accuracy is 0.468750, time 12.336 samples/sec
+epoch 12, total_step 195920, total loss is 13.98 , inference loss is 7.53, weight deacy loss is 6.45, training accuracy is 0.468750, time 299.026 samples/sec
+epoch 12, total_step 195940, total loss is 14.44 , inference loss is 7.99, weight deacy loss is 6.45, training accuracy is 0.406250, time 215.494 samples/sec
+epoch 12, total_step 195960, total loss is 16.63 , inference loss is 10.18, weight deacy loss is 6.45, training accuracy is 0.375000, time 213.091 samples/sec
+epoch 12, total_step 195980, total loss is 11.71 , inference loss is 5.26, weight deacy loss is 6.45, training accuracy is 0.500000, time 210.804 samples/sec
+testing verification..
+(12000, 512)
+infer time 26.41791999999998
+best_threshold_index 113 0.9855555555555555
+best_threshold_index 108 0.9848148148148148
+best_threshold_index 109 0.9853703703703703
+best_threshold_index 113 0.987037037037037
+best_threshold_index 109 0.9851851851851852
+best_threshold_index 113 0.9851851851851852
+best_threshold_index 109 0.9857407407407407
+best_threshold_index 106 0.9848148148148148
+best_threshold_index 109 0.9835185185185186
+best_threshold_index 113 0.9846296296296296
+[lfw][196000]XNorm: 22.314521
+[lfw][196000]Accuracy-Flip: 0.98283+-0.00753
+test accuracy is: 0.9828333333333333
+epoch 12, total_step 196000, total loss is 16.71 , inference loss is 10.26, weight deacy loss is 6.45, training accuracy is 0.281250, time 15.874 samples/sec
+epoch 12, total_step 196020, total loss is 13.93 , inference loss is 7.48, weight deacy loss is 6.45, training accuracy is 0.531250, time 214.533 samples/sec
+epoch 12, total_step 196040, total loss is 15.53 , inference loss is 9.08, weight deacy loss is 6.45, training accuracy is 0.437500, time 221.005 samples/sec
+epoch 12, total_step 196060, total loss is 11.28 , inference loss is 4.83, weight deacy loss is 6.45, training accuracy is 0.531250, time 219.379 samples/sec
+epoch 12, total_step 196080, total loss is 13.04 , inference loss is 6.59, weight deacy loss is 6.45, training accuracy is 0.375000, time 212.307 samples/sec
+epoch 12, total_step 196100, total loss is 14.72 , inference loss is 8.27, weight deacy loss is 6.45, training accuracy is 0.406250, time 212.883 samples/sec
+epoch 12, total_step 196120, total loss is 16.72 , inference loss is 10.27, weight deacy loss is 6.45, training accuracy is 0.406250, time 222.964 samples/sec
+epoch 12, total_step 196140, total loss is 15.60 , inference loss is 9.15, weight deacy loss is 6.45, training accuracy is 0.468750, time 218.771 samples/sec
+epoch 12, total_step 196160, total loss is 13.85 , inference loss is 7.40, weight deacy loss is 6.45, training accuracy is 0.406250, time 218.528 samples/sec
+epoch 12, total_step 196180, total loss is 12.22 , inference loss is 5.77, weight deacy loss is 6.45, training accuracy is 0.437500, time 215.235 samples/sec
+epoch 12, total_step 196200, total loss is 15.90 , inference loss is 9.45, weight deacy loss is 6.45, training accuracy is 0.281250, time 12.078 samples/sec
+epoch 12, total_step 196220, total loss is 13.86 , inference loss is 7.41, weight deacy loss is 6.45, training accuracy is 0.375000, time 217.321 samples/sec
+epoch 12, total_step 196240, total loss is 13.71 , inference loss is 7.26, weight deacy loss is 6.45, training accuracy is 0.468750, time 223.593 samples/sec
+epoch 12, total_step 196260, total loss is 11.25 , inference loss is 4.80, weight deacy loss is 6.45, training accuracy is 0.406250, time 222.147 samples/sec
+epoch 12, total_step 196280, total loss is 15.49 , inference loss is 9.04, weight deacy loss is 6.45, training accuracy is 0.375000, time 297.156 samples/sec
+epoch 12, total_step 196300, total loss is 11.17 , inference loss is 4.72, weight deacy loss is 6.45, training accuracy is 0.531250, time 216.638 samples/sec
+epoch 12, total_step 196320, total loss is 15.13 , inference loss is 8.68, weight deacy loss is 6.45, training accuracy is 0.312500, time 219.295 samples/sec
+epoch 12, total_step 196340, total loss is 14.00 , inference loss is 7.55, weight deacy loss is 6.45, training accuracy is 0.343750, time 215.288 samples/sec
+epoch 12, total_step 196360, total loss is 17.99 , inference loss is 11.55, weight deacy loss is 6.45, training accuracy is 0.375000, time 230.078 samples/sec
+epoch 12, total_step 196380, total loss is 15.06 , inference loss is 8.62, weight deacy loss is 6.45, training accuracy is 0.468750, time 220.339 samples/sec
+epoch 12, total_step 196400, total loss is 16.12 , inference loss is 9.67, weight deacy loss is 6.45, training accuracy is 0.250000, time 213.268 samples/sec
+epoch 12, total_step 196420, total loss is 11.11 , inference loss is 4.67, weight deacy loss is 6.45, training accuracy is 0.531250, time 215.745 samples/sec
+epoch 12, total_step 196440, total loss is 13.01 , inference loss is 6.57, weight deacy loss is 6.45, training accuracy is 0.468750, time 223.797 samples/sec
+epoch 12, total_step 196460, total loss is 14.25 , inference loss is 7.80, weight deacy loss is 6.45, training accuracy is 0.312500, time 219.349 samples/sec
+epoch 12, total_step 196480, total loss is 14.82 , inference loss is 8.37, weight deacy loss is 6.45, training accuracy is 0.468750, time 214.579 samples/sec
+epoch 12, total_step 196500, total loss is 14.83 , inference loss is 8.38, weight deacy loss is 6.45, training accuracy is 0.312500, time 13.271 samples/sec
+epoch 12, total_step 196520, total loss is 16.03 , inference loss is 9.59, weight deacy loss is 6.45, training accuracy is 0.343750, time 219.997 samples/sec
+epoch 12, total_step 196540, total loss is 18.01 , inference loss is 11.56, weight deacy loss is 6.45, training accuracy is 0.312500, time 216.594 samples/sec
+epoch 12, total_step 196560, total loss is 13.08 , inference loss is 6.63, weight deacy loss is 6.45, training accuracy is 0.343750, time 228.651 samples/sec
+epoch 12, total_step 196580, total loss is 10.20 , inference loss is 3.75, weight deacy loss is 6.45, training accuracy is 0.562500, time 206.509 samples/sec
+epoch 12, total_step 196600, total loss is 16.02 , inference loss is 9.58, weight deacy loss is 6.45, training accuracy is 0.218750, time 220.481 samples/sec
+epoch 12, total_step 196620, total loss is 14.77 , inference loss is 8.33, weight deacy loss is 6.45, training accuracy is 0.562500, time 212.812 samples/sec
+epoch 12, total_step 196640, total loss is 16.83 , inference loss is 10.38, weight deacy loss is 6.45, training accuracy is 0.437500, time 217.015 samples/sec
+epoch 12, total_step 196660, total loss is 12.57 , inference loss is 6.12, weight deacy loss is 6.45, training accuracy is 0.406250, time 213.376 samples/sec
+epoch 12, total_step 196680, total loss is 10.13 , inference loss is 3.69, weight deacy loss is 6.45, training accuracy is 0.437500, time 216.801 samples/sec
+epoch 12, total_step 196700, total loss is 13.29 , inference loss is 6.84, weight deacy loss is 6.45, training accuracy is 0.500000, time 223.004 samples/sec
+epoch 12, total_step 196720, total loss is 14.98 , inference loss is 8.53, weight deacy loss is 6.45, training accuracy is 0.281250, time 296.073 samples/sec
+epoch 12, total_step 196740, total loss is 12.06 , inference loss is 5.62, weight deacy loss is 6.45, training accuracy is 0.375000, time 214.416 samples/sec
+epoch 12, total_step 196760, total loss is 14.91 , inference loss is 8.47, weight deacy loss is 6.45, training accuracy is 0.312500, time 228.778 samples/sec
+epoch 12, total_step 196780, total loss is 15.33 , inference loss is 8.88, weight deacy loss is 6.45, training accuracy is 0.375000, time 218.854 samples/sec
+epoch 12, total_step 196800, total loss is 15.74 , inference loss is 9.30, weight deacy loss is 6.45, training accuracy is 0.562500, time 12.118 samples/sec
+epoch 12, total_step 196820, total loss is 14.63 , inference loss is 8.18, weight deacy loss is 6.45, training accuracy is 0.468750, time 219.840 samples/sec
+epoch 12, total_step 196840, total loss is 15.15 , inference loss is 8.70, weight deacy loss is 6.44, training accuracy is 0.281250, time 224.044 samples/sec
+epoch 12, total_step 196860, total loss is 12.54 , inference loss is 6.09, weight deacy loss is 6.44, training accuracy is 0.343750, time 220.689 samples/sec
+epoch 12, total_step 196880, total loss is 14.42 , inference loss is 7.97, weight deacy loss is 6.44, training accuracy is 0.437500, time 215.859 samples/sec
+epoch 12, total_step 196900, total loss is 14.01 , inference loss is 7.56, weight deacy loss is 6.44, training accuracy is 0.406250, time 220.042 samples/sec
+epoch 12, total_step 196920, total loss is 15.21 , inference loss is 8.77, weight deacy loss is 6.44, training accuracy is 0.375000, time 213.984 samples/sec
+epoch 12, total_step 196940, total loss is 13.30 , inference loss is 6.86, weight deacy loss is 6.44, training accuracy is 0.406250, time 219.215 samples/sec
+epoch 12, total_step 196960, total loss is 19.54 , inference loss is 13.10, weight deacy loss is 6.44, training accuracy is 0.312500, time 219.190 samples/sec
+epoch 12, total_step 196980, total loss is 12.14 , inference loss is 5.69, weight deacy loss is 6.44, training accuracy is 0.375000, time 298.990 samples/sec
+epoch 12, total_step 197000, total loss is 11.11 , inference loss is 4.67, weight deacy loss is 6.44, training accuracy is 0.500000, time 212.186 samples/sec
+epoch 12, total_step 197020, total loss is 13.94 , inference loss is 7.49, weight deacy loss is 6.44, training accuracy is 0.406250, time 219.690 samples/sec
+epoch 12, total_step 197040, total loss is 13.40 , inference loss is 6.96, weight deacy loss is 6.44, training accuracy is 0.531250, time 213.027 samples/sec
+epoch 12, total_step 197060, total loss is 14.66 , inference loss is 8.22, weight deacy loss is 6.44, training accuracy is 0.343750, time 210.440 samples/sec
+epoch 12, total_step 197080, total loss is 10.00 , inference loss is 3.56, weight deacy loss is 6.44, training accuracy is 0.625000, time 212.517 samples/sec
+epoch 12, total_step 197100, total loss is 15.98 , inference loss is 9.54, weight deacy loss is 6.44, training accuracy is 0.250000, time 11.772 samples/sec
+epoch 12, total_step 197120, total loss is 18.01 , inference loss is 11.57, weight deacy loss is 6.44, training accuracy is 0.281250, time 211.707 samples/sec
+epoch 12, total_step 197140, total loss is 12.80 , inference loss is 6.35, weight deacy loss is 6.44, training accuracy is 0.375000, time 222.569 samples/sec
+epoch 12, total_step 197160, total loss is 12.27 , inference loss is 5.83, weight deacy loss is 6.44, training accuracy is 0.281250, time 213.480 samples/sec
+epoch 12, total_step 197180, total loss is 16.31 , inference loss is 9.87, weight deacy loss is 6.44, training accuracy is 0.312500, time 217.288 samples/sec
+epoch 12, total_step 197200, total loss is 14.94 , inference loss is 8.49, weight deacy loss is 6.44, training accuracy is 0.437500, time 215.159 samples/sec
+epoch 12, total_step 197220, total loss is 13.44 , inference loss is 7.00, weight deacy loss is 6.44, training accuracy is 0.406250, time 217.981 samples/sec
+epoch 12, total_step 197240, total loss is 12.54 , inference loss is 6.10, weight deacy loss is 6.44, training accuracy is 0.312500, time 298.871 samples/sec
+epoch 12, total_step 197260, total loss is 16.16 , inference loss is 9.72, weight deacy loss is 6.44, training accuracy is 0.312500, time 216.519 samples/sec
+epoch 12, total_step 197280, total loss is 15.66 , inference loss is 9.22, weight deacy loss is 6.44, training accuracy is 0.281250, time 206.319 samples/sec
+epoch 12, total_step 197300, total loss is 12.65 , inference loss is 6.21, weight deacy loss is 6.44, training accuracy is 0.500000, time 209.452 samples/sec
+epoch 12, total_step 197320, total loss is 14.74 , inference loss is 8.30, weight deacy loss is 6.44, training accuracy is 0.375000, time 213.345 samples/sec
+epoch 12, total_step 197340, total loss is 13.60 , inference loss is 7.16, weight deacy loss is 6.44, training accuracy is 0.281250, time 219.159 samples/sec
+epoch 12, total_step 197360, total loss is 16.74 , inference loss is 10.30, weight deacy loss is 6.44, training accuracy is 0.343750, time 245.966 samples/sec
+epoch 12, total_step 197380, total loss is 11.21 , inference loss is 4.77, weight deacy loss is 6.44, training accuracy is 0.531250, time 213.461 samples/sec
+epoch 12, total_step 197400, total loss is 14.24 , inference loss is 7.80, weight deacy loss is 6.44, training accuracy is 0.312500, time 12.425 samples/sec
+epoch 12, total_step 197420, total loss is 13.99 , inference loss is 7.55, weight deacy loss is 6.44, training accuracy is 0.312500, time 226.538 samples/sec
+epoch 12, total_step 197440, total loss is 13.04 , inference loss is 6.60, weight deacy loss is 6.44, training accuracy is 0.343750, time 215.648 samples/sec
+epoch 12, total_step 197460, total loss is 13.00 , inference loss is 6.56, weight deacy loss is 6.44, training accuracy is 0.375000, time 220.473 samples/sec
+epoch 12, total_step 197480, total loss is 18.56 , inference loss is 12.12, weight deacy loss is 6.44, training accuracy is 0.281250, time 211.616 samples/sec
+epoch 12, total_step 197500, total loss is 17.70 , inference loss is 11.26, weight deacy loss is 6.44, training accuracy is 0.375000, time 299.623 samples/sec
+epoch 12, total_step 197520, total loss is 16.80 , inference loss is 10.36, weight deacy loss is 6.44, training accuracy is 0.406250, time 215.894 samples/sec
+epoch 12, total_step 197540, total loss is 11.91 , inference loss is 5.48, weight deacy loss is 6.44, training accuracy is 0.406250, time 217.949 samples/sec
+epoch 12, total_step 197560, total loss is 12.73 , inference loss is 6.29, weight deacy loss is 6.44, training accuracy is 0.343750, time 223.931 samples/sec
+epoch 12, total_step 197580, total loss is 13.77 , inference loss is 7.33, weight deacy loss is 6.44, training accuracy is 0.468750, time 221.110 samples/sec
+epoch 12, total_step 197600, total loss is 13.11 , inference loss is 6.67, weight deacy loss is 6.44, training accuracy is 0.468750, time 221.568 samples/sec
+epoch 12, total_step 197620, total loss is 15.36 , inference loss is 8.92, weight deacy loss is 6.44, training accuracy is 0.406250, time 214.603 samples/sec
+epoch 12, total_step 197640, total loss is 16.30 , inference loss is 9.86, weight deacy loss is 6.44, training accuracy is 0.406250, time 219.472 samples/sec
+epoch 12, total_step 197660, total loss is 14.33 , inference loss is 7.90, weight deacy loss is 6.44, training accuracy is 0.375000, time 214.257 samples/sec
+epoch 12, total_step 197680, total loss is 13.09 , inference loss is 6.65, weight deacy loss is 6.44, training accuracy is 0.375000, time 215.489 samples/sec
+epoch 12, total_step 197700, total loss is 12.80 , inference loss is 6.36, weight deacy loss is 6.44, training accuracy is 0.281250, time 12.194 samples/sec
+epoch 12, total_step 197720, total loss is 17.07 , inference loss is 10.64, weight deacy loss is 6.44, training accuracy is 0.218750, time 220.977 samples/sec
+epoch 12, total_step 197740, total loss is 17.99 , inference loss is 11.55, weight deacy loss is 6.44, training accuracy is 0.343750, time 204.731 samples/sec
+epoch 12, total_step 197760, total loss is 17.52 , inference loss is 11.09, weight deacy loss is 6.44, training accuracy is 0.281250, time 297.470 samples/sec
+epoch 12, total_step 197780, total loss is 11.98 , inference loss is 5.55, weight deacy loss is 6.44, training accuracy is 0.375000, time 216.711 samples/sec
+epoch 12, total_step 197800, total loss is 14.06 , inference loss is 7.62, weight deacy loss is 6.44, training accuracy is 0.437500, time 224.351 samples/sec
+epoch 12, total_step 197820, total loss is 15.69 , inference loss is 9.26, weight deacy loss is 6.44, training accuracy is 0.375000, time 218.860 samples/sec
+epoch 12, total_step 197840, total loss is 16.20 , inference loss is 9.76, weight deacy loss is 6.44, training accuracy is 0.437500, time 217.325 samples/sec
+epoch 12, total_step 197860, total loss is 14.88 , inference loss is 8.45, weight deacy loss is 6.44, training accuracy is 0.312500, time 220.677 samples/sec
+epoch 12, total_step 197880, total loss is 20.63 , inference loss is 14.19, weight deacy loss is 6.44, training accuracy is 0.312500, time 223.372 samples/sec
+epoch 12, total_step 197900, total loss is 12.95 , inference loss is 6.51, weight deacy loss is 6.44, training accuracy is 0.406250, time 215.136 samples/sec
+epoch 12, total_step 197920, total loss is 14.99 , inference loss is 8.56, weight deacy loss is 6.44, training accuracy is 0.343750, time 207.790 samples/sec
+epoch 12, total_step 197940, total loss is 14.97 , inference loss is 8.53, weight deacy loss is 6.44, training accuracy is 0.343750, time 211.973 samples/sec
+epoch 12, total_step 197960, total loss is 12.01 , inference loss is 5.58, weight deacy loss is 6.44, training accuracy is 0.468750, time 219.438 samples/sec
+epoch 12, total_step 197980, total loss is 20.09 , inference loss is 13.65, weight deacy loss is 6.44, training accuracy is 0.218750, time 208.891 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.52494500000003
+best_threshold_index 114 0.9857407407407407
+best_threshold_index 114 0.9848148148148148
+best_threshold_index 114 0.9844444444444445
+best_threshold_index 108 0.9855555555555555
+best_threshold_index 111 0.9848148148148148
+best_threshold_index 114 0.9840740740740741
+best_threshold_index 114 0.9842592592592593
+best_threshold_index 114 0.9846296296296296
+best_threshold_index 114 0.9829629629629629
+best_threshold_index 111 0.9840740740740741
+[lfw][198000]XNorm: 23.293579
+[lfw][198000]Accuracy-Flip: 0.98283+-0.00850
+test accuracy is: 0.9828333333333333
+epoch 12, total_step 198000, total loss is 16.61 , inference loss is 10.18, weight deacy loss is 6.44, training accuracy is 0.375000, time 18.494 samples/sec
+epoch 12, total_step 198020, total loss is 14.85 , inference loss is 8.42, weight deacy loss is 6.44, training accuracy is 0.312500, time 212.345 samples/sec
+epoch 12, total_step 198040, total loss is 15.68 , inference loss is 9.25, weight deacy loss is 6.43, training accuracy is 0.218750, time 216.835 samples/sec
+epoch 12, total_step 198060, total loss is 12.05 , inference loss is 5.62, weight deacy loss is 6.43, training accuracy is 0.593750, time 213.996 samples/sec
+epoch 12, total_step 198080, total loss is 15.21 , inference loss is 8.77, weight deacy loss is 6.43, training accuracy is 0.312500, time 224.995 samples/sec
+epoch 12, total_step 198100, total loss is 17.51 , inference loss is 11.08, weight deacy loss is 6.43, training accuracy is 0.406250, time 222.717 samples/sec
+epoch 12, total_step 198120, total loss is 17.38 , inference loss is 10.95, weight deacy loss is 6.43, training accuracy is 0.218750, time 219.779 samples/sec
+epoch 12, total_step 198140, total loss is 12.82 , inference loss is 6.39, weight deacy loss is 6.43, training accuracy is 0.375000, time 299.072 samples/sec
+epoch 12, total_step 198160, total loss is 14.60 , inference loss is 8.16, weight deacy loss is 6.43, training accuracy is 0.312500, time 216.008 samples/sec
+epoch 12, total_step 198180, total loss is 14.97 , inference loss is 8.54, weight deacy loss is 6.43, training accuracy is 0.375000, time 219.076 samples/sec
+epoch 12, total_step 198200, total loss is 11.64 , inference loss is 5.20, weight deacy loss is 6.43, training accuracy is 0.343750, time 216.389 samples/sec
+epoch 12, total_step 198220, total loss is 15.79 , inference loss is 9.36, weight deacy loss is 6.43, training accuracy is 0.218750, time 225.193 samples/sec
+epoch 12, total_step 198240, total loss is 14.41 , inference loss is 7.98, weight deacy loss is 6.43, training accuracy is 0.468750, time 215.800 samples/sec
+epoch 12, total_step 198260, total loss is 14.16 , inference loss is 7.73, weight deacy loss is 6.43, training accuracy is 0.187500, time 211.931 samples/sec
+epoch 12, total_step 198280, total loss is 13.74 , inference loss is 7.31, weight deacy loss is 6.43, training accuracy is 0.406250, time 222.235 samples/sec
+epoch 12, total_step 198300, total loss is 13.59 , inference loss is 7.15, weight deacy loss is 6.43, training accuracy is 0.375000, time 12.354 samples/sec
+epoch 12, total_step 198320, total loss is 10.74 , inference loss is 4.31, weight deacy loss is 6.43, training accuracy is 0.406250, time 214.880 samples/sec
+epoch 12, total_step 198340, total loss is 19.76 , inference loss is 13.32, weight deacy loss is 6.43, training accuracy is 0.312500, time 216.867 samples/sec
+epoch 12, total_step 198360, total loss is 23.46 , inference loss is 17.03, weight deacy loss is 6.43, training accuracy is 0.250000, time 220.203 samples/sec
+epoch 12, total_step 198380, total loss is 20.39 , inference loss is 13.96, weight deacy loss is 6.43, training accuracy is 0.312500, time 212.947 samples/sec
+epoch 12, total_step 198400, total loss is 17.35 , inference loss is 10.91, weight deacy loss is 6.43, training accuracy is 0.406250, time 224.438 samples/sec
+epoch 12, total_step 198420, total loss is 18.82 , inference loss is 12.39, weight deacy loss is 6.43, training accuracy is 0.250000, time 221.525 samples/sec
+epoch 12, total_step 198440, total loss is 11.84 , inference loss is 5.40, weight deacy loss is 6.43, training accuracy is 0.406250, time 208.068 samples/sec
+epoch 12, total_step 198460, total loss is 13.29 , inference loss is 6.86, weight deacy loss is 6.43, training accuracy is 0.375000, time 209.586 samples/sec
+epoch 12, total_step 198480, total loss is 11.41 , inference loss is 4.98, weight deacy loss is 6.43, training accuracy is 0.437500, time 211.307 samples/sec
+epoch 12, total_step 198500, total loss is 11.04 , inference loss is 4.61, weight deacy loss is 6.43, training accuracy is 0.437500, time 216.620 samples/sec
+epoch 12, total_step 198520, total loss is 15.55 , inference loss is 9.11, weight deacy loss is 6.43, training accuracy is 0.406250, time 213.515 samples/sec
+epoch 12, total_step 198540, total loss is 14.46 , inference loss is 8.03, weight deacy loss is 6.43, training accuracy is 0.343750, time 214.133 samples/sec
+epoch 12, total_step 198560, total loss is 18.26 , inference loss is 11.83, weight deacy loss is 6.43, training accuracy is 0.343750, time 213.286 samples/sec
+epoch 12, total_step 198580, total loss is 13.74 , inference loss is 7.31, weight deacy loss is 6.43, training accuracy is 0.437500, time 299.692 samples/sec
+epoch 12, total_step 198600, total loss is 16.29 , inference loss is 9.86, weight deacy loss is 6.43, training accuracy is 0.218750, time 12.370 samples/sec
+epoch 12, total_step 198620, total loss is 13.01 , inference loss is 6.58, weight deacy loss is 6.43, training accuracy is 0.281250, time 218.554 samples/sec
+epoch 12, total_step 198640, total loss is 14.54 , inference loss is 8.11, weight deacy loss is 6.43, training accuracy is 0.312500, time 220.611 samples/sec
+epoch 12, total_step 198660, total loss is 14.96 , inference loss is 8.53, weight deacy loss is 6.43, training accuracy is 0.406250, time 217.236 samples/sec
+epoch 12, total_step 198680, total loss is 16.71 , inference loss is 10.28, weight deacy loss is 6.43, training accuracy is 0.250000, time 217.425 samples/sec
+epoch 12, total_step 198700, total loss is 17.77 , inference loss is 11.34, weight deacy loss is 6.43, training accuracy is 0.218750, time 215.050 samples/sec
+epoch 12, total_step 198720, total loss is 14.60 , inference loss is 8.17, weight deacy loss is 6.43, training accuracy is 0.406250, time 215.475 samples/sec
+epoch 12, total_step 198740, total loss is 15.22 , inference loss is 8.79, weight deacy loss is 6.43, training accuracy is 0.312500, time 215.712 samples/sec
+epoch 12, total_step 198760, total loss is 16.05 , inference loss is 9.62, weight deacy loss is 6.43, training accuracy is 0.375000, time 212.793 samples/sec
+epoch 12, total_step 198780, total loss is 11.62 , inference loss is 5.19, weight deacy loss is 6.43, training accuracy is 0.500000, time 223.160 samples/sec
+epoch 12, total_step 198800, total loss is 14.13 , inference loss is 7.70, weight deacy loss is 6.43, training accuracy is 0.281250, time 216.443 samples/sec
+epoch 12, total_step 198820, total loss is 12.85 , inference loss is 6.42, weight deacy loss is 6.43, training accuracy is 0.437500, time 225.367 samples/sec
+epoch 12, total_step 198840, total loss is 11.84 , inference loss is 5.41, weight deacy loss is 6.43, training accuracy is 0.531250, time 296.139 samples/sec
+epoch 12, total_step 198860, total loss is 12.49 , inference loss is 6.06, weight deacy loss is 6.43, training accuracy is 0.375000, time 215.217 samples/sec
+epoch 12, total_step 198880, total loss is 12.88 , inference loss is 6.45, weight deacy loss is 6.43, training accuracy is 0.375000, time 218.571 samples/sec
+epoch 12, total_step 198900, total loss is 20.65 , inference loss is 14.23, weight deacy loss is 6.43, training accuracy is 0.156250, time 13.002 samples/sec
+epoch 12, total_step 198920, total loss is 15.80 , inference loss is 9.37, weight deacy loss is 6.43, training accuracy is 0.281250, time 217.439 samples/sec
+epoch 12, total_step 198940, total loss is 15.21 , inference loss is 8.78, weight deacy loss is 6.43, training accuracy is 0.312500, time 219.400 samples/sec
+epoch 12, total_step 198960, total loss is 15.84 , inference loss is 9.42, weight deacy loss is 6.43, training accuracy is 0.187500, time 218.290 samples/sec
+epoch 12, total_step 198980, total loss is 14.37 , inference loss is 7.94, weight deacy loss is 6.43, training accuracy is 0.406250, time 213.373 samples/sec
+epoch 12, total_step 199000, total loss is 15.19 , inference loss is 8.77, weight deacy loss is 6.43, training accuracy is 0.437500, time 201.090 samples/sec
+epoch 12, total_step 199020, total loss is 11.92 , inference loss is 5.50, weight deacy loss is 6.43, training accuracy is 0.406250, time 198.634 samples/sec
+epoch 12, total_step 199040, total loss is 12.28 , inference loss is 5.86, weight deacy loss is 6.43, training accuracy is 0.343750, time 219.651 samples/sec
+epoch 12, total_step 199060, total loss is 15.45 , inference loss is 9.02, weight deacy loss is 6.43, training accuracy is 0.218750, time 182.907 samples/sec
+epoch 12, total_step 199080, total loss is 14.29 , inference loss is 7.86, weight deacy loss is 6.43, training accuracy is 0.281250, time 200.381 samples/sec2022-12-01 13:31:35.585015: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 6820 of 10000
+2022-12-01 13:31:40.179242: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 12, total_step 199100, total loss is 13.33 , inference loss is 6.91, weight deacy loss is 6.43, training accuracy is 0.312500, time 205.865 samples/sec
+epoch 12, total_step 199120, total loss is 16.22 , inference loss is 9.80, weight deacy loss is 6.43, training accuracy is 0.312500, time 175.513 samples/sec
+epoch 12, total_step 199140, total loss is 11.72 , inference loss is 5.29, weight deacy loss is 6.43, training accuracy is 0.468750, time 299.469 samples/sec
+epoch 12, total_step 199160, total loss is 14.53 , inference loss is 8.10, weight deacy loss is 6.43, training accuracy is 0.281250, time 224.454 samples/sec
+epoch 12, total_step 199180, total loss is 13.73 , inference loss is 7.30, weight deacy loss is 6.43, training accuracy is 0.375000, time 181.891 samples/sec
+epoch 12, total_step 199200, total loss is 14.01 , inference loss is 7.59, weight deacy loss is 6.43, training accuracy is 0.281250, time 12.160 samples/sec
+epoch 12, total_step 199220, total loss is 17.49 , inference loss is 11.06, weight deacy loss is 6.43, training accuracy is 0.343750, time 203.023 samples/sec
+epoch 12, total_step 199240, total loss is 14.99 , inference loss is 8.56, weight deacy loss is 6.43, training accuracy is 0.406250, time 220.045 samples/sec
+epoch 12, total_step 199260, total loss is 13.96 , inference loss is 7.54, weight deacy loss is 6.43, training accuracy is 0.437500, time 222.502 samples/sec
+epoch 12, total_step 199280, total loss is 15.92 , inference loss is 9.49, weight deacy loss is 6.42, training accuracy is 0.437500, time 177.379 samples/sec
+epoch 12, total_step 199300, total loss is 14.09 , inference loss is 7.67, weight deacy loss is 6.42, training accuracy is 0.375000, time 193.820 samples/sec
+End of epoch 12
+epoch 13, total_step 199320, total loss is 13.67 , inference loss is 7.25, weight deacy loss is 6.42, training accuracy is 0.500000, time 232.519 samples/sec
+epoch 13, total_step 199340, total loss is 9.19 , inference loss is 2.76, weight deacy loss is 6.42, training accuracy is 0.750000, time 297.225 samples/sec
+epoch 13, total_step 199360, total loss is 10.55 , inference loss is 4.12, weight deacy loss is 6.42, training accuracy is 0.750000, time 227.530 samples/sec
+epoch 13, total_step 199380, total loss is 9.99 , inference loss is 3.57, weight deacy loss is 6.42, training accuracy is 0.593750, time 228.101 samples/sec
+epoch 13, total_step 199400, total loss is 8.80 , inference loss is 2.38, weight deacy loss is 6.42, training accuracy is 0.781250, time 228.327 samples/sec
+epoch 13, total_step 199420, total loss is 8.67 , inference loss is 2.24, weight deacy loss is 6.42, training accuracy is 0.906250, time 221.201 samples/sec
+epoch 13, total_step 199440, total loss is 12.71 , inference loss is 6.29, weight deacy loss is 6.42, training accuracy is 0.718750, time 215.895 samples/sec
+epoch 13, total_step 199460, total loss is 11.04 , inference loss is 4.61, weight deacy loss is 6.42, training accuracy is 0.718750, time 222.991 samples/sec
+epoch 13, total_step 199480, total loss is 12.26 , inference loss is 5.83, weight deacy loss is 6.42, training accuracy is 0.656250, time 228.868 samples/sec
+epoch 13, total_step 199500, total loss is 10.18 , inference loss is 3.76, weight deacy loss is 6.42, training accuracy is 0.781250, time 11.966 samples/sec
+epoch 13, total_step 199520, total loss is 12.81 , inference loss is 6.39, weight deacy loss is 6.42, training accuracy is 0.656250, time 214.625 samples/sec
+epoch 13, total_step 199540, total loss is 11.93 , inference loss is 5.51, weight deacy loss is 6.42, training accuracy is 0.656250, time 210.720 samples/sec
+epoch 13, total_step 199560, total loss is 10.98 , inference loss is 4.56, weight deacy loss is 6.42, training accuracy is 0.625000, time 200.751 samples/sec
+epoch 13, total_step 199580, total loss is 10.75 , inference loss is 4.33, weight deacy loss is 6.42, training accuracy is 0.718750, time 211.992 samples/sec
+epoch 13, total_step 199600, total loss is 10.23 , inference loss is 3.81, weight deacy loss is 6.42, training accuracy is 0.687500, time 299.176 samples/sec
+epoch 13, total_step 199620, total loss is 10.14 , inference loss is 3.71, weight deacy loss is 6.42, training accuracy is 0.593750, time 237.664 samples/sec
+epoch 13, total_step 199640, total loss is 11.98 , inference loss is 5.56, weight deacy loss is 6.42, training accuracy is 0.718750, time 219.565 samples/sec
+epoch 13, total_step 199660, total loss is 10.66 , inference loss is 4.24, weight deacy loss is 6.42, training accuracy is 0.625000, time 213.441 samples/sec
+epoch 13, total_step 199680, total loss is 13.00 , inference loss is 6.58, weight deacy loss is 6.42, training accuracy is 0.656250, time 235.040 samples/sec
+epoch 13, total_step 199700, total loss is 7.87 , inference loss is 1.45, weight deacy loss is 6.42, training accuracy is 0.812500, time 217.770 samples/sec
+epoch 13, total_step 199720, total loss is 9.48 , inference loss is 3.05, weight deacy loss is 6.42, training accuracy is 0.656250, time 217.590 samples/sec
+epoch 13, total_step 199740, total loss is 9.34 , inference loss is 2.92, weight deacy loss is 6.42, training accuracy is 0.750000, time 210.013 samples/sec
+epoch 13, total_step 199760, total loss is 13.08 , inference loss is 6.66, weight deacy loss is 6.42, training accuracy is 0.593750, time 231.086 samples/sec
+epoch 13, total_step 199780, total loss is 10.40 , inference loss is 3.97, weight deacy loss is 6.42, training accuracy is 0.812500, time 221.137 samples/sec
+epoch 13, total_step 199800, total loss is 10.99 , inference loss is 4.57, weight deacy loss is 6.42, training accuracy is 0.625000, time 12.220 samples/sec
+epoch 13, total_step 199820, total loss is 13.26 , inference loss is 6.84, weight deacy loss is 6.42, training accuracy is 0.437500, time 220.581 samples/sec
+epoch 13, total_step 199840, total loss is 11.51 , inference loss is 5.09, weight deacy loss is 6.42, training accuracy is 0.562500, time 224.214 samples/sec
+epoch 13, total_step 199860, total loss is 17.30 , inference loss is 10.88, weight deacy loss is 6.42, training accuracy is 0.531250, time 295.701 samples/sec
+epoch 13, total_step 199880, total loss is 12.78 , inference loss is 6.36, weight deacy loss is 6.42, training accuracy is 0.687500, time 227.899 samples/sec
+epoch 13, total_step 199900, total loss is 11.68 , inference loss is 5.26, weight deacy loss is 6.42, training accuracy is 0.562500, time 217.670 samples/sec
+epoch 13, total_step 199920, total loss is 14.47 , inference loss is 8.05, weight deacy loss is 6.42, training accuracy is 0.656250, time 242.318 samples/sec
+epoch 13, total_step 199940, total loss is 18.01 , inference loss is 11.59, weight deacy loss is 6.42, training accuracy is 0.500000, time 223.695 samples/sec
+epoch 13, total_step 199960, total loss is 10.25 , inference loss is 3.83, weight deacy loss is 6.42, training accuracy is 0.718750, time 218.436 samples/sec
+epoch 13, total_step 199980, total loss is 10.40 , inference loss is 3.97, weight deacy loss is 6.42, training accuracy is 0.656250, time 214.329 samples/sec
+testing verification..
+(12000, 512)
+infer time 21.83346500000001
+best_threshold_index 115 0.9844444444444445
+best_threshold_index 115 0.9835185185185186
+best_threshold_index 115 0.9838888888888889
+best_threshold_index 115 0.985
+best_threshold_index 115 0.9846296296296296
+best_threshold_index 117 0.9829629629629629
+best_threshold_index 115 0.9844444444444445
+best_threshold_index 115 0.9840740740740741
+best_threshold_index 117 0.9822222222222222
+best_threshold_index 117 0.9837037037037037
+[lfw][200000]XNorm: 18.833636
+[lfw][200000]Accuracy-Flip: 0.98333+-0.00711
+test accuracy is: 0.9833333333333332
+epoch 13, total_step 200000, total loss is 14.51 , inference loss is 8.09, weight deacy loss is 6.42, training accuracy is 0.687500, time 18.194 samples/sec
+epoch 13, total_step 200020, total loss is 11.84 , inference loss is 5.42, weight deacy loss is 6.42, training accuracy is 0.687500, time 214.685 samples/sec
+epoch 13, total_step 200040, total loss is 12.46 , inference loss is 6.03, weight deacy loss is 6.42, training accuracy is 0.750000, time 179.323 samples/sec
+epoch 13, total_step 200060, total loss is 12.07 , inference loss is 5.64, weight deacy loss is 6.42, training accuracy is 0.625000, time 226.743 samples/sec
+epoch 13, total_step 200080, total loss is 11.03 , inference loss is 4.61, weight deacy loss is 6.43, training accuracy is 0.593750, time 232.849 samples/sec
+epoch 13, total_step 200100, total loss is 12.53 , inference loss is 6.11, weight deacy loss is 6.43, training accuracy is 0.500000, time 12.594 samples/sec
+epoch 13, total_step 200120, total loss is 11.76 , inference loss is 5.33, weight deacy loss is 6.43, training accuracy is 0.562500, time 222.611 samples/sec
+epoch 13, total_step 200140, total loss is 12.41 , inference loss is 5.98, weight deacy loss is 6.43, training accuracy is 0.562500, time 217.103 samples/sec
+epoch 13, total_step 200160, total loss is 11.87 , inference loss is 5.44, weight deacy loss is 6.43, training accuracy is 0.656250, time 230.886 samples/sec
+epoch 13, total_step 200180, total loss is 12.60 , inference loss is 6.18, weight deacy loss is 6.43, training accuracy is 0.656250, time 217.836 samples/sec
+epoch 13, total_step 200200, total loss is 10.01 , inference loss is 3.58, weight deacy loss is 6.43, training accuracy is 0.625000, time 221.980 samples/sec
+epoch 13, total_step 200220, total loss is 15.12 , inference loss is 8.69, weight deacy loss is 6.43, training accuracy is 0.593750, time 217.716 samples/sec
+epoch 13, total_step 200240, total loss is 10.90 , inference loss is 4.47, weight deacy loss is 6.43, training accuracy is 0.625000, time 223.421 samples/sec
+epoch 13, total_step 200260, total loss is 14.57 , inference loss is 8.14, weight deacy loss is 6.43, training accuracy is 0.562500, time 224.785 samples/sec
+epoch 13, total_step 200280, total loss is 11.94 , inference loss is 5.51, weight deacy loss is 6.43, training accuracy is 0.656250, time 225.982 samples/sec
+epoch 13, total_step 200300, total loss is 13.24 , inference loss is 6.81, weight deacy loss is 6.43, training accuracy is 0.468750, time 215.250 samples/sec
+epoch 13, total_step 200320, total loss is 11.75 , inference loss is 5.32, weight deacy loss is 6.43, training accuracy is 0.656250, time 228.983 samples/sec
+epoch 13, total_step 200340, total loss is 12.65 , inference loss is 6.23, weight deacy loss is 6.43, training accuracy is 0.531250, time 239.436 samples/sec
+epoch 13, total_step 200360, total loss is 11.64 , inference loss is 5.21, weight deacy loss is 6.43, training accuracy is 0.687500, time 213.765 samples/sec
+epoch 13, total_step 200380, total loss is 15.32 , inference loss is 8.89, weight deacy loss is 6.43, training accuracy is 0.687500, time 299.062 samples/sec
+epoch 13, total_step 200400, total loss is 16.39 , inference loss is 9.96, weight deacy loss is 6.43, training accuracy is 0.343750, time 12.114 samples/sec
+epoch 13, total_step 200420, total loss is 13.56 , inference loss is 7.13, weight deacy loss is 6.43, training accuracy is 0.531250, time 225.521 samples/sec
+epoch 13, total_step 200440, total loss is 11.59 , inference loss is 5.16, weight deacy loss is 6.43, training accuracy is 0.562500, time 212.821 samples/sec
+epoch 13, total_step 200460, total loss is 16.49 , inference loss is 10.07, weight deacy loss is 6.43, training accuracy is 0.437500, time 234.331 samples/sec
+epoch 13, total_step 200480, total loss is 10.87 , inference loss is 4.44, weight deacy loss is 6.43, training accuracy is 0.625000, time 236.229 samples/sec
+epoch 13, total_step 200500, total loss is 11.35 , inference loss is 4.92, weight deacy loss is 6.43, training accuracy is 0.406250, time 224.747 samples/sec
+epoch 13, total_step 200520, total loss is 11.27 , inference loss is 4.84, weight deacy loss is 6.43, training accuracy is 0.625000, time 224.929 samples/sec
+epoch 13, total_step 200540, total loss is 12.29 , inference loss is 5.86, weight deacy loss is 6.43, training accuracy is 0.593750, time 228.627 samples/sec
+epoch 13, total_step 200560, total loss is 13.70 , inference loss is 7.27, weight deacy loss is 6.43, training accuracy is 0.593750, time 224.143 samples/sec
+epoch 13, total_step 200580, total loss is 12.15 , inference loss is 5.72, weight deacy loss is 6.43, training accuracy is 0.562500, time 225.653 samples/sec
+epoch 13, total_step 200600, total loss is 11.83 , inference loss is 5.40, weight deacy loss is 6.43, training accuracy is 0.781250, time 196.468 samples/sec
+epoch 13, total_step 200620, total loss is 10.95 , inference loss is 4.52, weight deacy loss is 6.43, training accuracy is 0.718750, time 194.590 samples/sec
+epoch 13, total_step 200640, total loss is 15.26 , inference loss is 8.83, weight deacy loss is 6.43, training accuracy is 0.562500, time 206.987 samples/sec
+epoch 13, total_step 200660, total loss is 14.89 , inference loss is 8.46, weight deacy loss is 6.43, training accuracy is 0.437500, time 226.140 samples/sec
+epoch 13, total_step 200680, total loss is 12.14 , inference loss is 5.71, weight deacy loss is 6.43, training accuracy is 0.656250, time 217.998 samples/sec
+epoch 13, total_step 200700, total loss is 12.63 , inference loss is 6.20, weight deacy loss is 6.43, training accuracy is 0.593750, time 12.045 samples/sec
+epoch 13, total_step 200720, total loss is 14.49 , inference loss is 8.05, weight deacy loss is 6.43, training accuracy is 0.562500, time 218.374 samples/sec
+epoch 13, total_step 200740, total loss is 14.18 , inference loss is 7.75, weight deacy loss is 6.43, training accuracy is 0.531250, time 226.597 samples/sec
+epoch 13, total_step 200760, total loss is 10.98 , inference loss is 4.55, weight deacy loss is 6.43, training accuracy is 0.656250, time 232.665 samples/sec
+epoch 13, total_step 200780, total loss is 15.00 , inference loss is 8.56, weight deacy loss is 6.43, training accuracy is 0.625000, time 216.026 samples/sec
+epoch 13, total_step 200800, total loss is 11.15 , inference loss is 4.72, weight deacy loss is 6.43, training accuracy is 0.531250, time 223.989 samples/sec
+epoch 13, total_step 200820, total loss is 13.19 , inference loss is 6.76, weight deacy loss is 6.43, training accuracy is 0.343750, time 223.429 samples/sec
+epoch 13, total_step 200840, total loss is 13.84 , inference loss is 7.41, weight deacy loss is 6.43, training accuracy is 0.562500, time 217.288 samples/sec
+epoch 13, total_step 200860, total loss is 10.24 , inference loss is 3.81, weight deacy loss is 6.43, training accuracy is 0.687500, time 236.944 samples/sec
+epoch 13, total_step 200880, total loss is 12.77 , inference loss is 6.33, weight deacy loss is 6.43, training accuracy is 0.593750, time 276.874 samples/sec
+epoch 13, total_step 200900, total loss is 12.83 , inference loss is 6.40, weight deacy loss is 6.43, training accuracy is 0.562500, time 210.677 samples/sec
+epoch 13, total_step 200920, total loss is 15.44 , inference loss is 9.01, weight deacy loss is 6.43, training accuracy is 0.593750, time 225.930 samples/sec
+epoch 13, total_step 200940, total loss is 10.98 , inference loss is 4.55, weight deacy loss is 6.43, training accuracy is 0.656250, time 219.963 samples/sec
+epoch 13, total_step 200960, total loss is 10.84 , inference loss is 4.40, weight deacy loss is 6.43, training accuracy is 0.593750, time 223.168 samples/sec
+epoch 13, total_step 200980, total loss is 12.03 , inference loss is 5.60, weight deacy loss is 6.43, training accuracy is 0.562500, time 220.308 samples/sec
+epoch 13, total_step 201000, total loss is 9.92 , inference loss is 3.49, weight deacy loss is 6.43, training accuracy is 0.625000, time 12.465 samples/sec
+epoch 13, total_step 201020, total loss is 10.99 , inference loss is 4.55, weight deacy loss is 6.43, training accuracy is 0.625000, time 212.966 samples/sec
+epoch 13, total_step 201040, total loss is 11.96 , inference loss is 5.53, weight deacy loss is 6.43, training accuracy is 0.593750, time 211.472 samples/sec
+epoch 13, total_step 201060, total loss is 16.32 , inference loss is 9.89, weight deacy loss is 6.43, training accuracy is 0.468750, time 227.031 samples/sec
+epoch 13, total_step 201080, total loss is 11.80 , inference loss is 5.36, weight deacy loss is 6.43, training accuracy is 0.562500, time 238.290 samples/sec
+epoch 13, total_step 201100, total loss is 15.97 , inference loss is 9.54, weight deacy loss is 6.43, training accuracy is 0.406250, time 207.824 samples/sec
+epoch 13, total_step 201120, total loss is 11.50 , inference loss is 5.07, weight deacy loss is 6.43, training accuracy is 0.625000, time 219.363 samples/sec
+epoch 13, total_step 201140, total loss is 12.64 , inference loss is 6.20, weight deacy loss is 6.43, training accuracy is 0.656250, time 295.558 samples/sec
+epoch 13, total_step 201160, total loss is 14.06 , inference loss is 7.62, weight deacy loss is 6.43, training accuracy is 0.468750, time 219.732 samples/sec
+epoch 13, total_step 201180, total loss is 13.94 , inference loss is 7.51, weight deacy loss is 6.43, training accuracy is 0.531250, time 228.698 samples/sec
+epoch 13, total_step 201200, total loss is 12.25 , inference loss is 5.82, weight deacy loss is 6.43, training accuracy is 0.562500, time 239.344 samples/sec
+epoch 13, total_step 201220, total loss is 16.73 , inference loss is 10.29, weight deacy loss is 6.43, training accuracy is 0.531250, time 213.126 samples/sec
+epoch 13, total_step 201240, total loss is 11.92 , inference loss is 5.48, weight deacy loss is 6.43, training accuracy is 0.562500, time 215.134 samples/sec
+epoch 13, total_step 201260, total loss is 12.66 , inference loss is 6.22, weight deacy loss is 6.43, training accuracy is 0.531250, time 225.209 samples/sec
+epoch 13, total_step 201280, total loss is 13.17 , inference loss is 6.74, weight deacy loss is 6.43, training accuracy is 0.468750, time 225.647 samples/sec
+epoch 13, total_step 201300, total loss is 9.63 , inference loss is 3.20, weight deacy loss is 6.43, training accuracy is 0.656250, time 12.027 samples/sec
+epoch 13, total_step 201320, total loss is 13.91 , inference loss is 7.48, weight deacy loss is 6.43, training accuracy is 0.718750, time 227.001 samples/sec
+epoch 13, total_step 201340, total loss is 11.74 , inference loss is 5.31, weight deacy loss is 6.43, training accuracy is 0.625000, time 227.460 samples/sec
+epoch 13, total_step 201360, total loss is 13.10 , inference loss is 6.67, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.717 samples/sec
+epoch 13, total_step 201380, total loss is 11.93 , inference loss is 5.49, weight deacy loss is 6.43, training accuracy is 0.625000, time 235.075 samples/sec
+epoch 13, total_step 201400, total loss is 10.93 , inference loss is 4.50, weight deacy loss is 6.43, training accuracy is 0.593750, time 298.366 samples/sec
+epoch 13, total_step 201420, total loss is 13.94 , inference loss is 7.50, weight deacy loss is 6.43, training accuracy is 0.500000, time 225.258 samples/sec
+epoch 13, total_step 201440, total loss is 11.09 , inference loss is 4.65, weight deacy loss is 6.43, training accuracy is 0.718750, time 239.163 samples/sec
+epoch 13, total_step 201460, total loss is 15.22 , inference loss is 8.79, weight deacy loss is 6.43, training accuracy is 0.468750, time 226.457 samples/sec
+epoch 13, total_step 201480, total loss is 13.09 , inference loss is 6.65, weight deacy loss is 6.43, training accuracy is 0.687500, time 199.897 samples/sec
+epoch 13, total_step 201500, total loss is 13.20 , inference loss is 6.77, weight deacy loss is 6.43, training accuracy is 0.468750, time 223.565 samples/sec
+epoch 13, total_step 201520, total loss is 11.10 , inference loss is 4.67, weight deacy loss is 6.43, training accuracy is 0.562500, time 219.068 samples/sec
+epoch 13, total_step 201540, total loss is 12.25 , inference loss is 5.81, weight deacy loss is 6.43, training accuracy is 0.750000, time 223.502 samples/sec
+epoch 13, total_step 201560, total loss is 13.88 , inference loss is 7.45, weight deacy loss is 6.43, training accuracy is 0.625000, time 223.270 samples/sec
+epoch 13, total_step 201580, total loss is 10.23 , inference loss is 3.80, weight deacy loss is 6.43, training accuracy is 0.656250, time 227.075 samples/sec
+epoch 13, total_step 201600, total loss is 11.39 , inference loss is 4.96, weight deacy loss is 6.43, training accuracy is 0.593750, time 12.772 samples/sec
+epoch 13, total_step 201620, total loss is 15.98 , inference loss is 9.55, weight deacy loss is 6.43, training accuracy is 0.406250, time 229.383 samples/sec
+epoch 13, total_step 201640, total loss is 12.53 , inference loss is 6.09, weight deacy loss is 6.43, training accuracy is 0.625000, time 221.590 samples/sec
+epoch 13, total_step 201660, total loss is 9.30 , inference loss is 2.87, weight deacy loss is 6.43, training accuracy is 0.625000, time 296.647 samples/sec
+epoch 13, total_step 201680, total loss is 11.19 , inference loss is 4.76, weight deacy loss is 6.43, training accuracy is 0.625000, time 224.481 samples/sec
+epoch 13, total_step 201700, total loss is 8.61 , inference loss is 2.17, weight deacy loss is 6.43, training accuracy is 0.812500, time 220.046 samples/sec
+epoch 13, total_step 201720, total loss is 10.52 , inference loss is 4.08, weight deacy loss is 6.43, training accuracy is 0.562500, time 223.504 samples/sec
+epoch 13, total_step 201740, total loss is 12.06 , inference loss is 5.63, weight deacy loss is 6.43, training accuracy is 0.656250, time 235.239 samples/sec
+epoch 13, total_step 201760, total loss is 14.99 , inference loss is 8.56, weight deacy loss is 6.43, training accuracy is 0.531250, time 228.859 samples/sec
+epoch 13, total_step 201780, total loss is 10.17 , inference loss is 3.74, weight deacy loss is 6.43, training accuracy is 0.656250, time 211.881 samples/sec
+epoch 13, total_step 201800, total loss is 9.94 , inference loss is 3.51, weight deacy loss is 6.43, training accuracy is 0.656250, time 215.631 samples/sec
+epoch 13, total_step 201820, total loss is 10.34 , inference loss is 3.90, weight deacy loss is 6.43, training accuracy is 0.656250, time 220.032 samples/sec
+epoch 13, total_step 201840, total loss is 11.86 , inference loss is 5.43, weight deacy loss is 6.43, training accuracy is 0.625000, time 224.938 samples/sec
+epoch 13, total_step 201860, total loss is 8.03 , inference loss is 1.60, weight deacy loss is 6.43, training accuracy is 0.656250, time 212.834 samples/sec
+epoch 13, total_step 201880, total loss is 9.17 , inference loss is 2.74, weight deacy loss is 6.43, training accuracy is 0.687500, time 211.978 samples/sec
+epoch 13, total_step 201900, total loss is 13.19 , inference loss is 6.76, weight deacy loss is 6.43, training accuracy is 0.531250, time 12.936 samples/sec
+epoch 13, total_step 201920, total loss is 10.77 , inference loss is 4.34, weight deacy loss is 6.43, training accuracy is 0.531250, time 226.068 samples/sec
+epoch 13, total_step 201940, total loss is 14.80 , inference loss is 8.37, weight deacy loss is 6.43, training accuracy is 0.500000, time 237.252 samples/sec
+epoch 13, total_step 201960, total loss is 8.20 , inference loss is 1.77, weight deacy loss is 6.43, training accuracy is 0.625000, time 220.046 samples/sec
+epoch 13, total_step 201980, total loss is 13.72 , inference loss is 7.29, weight deacy loss is 6.43, training accuracy is 0.562500, time 215.427 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.354742999999992
+best_threshold_index 117 0.9851851851851852
+best_threshold_index 117 0.9846296296296296
+best_threshold_index 116 0.9837037037037037
+best_threshold_index 115 0.9851851851851852
+best_threshold_index 115 0.9842592592592593
+best_threshold_index 117 0.9833333333333333
+best_threshold_index 117 0.9848148148148148
+best_threshold_index 117 0.9831481481481481
+best_threshold_index 117 0.9837037037037037
+best_threshold_index 117 0.9837037037037037
+[lfw][202000]XNorm: 20.752503
+[lfw][202000]Accuracy-Flip: 0.98333+-0.00695
+test accuracy is: 0.9833333333333334
+epoch 13, total_step 202000, total loss is 14.89 , inference loss is 8.46, weight deacy loss is 6.43, training accuracy is 0.500000, time 17.447 samples/sec
+epoch 13, total_step 202020, total loss is 13.23 , inference loss is 6.80, weight deacy loss is 6.43, training accuracy is 0.468750, time 220.168 samples/sec
+epoch 13, total_step 202040, total loss is 13.64 , inference loss is 7.21, weight deacy loss is 6.43, training accuracy is 0.593750, time 220.775 samples/sec
+epoch 13, total_step 202060, total loss is 11.71 , inference loss is 5.28, weight deacy loss is 6.43, training accuracy is 0.687500, time 227.967 samples/sec
+epoch 13, total_step 202080, total loss is 12.78 , inference loss is 6.35, weight deacy loss is 6.43, training accuracy is 0.562500, time 227.922 samples/sec
+epoch 13, total_step 202100, total loss is 12.76 , inference loss is 6.32, weight deacy loss is 6.43, training accuracy is 0.562500, time 221.464 samples/sec
+epoch 13, total_step 202120, total loss is 10.31 , inference loss is 3.88, weight deacy loss is 6.43, training accuracy is 0.531250, time 216.155 samples/sec
+epoch 13, total_step 202140, total loss is 9.60 , inference loss is 3.17, weight deacy loss is 6.43, training accuracy is 0.625000, time 236.477 samples/sec
+epoch 13, total_step 202160, total loss is 10.87 , inference loss is 4.44, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.377 samples/sec
+epoch 13, total_step 202180, total loss is 8.44 , inference loss is 2.01, weight deacy loss is 6.43, training accuracy is 0.781250, time 230.614 samples/sec
+epoch 13, total_step 202200, total loss is 8.27 , inference loss is 1.84, weight deacy loss is 6.43, training accuracy is 0.750000, time 12.345 samples/sec
+epoch 13, total_step 202220, total loss is 11.13 , inference loss is 4.70, weight deacy loss is 6.43, training accuracy is 0.687500, time 232.464 samples/sec
+epoch 13, total_step 202240, total loss is 12.28 , inference loss is 5.85, weight deacy loss is 6.43, training accuracy is 0.593750, time 213.916 samples/sec
+epoch 13, total_step 202260, total loss is 11.22 , inference loss is 4.79, weight deacy loss is 6.43, training accuracy is 0.687500, time 222.522 samples/sec
+epoch 13, total_step 202280, total loss is 8.84 , inference loss is 2.41, weight deacy loss is 6.43, training accuracy is 0.687500, time 217.400 samples/sec
+epoch 13, total_step 202300, total loss is 9.70 , inference loss is 3.27, weight deacy loss is 6.43, training accuracy is 0.656250, time 223.529 samples/sec
+epoch 13, total_step 202320, total loss is 11.45 , inference loss is 5.02, weight deacy loss is 6.43, training accuracy is 0.593750, time 238.184 samples/sec
+epoch 13, total_step 202340, total loss is 11.31 , inference loss is 4.88, weight deacy loss is 6.43, training accuracy is 0.593750, time 224.670 samples/sec
+epoch 13, total_step 202360, total loss is 11.92 , inference loss is 5.49, weight deacy loss is 6.43, training accuracy is 0.625000, time 224.164 samples/sec
+epoch 13, total_step 202380, total loss is 11.20 , inference loss is 4.78, weight deacy loss is 6.43, training accuracy is 0.562500, time 221.038 samples/sec
+epoch 13, total_step 202400, total loss is 8.70 , inference loss is 2.27, weight deacy loss is 6.43, training accuracy is 0.750000, time 238.661 samples/sec
+epoch 13, total_step 202420, total loss is 13.82 , inference loss is 7.39, weight deacy loss is 6.43, training accuracy is 0.468750, time 226.401 samples/sec
+epoch 13, total_step 202440, total loss is 13.26 , inference loss is 6.83, weight deacy loss is 6.43, training accuracy is 0.531250, time 222.895 samples/sec
+epoch 13, total_step 202460, total loss is 12.58 , inference loss is 6.15, weight deacy loss is 6.43, training accuracy is 0.593750, time 296.385 samples/sec
+epoch 13, total_step 202480, total loss is 9.92 , inference loss is 3.49, weight deacy loss is 6.43, training accuracy is 0.562500, time 219.642 samples/sec
+epoch 13, total_step 202500, total loss is 9.51 , inference loss is 3.08, weight deacy loss is 6.43, training accuracy is 0.750000, time 13.353 samples/sec
+epoch 13, total_step 202520, total loss is 13.01 , inference loss is 6.58, weight deacy loss is 6.43, training accuracy is 0.468750, time 219.516 samples/sec
+epoch 13, total_step 202540, total loss is 10.54 , inference loss is 4.11, weight deacy loss is 6.43, training accuracy is 0.687500, time 210.942 samples/sec
+epoch 13, total_step 202560, total loss is 12.64 , inference loss is 6.21, weight deacy loss is 6.43, training accuracy is 0.625000, time 196.916 samples/sec
+epoch 13, total_step 202580, total loss is 12.24 , inference loss is 5.81, weight deacy loss is 6.43, training accuracy is 0.625000, time 236.525 samples/sec
+epoch 13, total_step 202600, total loss is 11.37 , inference loss is 4.94, weight deacy loss is 6.43, training accuracy is 0.625000, time 221.047 samples/sec
+epoch 13, total_step 202620, total loss is 13.51 , inference loss is 7.08, weight deacy loss is 6.43, training accuracy is 0.562500, time 178.336 samples/sec
+epoch 13, total_step 202640, total loss is 13.64 , inference loss is 7.21, weight deacy loss is 6.43, training accuracy is 0.468750, time 223.853 samples/sec
+epoch 13, total_step 202660, total loss is 13.78 , inference loss is 7.35, weight deacy loss is 6.43, training accuracy is 0.468750, time 231.361 samples/sec
+epoch 13, total_step 202680, total loss is 11.84 , inference loss is 5.41, weight deacy loss is 6.43, training accuracy is 0.656250, time 175.712 samples/sec
+epoch 13, total_step 202700, total loss is 14.67 , inference loss is 8.24, weight deacy loss is 6.43, training accuracy is 0.468750, time 232.026 samples/sec
+epoch 13, total_step 202720, total loss is 14.34 , inference loss is 7.91, weight deacy loss is 6.43, training accuracy is 0.531250, time 298.785 samples/sec
+epoch 13, total_step 202740, total loss is 15.69 , inference loss is 9.26, weight deacy loss is 6.43, training accuracy is 0.531250, time 225.971 samples/sec
+epoch 13, total_step 202760, total loss is 13.93 , inference loss is 7.50, weight deacy loss is 6.43, training accuracy is 0.562500, time 216.926 samples/sec
+epoch 13, total_step 202780, total loss is 14.16 , inference loss is 7.74, weight deacy loss is 6.43, training accuracy is 0.500000, time 224.980 samples/sec
+epoch 13, total_step 202800, total loss is 14.21 , inference loss is 7.78, weight deacy loss is 6.43, training accuracy is 0.500000, time 12.453 samples/sec
+epoch 13, total_step 202820, total loss is 11.71 , inference loss is 5.28, weight deacy loss is 6.43, training accuracy is 0.562500, time 226.642 samples/sec
+epoch 13, total_step 202840, total loss is 11.87 , inference loss is 5.44, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.051 samples/sec
+epoch 13, total_step 202860, total loss is 9.66 , inference loss is 3.23, weight deacy loss is 6.43, training accuracy is 0.687500, time 240.935 samples/sec
+epoch 13, total_step 202880, total loss is 14.80 , inference loss is 8.37, weight deacy loss is 6.43, training accuracy is 0.406250, time 224.541 samples/sec
+epoch 13, total_step 202900, total loss is 14.73 , inference loss is 8.31, weight deacy loss is 6.43, training accuracy is 0.625000, time 221.523 samples/sec
+epoch 13, total_step 202920, total loss is 15.45 , inference loss is 9.02, weight deacy loss is 6.43, training accuracy is 0.468750, time 225.002 samples/sec
+epoch 13, total_step 202940, total loss is 12.12 , inference loss is 5.69, weight deacy loss is 6.43, training accuracy is 0.625000, time 226.092 samples/sec
+epoch 13, total_step 202960, total loss is 14.74 , inference loss is 8.32, weight deacy loss is 6.43, training accuracy is 0.468750, time 229.632 samples/sec
+epoch 13, total_step 202980, total loss is 12.00 , inference loss is 5.57, weight deacy loss is 6.43, training accuracy is 0.593750, time 296.833 samples/sec
+epoch 13, total_step 203000, total loss is 12.09 , inference loss is 5.67, weight deacy loss is 6.43, training accuracy is 0.468750, time 224.340 samples/sec
+epoch 13, total_step 203020, total loss is 11.77 , inference loss is 5.34, weight deacy loss is 6.43, training accuracy is 0.687500, time 220.733 samples/sec
+epoch 13, total_step 203040, total loss is 16.33 , inference loss is 9.90, weight deacy loss is 6.43, training accuracy is 0.531250, time 232.192 samples/sec
+epoch 13, total_step 203060, total loss is 11.75 , inference loss is 5.32, weight deacy loss is 6.43, training accuracy is 0.718750, time 225.642 samples/sec
+epoch 13, total_step 203080, total loss is 11.58 , inference loss is 5.15, weight deacy loss is 6.43, training accuracy is 0.468750, time 227.792 samples/sec
+epoch 13, total_step 203100, total loss is 11.92 , inference loss is 5.49, weight deacy loss is 6.43, training accuracy is 0.562500, time 12.143 samples/sec
+epoch 13, total_step 203120, total loss is 12.29 , inference loss is 5.86, weight deacy loss is 6.43, training accuracy is 0.593750, time 224.756 samples/sec
+epoch 13, total_step 203140, total loss is 13.72 , inference loss is 7.29, weight deacy loss is 6.43, training accuracy is 0.500000, time 209.581 samples/sec
+epoch 13, total_step 203160, total loss is 13.88 , inference loss is 7.46, weight deacy loss is 6.43, training accuracy is 0.562500, time 219.574 samples/sec
+epoch 13, total_step 203180, total loss is 18.69 , inference loss is 12.26, weight deacy loss is 6.43, training accuracy is 0.437500, time 225.145 samples/sec
+epoch 13, total_step 203200, total loss is 13.32 , inference loss is 6.89, weight deacy loss is 6.43, training accuracy is 0.437500, time 223.462 samples/sec
+epoch 13, total_step 203220, total loss is 14.88 , inference loss is 8.45, weight deacy loss is 6.43, training accuracy is 0.531250, time 224.385 samples/sec
+epoch 13, total_step 203240, total loss is 10.59 , inference loss is 4.17, weight deacy loss is 6.43, training accuracy is 0.593750, time 299.850 samples/sec
+epoch 13, total_step 203260, total loss is 15.53 , inference loss is 9.10, weight deacy loss is 6.43, training accuracy is 0.375000, time 214.646 samples/sec
+epoch 13, total_step 203280, total loss is 17.72 , inference loss is 11.29, weight deacy loss is 6.43, training accuracy is 0.375000, time 220.438 samples/sec
+epoch 13, total_step 203300, total loss is 10.22 , inference loss is 3.79, weight deacy loss is 6.43, training accuracy is 0.531250, time 217.595 samples/sec
+epoch 13, total_step 203320, total loss is 13.71 , inference loss is 7.28, weight deacy loss is 6.43, training accuracy is 0.562500, time 217.640 samples/sec
+epoch 13, total_step 203340, total loss is 9.67 , inference loss is 3.24, weight deacy loss is 6.43, training accuracy is 0.656250, time 222.581 samples/sec
+epoch 13, total_step 203360, total loss is 13.86 , inference loss is 7.43, weight deacy loss is 6.43, training accuracy is 0.562500, time 211.164 samples/sec
+epoch 13, total_step 203380, total loss is 11.84 , inference loss is 5.41, weight deacy loss is 6.43, training accuracy is 0.531250, time 223.249 samples/sec
+epoch 13, total_step 203400, total loss is 15.93 , inference loss is 9.50, weight deacy loss is 6.43, training accuracy is 0.468750, time 13.472 samples/sec
+epoch 13, total_step 203420, total loss is 10.64 , inference loss is 4.21, weight deacy loss is 6.43, training accuracy is 0.562500, time 225.575 samples/sec
+epoch 13, total_step 203440, total loss is 15.40 , inference loss is 8.97, weight deacy loss is 6.43, training accuracy is 0.468750, time 222.297 samples/sec
+epoch 13, total_step 203460, total loss is 11.89 , inference loss is 5.46, weight deacy loss is 6.43, training accuracy is 0.531250, time 205.138 samples/sec
+epoch 13, total_step 203480, total loss is 13.42 , inference loss is 6.99, weight deacy loss is 6.43, training accuracy is 0.468750, time 209.703 samples/sec
+epoch 13, total_step 203500, total loss is 13.02 , inference loss is 6.59, weight deacy loss is 6.43, training accuracy is 0.500000, time 299.446 samples/sec
+epoch 13, total_step 203520, total loss is 14.48 , inference loss is 8.05, weight deacy loss is 6.43, training accuracy is 0.500000, time 216.892 samples/sec
+epoch 13, total_step 203540, total loss is 19.19 , inference loss is 12.76, weight deacy loss is 6.43, training accuracy is 0.343750, time 219.992 samples/sec
+epoch 13, total_step 203560, total loss is 11.43 , inference loss is 5.00, weight deacy loss is 6.43, training accuracy is 0.562500, time 227.500 samples/sec
+epoch 13, total_step 203580, total loss is 12.02 , inference loss is 5.59, weight deacy loss is 6.43, training accuracy is 0.468750, time 217.057 samples/sec
+epoch 13, total_step 203600, total loss is 13.91 , inference loss is 7.48, weight deacy loss is 6.43, training accuracy is 0.531250, time 212.660 samples/sec
+epoch 13, total_step 203620, total loss is 14.35 , inference loss is 7.92, weight deacy loss is 6.43, training accuracy is 0.406250, time 213.123 samples/sec
+epoch 13, total_step 203640, total loss is 12.17 , inference loss is 5.75, weight deacy loss is 6.43, training accuracy is 0.562500, time 225.349 samples/sec
+epoch 13, total_step 203660, total loss is 9.99 , inference loss is 3.57, weight deacy loss is 6.43, training accuracy is 0.656250, time 298.204 samples/sec
+epoch 13, total_step 203680, total loss is 12.90 , inference loss is 6.47, weight deacy loss is 6.43, training accuracy is 0.562500, time 296.152 samples/sec
+epoch 13, total_step 203700, total loss is 15.91 , inference loss is 9.48, weight deacy loss is 6.43, training accuracy is 0.375000, time 11.882 samples/sec
+epoch 13, total_step 203720, total loss is 13.32 , inference loss is 6.89, weight deacy loss is 6.43, training accuracy is 0.625000, time 221.336 samples/sec
+epoch 13, total_step 203740, total loss is 12.14 , inference loss is 5.71, weight deacy loss is 6.43, training accuracy is 0.593750, time 220.513 samples/sec
+epoch 13, total_step 203760, total loss is 11.08 , inference loss is 4.66, weight deacy loss is 6.43, training accuracy is 0.593750, time 228.690 samples/sec
+epoch 13, total_step 203780, total loss is 12.28 , inference loss is 5.85, weight deacy loss is 6.43, training accuracy is 0.437500, time 221.263 samples/sec
+epoch 13, total_step 203800, total loss is 13.33 , inference loss is 6.90, weight deacy loss is 6.43, training accuracy is 0.406250, time 224.827 samples/sec
+epoch 13, total_step 203820, total loss is 15.57 , inference loss is 9.15, weight deacy loss is 6.43, training accuracy is 0.375000, time 219.810 samples/sec
+epoch 13, total_step 203840, total loss is 15.82 , inference loss is 9.39, weight deacy loss is 6.43, training accuracy is 0.500000, time 224.974 samples/sec
+epoch 13, total_step 203860, total loss is 16.95 , inference loss is 10.52, weight deacy loss is 6.43, training accuracy is 0.437500, time 219.483 samples/sec
+epoch 13, total_step 203880, total loss is 15.41 , inference loss is 8.98, weight deacy loss is 6.43, training accuracy is 0.562500, time 220.408 samples/sec
+epoch 13, total_step 203900, total loss is 10.24 , inference loss is 3.81, weight deacy loss is 6.43, training accuracy is 0.562500, time 228.195 samples/sec
+epoch 13, total_step 203920, total loss is 12.50 , inference loss is 6.07, weight deacy loss is 6.43, training accuracy is 0.468750, time 202.202 samples/sec
+epoch 13, total_step 203940, total loss is 12.71 , inference loss is 6.28, weight deacy loss is 6.43, training accuracy is 0.500000, time 227.262 samples/sec
+epoch 13, total_step 203960, total loss is 12.46 , inference loss is 6.03, weight deacy loss is 6.43, training accuracy is 0.531250, time 206.187 samples/sec
+epoch 13, total_step 203980, total loss is 14.25 , inference loss is 7.82, weight deacy loss is 6.43, training accuracy is 0.500000, time 217.206 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.903029000000025
+best_threshold_index 117 0.9892592592592593
+best_threshold_index 120 0.9885185185185185
+best_threshold_index 117 0.9887037037037038
+best_threshold_index 117 0.9888888888888889
+best_threshold_index 117 0.9888888888888889
+best_threshold_index 117 0.9877777777777778
+best_threshold_index 117 0.9888888888888889
+best_threshold_index 117 0.9874074074074074
+best_threshold_index 117 0.9872222222222222
+best_threshold_index 117 0.9881481481481481
+[lfw][204000]XNorm: 21.583302
+[lfw][204000]Accuracy-Flip: 0.98783+-0.00601
+test accuracy is: 0.9878333333333333
+epoch 13, total_step 204000, total loss is 11.89 , inference loss is 5.46, weight deacy loss is 6.43, training accuracy is 0.562500, time 17.480 samples/sec
+epoch 13, total_step 204020, total loss is 13.17 , inference loss is 6.74, weight deacy loss is 6.43, training accuracy is 0.656250, time 218.288 samples/sec
+epoch 13, total_step 204040, total loss is 12.99 , inference loss is 6.57, weight deacy loss is 6.43, training accuracy is 0.437500, time 226.899 samples/sec
+epoch 13, total_step 204060, total loss is 12.48 , inference loss is 6.05, weight deacy loss is 6.43, training accuracy is 0.500000, time 223.651 samples/sec
+epoch 13, total_step 204080, total loss is 10.89 , inference loss is 4.46, weight deacy loss is 6.43, training accuracy is 0.500000, time 223.729 samples/sec
+epoch 13, total_step 204100, total loss is 15.23 , inference loss is 8.81, weight deacy loss is 6.43, training accuracy is 0.593750, time 227.228 samples/sec
+epoch 13, total_step 204120, total loss is 16.03 , inference loss is 9.60, weight deacy loss is 6.43, training accuracy is 0.531250, time 233.368 samples/sec
+epoch 13, total_step 204140, total loss is 13.94 , inference loss is 7.51, weight deacy loss is 6.43, training accuracy is 0.468750, time 213.846 samples/sec
+epoch 13, total_step 204160, total loss is 11.58 , inference loss is 5.15, weight deacy loss is 6.43, training accuracy is 0.562500, time 224.841 samples/sec
+epoch 13, total_step 204180, total loss is 11.87 , inference loss is 5.44, weight deacy loss is 6.43, training accuracy is 0.437500, time 222.328 samples/sec
+epoch 13, total_step 204200, total loss is 10.46 , inference loss is 4.03, weight deacy loss is 6.43, training accuracy is 0.593750, time 223.141 samples/sec
+epoch 13, total_step 204220, total loss is 11.94 , inference loss is 5.51, weight deacy loss is 6.43, training accuracy is 0.500000, time 222.538 samples/sec
+epoch 13, total_step 204240, total loss is 15.87 , inference loss is 9.44, weight deacy loss is 6.43, training accuracy is 0.343750, time 218.516 samples/sec
+epoch 13, total_step 204260, total loss is 13.81 , inference loss is 7.38, weight deacy loss is 6.43, training accuracy is 0.468750, time 239.279 samples/sec
+epoch 13, total_step 204280, total loss is 14.53 , inference loss is 8.11, weight deacy loss is 6.43, training accuracy is 0.375000, time 222.918 samples/sec
+epoch 13, total_step 204300, total loss is 11.22 , inference loss is 4.79, weight deacy loss is 6.43, training accuracy is 0.500000, time 12.881 samples/sec
+epoch 13, total_step 204320, total loss is 13.54 , inference loss is 7.11, weight deacy loss is 6.43, training accuracy is 0.406250, time 222.209 samples/sec
+epoch 13, total_step 204340, total loss is 16.94 , inference loss is 10.52, weight deacy loss is 6.43, training accuracy is 0.593750, time 221.034 samples/sec
+epoch 13, total_step 204360, total loss is 14.88 , inference loss is 8.46, weight deacy loss is 6.43, training accuracy is 0.531250, time 223.271 samples/sec
+epoch 13, total_step 204380, total loss is 9.69 , inference loss is 3.26, weight deacy loss is 6.43, training accuracy is 0.656250, time 223.591 samples/sec
+epoch 13, total_step 204400, total loss is 14.82 , inference loss is 8.40, weight deacy loss is 6.43, training accuracy is 0.406250, time 227.421 samples/sec
+epoch 13, total_step 204420, total loss is 12.11 , inference loss is 5.68, weight deacy loss is 6.43, training accuracy is 0.406250, time 226.849 samples/sec
+epoch 13, total_step 204440, total loss is 16.64 , inference loss is 10.22, weight deacy loss is 6.43, training accuracy is 0.406250, time 224.180 samples/sec
+epoch 13, total_step 204460, total loss is 15.21 , inference loss is 8.79, weight deacy loss is 6.43, training accuracy is 0.500000, time 223.482 samples/sec
+epoch 13, total_step 204480, total loss is 13.94 , inference loss is 7.51, weight deacy loss is 6.43, training accuracy is 0.500000, time 201.556 samples/sec
+epoch 13, total_step 204500, total loss is 13.53 , inference loss is 7.11, weight deacy loss is 6.43, training accuracy is 0.562500, time 242.056 samples/sec
+epoch 13, total_step 204520, total loss is 11.18 , inference loss is 4.75, weight deacy loss is 6.43, training accuracy is 0.687500, time 213.600 samples/sec
+epoch 13, total_step 204540, total loss is 12.95 , inference loss is 6.53, weight deacy loss is 6.43, training accuracy is 0.562500, time 227.578 samples/sec
+epoch 13, total_step 204560, total loss is 9.53 , inference loss is 3.11, weight deacy loss is 6.43, training accuracy is 0.656250, time 215.489 samples/sec
+epoch 13, total_step 204580, total loss is 11.30 , inference loss is 4.88, weight deacy loss is 6.43, training accuracy is 0.406250, time 294.322 samples/sec
+epoch 13, total_step 204600, total loss is 12.99 , inference loss is 6.57, weight deacy loss is 6.43, training accuracy is 0.406250, time 12.374 samples/sec
+epoch 13, total_step 204620, total loss is 12.54 , inference loss is 6.11, weight deacy loss is 6.43, training accuracy is 0.656250, time 224.399 samples/sec
+epoch 13, total_step 204640, total loss is 12.02 , inference loss is 5.60, weight deacy loss is 6.43, training accuracy is 0.468750, time 224.519 samples/sec
+epoch 13, total_step 204660, total loss is 17.15 , inference loss is 10.72, weight deacy loss is 6.43, training accuracy is 0.375000, time 225.263 samples/sec
+epoch 13, total_step 204680, total loss is 13.44 , inference loss is 7.01, weight deacy loss is 6.43, training accuracy is 0.500000, time 229.021 samples/sec
+epoch 13, total_step 204700, total loss is 13.10 , inference loss is 6.68, weight deacy loss is 6.43, training accuracy is 0.468750, time 224.360 samples/sec
+epoch 13, total_step 204720, total loss is 11.36 , inference loss is 4.94, weight deacy loss is 6.43, training accuracy is 0.343750, time 197.748 samples/sec
+epoch 13, total_step 204740, total loss is 13.09 , inference loss is 6.67, weight deacy loss is 6.42, training accuracy is 0.562500, time 223.520 samples/sec
+epoch 13, total_step 204760, total loss is 15.24 , inference loss is 8.82, weight deacy loss is 6.42, training accuracy is 0.406250, time 213.776 samples/sec
+epoch 13, total_step 204780, total loss is 13.52 , inference loss is 7.09, weight deacy loss is 6.42, training accuracy is 0.531250, time 223.381 samples/sec
+epoch 13, total_step 204800, total loss is 16.46 , inference loss is 10.04, weight deacy loss is 6.42, training accuracy is 0.375000, time 225.030 samples/sec
+epoch 13, total_step 204820, total loss is 11.50 , inference loss is 5.07, weight deacy loss is 6.42, training accuracy is 0.625000, time 223.075 samples/sec
+epoch 13, total_step 204840, total loss is 12.57 , inference loss is 6.15, weight deacy loss is 6.42, training accuracy is 0.531250, time 206.779 samples/sec
+epoch 13, total_step 204860, total loss is 13.34 , inference loss is 6.92, weight deacy loss is 6.42, training accuracy is 0.375000, time 221.982 samples/sec
+epoch 13, total_step 204880, total loss is 14.20 , inference loss is 7.78, weight deacy loss is 6.42, training accuracy is 0.562500, time 242.323 samples/sec
+epoch 13, total_step 204900, total loss is 11.83 , inference loss is 5.41, weight deacy loss is 6.42, training accuracy is 0.468750, time 13.629 samples/sec
+epoch 13, total_step 204920, total loss is 15.02 , inference loss is 8.60, weight deacy loss is 6.42, training accuracy is 0.437500, time 226.225 samples/sec
+epoch 13, total_step 204940, total loss is 9.37 , inference loss is 2.95, weight deacy loss is 6.42, training accuracy is 0.687500, time 231.983 samples/sec
+epoch 13, total_step 204960, total loss is 12.80 , inference loss is 6.38, weight deacy loss is 6.42, training accuracy is 0.593750, time 223.525 samples/sec
+epoch 13, total_step 204980, total loss is 11.43 , inference loss is 5.00, weight deacy loss is 6.42, training accuracy is 0.593750, time 242.162 samples/sec
+epoch 13, total_step 205000, total loss is 9.83 , inference loss is 3.41, weight deacy loss is 6.42, training accuracy is 0.531250, time 225.333 samples/sec
+epoch 13, total_step 205020, total loss is 9.56 , inference loss is 3.13, weight deacy loss is 6.42, training accuracy is 0.656250, time 223.174 samples/sec
+epoch 13, total_step 205040, total loss is 15.25 , inference loss is 8.83, weight deacy loss is 6.42, training accuracy is 0.343750, time 216.191 samples/sec
+epoch 13, total_step 205060, total loss is 18.04 , inference loss is 11.61, weight deacy loss is 6.42, training accuracy is 0.281250, time 225.659 samples/sec
+epoch 13, total_step 205080, total loss is 10.47 , inference loss is 4.05, weight deacy loss is 6.42, training accuracy is 0.625000, time 219.869 samples/sec
+epoch 13, total_step 205100, total loss is 15.32 , inference loss is 8.89, weight deacy loss is 6.42, training accuracy is 0.406250, time 301.155 samples/sec
+epoch 13, total_step 205120, total loss is 13.95 , inference loss is 7.52, weight deacy loss is 6.42, training accuracy is 0.500000, time 222.865 samples/sec
+epoch 13, total_step 205140, total loss is 13.56 , inference loss is 7.14, weight deacy loss is 6.42, training accuracy is 0.468750, time 226.279 samples/sec
+epoch 13, total_step 205160, total loss is 15.23 , inference loss is 8.81, weight deacy loss is 6.42, training accuracy is 0.437500, time 215.512 samples/sec
+epoch 13, total_step 205180, total loss is 10.89 , inference loss is 4.47, weight deacy loss is 6.42, training accuracy is 0.656250, time 224.906 samples/sec
+epoch 13, total_step 205200, total loss is 12.37 , inference loss is 5.95, weight deacy loss is 6.42, training accuracy is 0.468750, time 12.231 samples/sec
+epoch 13, total_step 205220, total loss is 13.91 , inference loss is 7.48, weight deacy loss is 6.42, training accuracy is 0.406250, time 229.148 samples/sec
+epoch 13, total_step 205240, total loss is 13.98 , inference loss is 7.56, weight deacy loss is 6.42, training accuracy is 0.437500, time 226.090 samples/sec
+epoch 13, total_step 205260, total loss is 14.91 , inference loss is 8.49, weight deacy loss is 6.42, training accuracy is 0.562500, time 220.402 samples/sec
+epoch 13, total_step 205280, total loss is 15.16 , inference loss is 8.74, weight deacy loss is 6.42, training accuracy is 0.406250, time 219.308 samples/sec
+epoch 13, total_step 205300, total loss is 12.17 , inference loss is 5.74, weight deacy loss is 6.42, training accuracy is 0.343750, time 223.381 samples/sec
+epoch 13, total_step 205320, total loss is 13.73 , inference loss is 7.31, weight deacy loss is 6.42, training accuracy is 0.437500, time 217.291 samples/sec
+epoch 13, total_step 205340, total loss is 10.71 , inference loss is 4.29, weight deacy loss is 6.42, training accuracy is 0.625000, time 219.009 samples/sec
+epoch 13, total_step 205360, total loss is 10.07 , inference loss is 3.64, weight deacy loss is 6.42, training accuracy is 0.718750, time 215.450 samples/sec
+epoch 13, total_step 205380, total loss is 13.84 , inference loss is 7.42, weight deacy loss is 6.42, training accuracy is 0.468750, time 219.329 samples/sec
+epoch 13, total_step 205400, total loss is 17.41 , inference loss is 10.99, weight deacy loss is 6.42, training accuracy is 0.343750, time 296.058 samples/sec
+epoch 13, total_step 205420, total loss is 11.95 , inference loss is 5.52, weight deacy loss is 6.42, training accuracy is 0.531250, time 299.807 samples/sec
+epoch 13, total_step 205440, total loss is 11.09 , inference loss is 4.67, weight deacy loss is 6.42, training accuracy is 0.687500, time 219.390 samples/sec
+epoch 13, total_step 205460, total loss is 12.40 , inference loss is 5.98, weight deacy loss is 6.42, training accuracy is 0.625000, time 231.254 samples/sec
+epoch 13, total_step 205480, total loss is 13.29 , inference loss is 6.87, weight deacy loss is 6.42, training accuracy is 0.531250, time 212.581 samples/sec
+epoch 13, total_step 205500, total loss is 14.36 , inference loss is 7.94, weight deacy loss is 6.42, training accuracy is 0.468750, time 12.242 samples/sec
+epoch 13, total_step 205520, total loss is 11.38 , inference loss is 4.96, weight deacy loss is 6.42, training accuracy is 0.500000, time 229.998 samples/sec
+epoch 13, total_step 205540, total loss is 11.41 , inference loss is 4.99, weight deacy loss is 6.42, training accuracy is 0.562500, time 219.883 samples/sec
+epoch 13, total_step 205560, total loss is 10.88 , inference loss is 4.46, weight deacy loss is 6.42, training accuracy is 0.593750, time 219.612 samples/sec
+epoch 13, total_step 205580, total loss is 14.65 , inference loss is 8.23, weight deacy loss is 6.42, training accuracy is 0.531250, time 221.534 samples/sec
+epoch 13, total_step 205600, total loss is 13.11 , inference loss is 6.69, weight deacy loss is 6.42, training accuracy is 0.500000, time 220.437 samples/sec
+epoch 13, total_step 205620, total loss is 11.16 , inference loss is 4.74, weight deacy loss is 6.42, training accuracy is 0.531250, time 212.927 samples/sec
+epoch 13, total_step 205640, total loss is 15.03 , inference loss is 8.61, weight deacy loss is 6.42, training accuracy is 0.281250, time 237.604 samples/sec
+epoch 13, total_step 205660, total loss is 12.99 , inference loss is 6.57, weight deacy loss is 6.42, training accuracy is 0.468750, time 219.205 samples/sec
+epoch 13, total_step 205680, total loss is 11.04 , inference loss is 4.62, weight deacy loss is 6.42, training accuracy is 0.468750, time 221.676 samples/sec
+epoch 13, total_step 205700, total loss is 11.31 , inference loss is 4.89, weight deacy loss is 6.42, training accuracy is 0.562500, time 218.302 samples/sec
+epoch 13, total_step 205720, total loss is 13.69 , inference loss is 7.27, weight deacy loss is 6.42, training accuracy is 0.562500, time 228.725 samples/sec
+epoch 13, total_step 205740, total loss is 19.66 , inference loss is 13.24, weight deacy loss is 6.42, training accuracy is 0.281250, time 221.096 samples/sec
+epoch 13, total_step 205760, total loss is 12.85 , inference loss is 6.43, weight deacy loss is 6.42, training accuracy is 0.468750, time 217.371 samples/sec
+epoch 13, total_step 205780, total loss is 11.79 , inference loss is 5.37, weight deacy loss is 6.42, training accuracy is 0.468750, time 221.705 samples/sec
+epoch 13, total_step 205800, total loss is 12.77 , inference loss is 6.35, weight deacy loss is 6.42, training accuracy is 0.468750, time 12.002 samples/sec
+epoch 13, total_step 205820, total loss is 11.77 , inference loss is 5.35, weight deacy loss is 6.42, training accuracy is 0.593750, time 221.374 samples/sec
+epoch 13, total_step 205840, total loss is 17.56 , inference loss is 11.14, weight deacy loss is 6.42, training accuracy is 0.437500, time 217.299 samples/sec
+epoch 13, total_step 205860, total loss is 14.92 , inference loss is 8.50, weight deacy loss is 6.42, training accuracy is 0.437500, time 215.294 samples/sec
+epoch 13, total_step 205880, total loss is 13.43 , inference loss is 7.01, weight deacy loss is 6.42, training accuracy is 0.468750, time 224.768 samples/sec
+epoch 13, total_step 205900, total loss is 11.79 , inference loss is 5.37, weight deacy loss is 6.42, training accuracy is 0.531250, time 214.594 samples/sec
+epoch 13, total_step 205920, total loss is 16.85 , inference loss is 10.43, weight deacy loss is 6.42, training accuracy is 0.312500, time 225.310 samples/sec
+epoch 13, total_step 205940, total loss is 13.02 , inference loss is 6.60, weight deacy loss is 6.42, training accuracy is 0.500000, time 219.250 samples/sec
+epoch 13, total_step 205960, total loss is 12.39 , inference loss is 5.97, weight deacy loss is 6.42, training accuracy is 0.437500, time 224.506 samples/sec
+epoch 13, total_step 205980, total loss is 13.26 , inference loss is 6.84, weight deacy loss is 6.42, training accuracy is 0.500000, time 216.045 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.786638000000018
+best_threshold_index 110 0.9861111111111112
+best_threshold_index 111 0.9866666666666667
+best_threshold_index 111 0.9853703703703703
+best_threshold_index 107 0.9874074074074074
+best_threshold_index 122 0.9862962962962963
+best_threshold_index 111 0.9848148148148148
+best_threshold_index 111 0.9857407407407407
+best_threshold_index 111 0.985
+best_threshold_index 111 0.9846296296296296
+best_threshold_index 111 0.985
+[lfw][206000]XNorm: 22.059153
+[lfw][206000]Accuracy-Flip: 0.98383+-0.00969
+test accuracy is: 0.9838333333333333
+epoch 13, total_step 206000, total loss is 12.45 , inference loss is 6.03, weight deacy loss is 6.42, training accuracy is 0.593750, time 16.600 samples/sec
+epoch 13, total_step 206020, total loss is 16.12 , inference loss is 9.70, weight deacy loss is 6.42, training accuracy is 0.468750, time 213.482 samples/sec
+epoch 13, total_step 206040, total loss is 11.44 , inference loss is 5.02, weight deacy loss is 6.42, training accuracy is 0.468750, time 211.017 samples/sec
+epoch 13, total_step 206060, total loss is 12.89 , inference loss is 6.47, weight deacy loss is 6.42, training accuracy is 0.562500, time 219.868 samples/sec
+epoch 13, total_step 206080, total loss is 10.57 , inference loss is 4.15, weight deacy loss is 6.42, training accuracy is 0.687500, time 219.554 samples/sec
+epoch 13, total_step 206100, total loss is 12.68 , inference loss is 6.26, weight deacy loss is 6.42, training accuracy is 0.656250, time 13.588 samples/sec
+epoch 13, total_step 206120, total loss is 10.57 , inference loss is 4.15, weight deacy loss is 6.42, training accuracy is 0.625000, time 211.595 samples/sec
+epoch 13, total_step 206140, total loss is 17.02 , inference loss is 10.60, weight deacy loss is 6.42, training accuracy is 0.468750, time 219.140 samples/sec
+epoch 13, total_step 206160, total loss is 13.67 , inference loss is 7.25, weight deacy loss is 6.42, training accuracy is 0.500000, time 232.820 samples/sec
+epoch 13, total_step 206180, total loss is 12.42 , inference loss is 6.00, weight deacy loss is 6.42, training accuracy is 0.343750, time 219.273 samples/sec
+epoch 13, total_step 206200, total loss is 17.21 , inference loss is 10.79, weight deacy loss is 6.42, training accuracy is 0.468750, time 233.410 samples/sec
+epoch 13, total_step 206220, total loss is 15.47 , inference loss is 9.05, weight deacy loss is 6.42, training accuracy is 0.437500, time 221.122 samples/sec
+epoch 13, total_step 206240, total loss is 13.66 , inference loss is 7.24, weight deacy loss is 6.42, training accuracy is 0.375000, time 225.257 samples/sec
+epoch 13, total_step 206260, total loss is 13.45 , inference loss is 7.03, weight deacy loss is 6.42, training accuracy is 0.468750, time 220.317 samples/sec
+epoch 13, total_step 206280, total loss is 13.13 , inference loss is 6.71, weight deacy loss is 6.42, training accuracy is 0.375000, time 216.370 samples/sec
+epoch 13, total_step 206300, total loss is 13.30 , inference loss is 6.88, weight deacy loss is 6.42, training accuracy is 0.500000, time 221.964 samples/sec
+epoch 13, total_step 206320, total loss is 9.44 , inference loss is 3.02, weight deacy loss is 6.42, training accuracy is 0.500000, time 221.971 samples/sec
+epoch 13, total_step 206340, total loss is 14.07 , inference loss is 7.65, weight deacy loss is 6.42, training accuracy is 0.500000, time 219.876 samples/sec
+epoch 13, total_step 206360, total loss is 14.02 , inference loss is 7.60, weight deacy loss is 6.42, training accuracy is 0.656250, time 216.522 samples/sec
+epoch 13, total_step 206380, total loss is 11.02 , inference loss is 4.60, weight deacy loss is 6.42, training accuracy is 0.500000, time 233.350 samples/sec
+epoch 13, total_step 206400, total loss is 16.52 , inference loss is 10.10, weight deacy loss is 6.42, training accuracy is 0.500000, time 12.450 samples/sec
+epoch 13, total_step 206420, total loss is 13.63 , inference loss is 7.21, weight deacy loss is 6.42, training accuracy is 0.375000, time 226.350 samples/sec
+epoch 13, total_step 206440, total loss is 14.03 , inference loss is 7.62, weight deacy loss is 6.42, training accuracy is 0.406250, time 225.669 samples/sec
+epoch 13, total_step 206460, total loss is 10.90 , inference loss is 4.49, weight deacy loss is 6.42, training accuracy is 0.437500, time 219.288 samples/sec
+epoch 13, total_step 206480, total loss is 12.38 , inference loss is 5.96, weight deacy loss is 6.42, training accuracy is 0.625000, time 213.326 samples/sec
+epoch 13, total_step 206500, total loss is 13.73 , inference loss is 7.31, weight deacy loss is 6.42, training accuracy is 0.562500, time 218.612 samples/sec
+epoch 13, total_step 206520, total loss is 14.32 , inference loss is 7.90, weight deacy loss is 6.42, training accuracy is 0.468750, time 217.028 samples/sec
+epoch 13, total_step 206540, total loss is 12.76 , inference loss is 6.34, weight deacy loss is 6.42, training accuracy is 0.468750, time 221.123 samples/sec
+epoch 13, total_step 206560, total loss is 15.75 , inference loss is 9.33, weight deacy loss is 6.42, training accuracy is 0.312500, time 221.378 samples/sec
+epoch 13, total_step 206580, total loss is 11.46 , inference loss is 5.05, weight deacy loss is 6.42, training accuracy is 0.625000, time 213.106 samples/sec
+epoch 13, total_step 206600, total loss is 10.53 , inference loss is 4.11, weight deacy loss is 6.42, training accuracy is 0.625000, time 213.637 samples/sec
+epoch 13, total_step 206620, total loss is 17.53 , inference loss is 11.11, weight deacy loss is 6.42, training accuracy is 0.406250, time 223.904 samples/sec
+epoch 13, total_step 206640, total loss is 16.53 , inference loss is 10.12, weight deacy loss is 6.42, training accuracy is 0.437500, time 187.975 samples/sec
+epoch 13, total_step 206660, total loss is 12.93 , inference loss is 6.51, weight deacy loss is 6.42, training accuracy is 0.500000, time 217.151 samples/sec
+epoch 13, total_step 206680, total loss is 11.13 , inference loss is 4.71, weight deacy loss is 6.42, training accuracy is 0.562500, time 295.791 samples/sec
+epoch 13, total_step 206700, total loss is 12.16 , inference loss is 5.74, weight deacy loss is 6.42, training accuracy is 0.562500, time 12.288 samples/sec
+epoch 13, total_step 206720, total loss is 13.16 , inference loss is 6.74, weight deacy loss is 6.42, training accuracy is 0.593750, time 205.764 samples/sec
+epoch 13, total_step 206740, total loss is 12.69 , inference loss is 6.27, weight deacy loss is 6.42, training accuracy is 0.500000, time 215.197 samples/sec
+epoch 13, total_step 206760, total loss is 15.53 , inference loss is 9.11, weight deacy loss is 6.42, training accuracy is 0.437500, time 217.807 samples/sec
+epoch 13, total_step 206780, total loss is 12.01 , inference loss is 5.59, weight deacy loss is 6.42, training accuracy is 0.593750, time 223.748 samples/sec
+epoch 13, total_step 206800, total loss is 9.86 , inference loss is 3.45, weight deacy loss is 6.42, training accuracy is 0.562500, time 223.699 samples/sec
+epoch 13, total_step 206820, total loss is 9.95 , inference loss is 3.54, weight deacy loss is 6.42, training accuracy is 0.500000, time 237.445 samples/sec
+epoch 13, total_step 206840, total loss is 12.27 , inference loss is 5.86, weight deacy loss is 6.42, training accuracy is 0.562500, time 219.068 samples/sec
+epoch 13, total_step 206860, total loss is 13.52 , inference loss is 7.11, weight deacy loss is 6.42, training accuracy is 0.500000, time 205.739 samples/sec
+epoch 13, total_step 206880, total loss is 13.64 , inference loss is 7.22, weight deacy loss is 6.42, training accuracy is 0.406250, time 217.464 samples/sec
+epoch 13, total_step 206900, total loss is 11.07 , inference loss is 4.65, weight deacy loss is 6.42, training accuracy is 0.500000, time 230.190 samples/sec
+epoch 13, total_step 206920, total loss is 15.07 , inference loss is 8.66, weight deacy loss is 6.42, training accuracy is 0.500000, time 224.847 samples/sec
+epoch 13, total_step 206940, total loss is 12.31 , inference loss is 5.89, weight deacy loss is 6.42, training accuracy is 0.562500, time 294.517 samples/sec
+epoch 13, total_step 206960, total loss is 14.20 , inference loss is 7.79, weight deacy loss is 6.42, training accuracy is 0.500000, time 211.828 samples/sec
+epoch 13, total_step 206980, total loss is 10.00 , inference loss is 3.58, weight deacy loss is 6.42, training accuracy is 0.593750, time 229.383 samples/sec
+epoch 13, total_step 207000, total loss is 10.95 , inference loss is 4.53, weight deacy loss is 6.42, training accuracy is 0.625000, time 13.438 samples/sec
+epoch 13, total_step 207020, total loss is 14.55 , inference loss is 8.14, weight deacy loss is 6.42, training accuracy is 0.437500, time 213.416 samples/sec
+epoch 13, total_step 207040, total loss is 11.22 , inference loss is 4.80, weight deacy loss is 6.42, training accuracy is 0.500000, time 215.395 samples/sec
+epoch 13, total_step 207060, total loss is 13.76 , inference loss is 7.35, weight deacy loss is 6.42, training accuracy is 0.531250, time 211.711 samples/sec
+epoch 13, total_step 207080, total loss is 10.42 , inference loss is 4.00, weight deacy loss is 6.42, training accuracy is 0.562500, time 213.412 samples/sec
+epoch 13, total_step 207100, total loss is 11.05 , inference loss is 4.64, weight deacy loss is 6.41, training accuracy is 0.593750, time 212.605 samples/sec
+epoch 13, total_step 207120, total loss is 16.60 , inference loss is 10.18, weight deacy loss is 6.41, training accuracy is 0.437500, time 226.019 samples/sec
+epoch 13, total_step 207140, total loss is 10.52 , inference loss is 4.11, weight deacy loss is 6.41, training accuracy is 0.562500, time 298.229 samples/sec
+epoch 13, total_step 207160, total loss is 16.13 , inference loss is 9.71, weight deacy loss is 6.41, training accuracy is 0.531250, time 276.174 samples/sec
+epoch 13, total_step 207180, total loss is 10.10 , inference loss is 3.69, weight deacy loss is 6.41, training accuracy is 0.593750, time 223.682 samples/sec
+epoch 13, total_step 207200, total loss is 13.13 , inference loss is 6.72, weight deacy loss is 6.41, training accuracy is 0.468750, time 221.650 samples/sec
+epoch 13, total_step 207220, total loss is 15.79 , inference loss is 9.37, weight deacy loss is 6.41, training accuracy is 0.531250, time 214.784 samples/sec
+epoch 13, total_step 207240, total loss is 11.57 , inference loss is 5.16, weight deacy loss is 6.41, training accuracy is 0.531250, time 225.537 samples/sec
+epoch 13, total_step 207260, total loss is 14.98 , inference loss is 8.56, weight deacy loss is 6.41, training accuracy is 0.468750, time 228.420 samples/sec
+epoch 13, total_step 207280, total loss is 11.86 , inference loss is 5.44, weight deacy loss is 6.41, training accuracy is 0.562500, time 221.235 samples/sec
+epoch 13, total_step 207300, total loss is 14.79 , inference loss is 8.37, weight deacy loss is 6.41, training accuracy is 0.437500, time 12.210 samples/sec
+epoch 13, total_step 207320, total loss is 13.73 , inference loss is 7.31, weight deacy loss is 6.41, training accuracy is 0.437500, time 297.815 samples/sec
+epoch 13, total_step 207340, total loss is 10.55 , inference loss is 4.14, weight deacy loss is 6.41, training accuracy is 0.625000, time 216.764 samples/sec
+epoch 13, total_step 207360, total loss is 15.44 , inference loss is 9.02, weight deacy loss is 6.41, training accuracy is 0.531250, time 221.296 samples/sec
+epoch 13, total_step 207380, total loss is 16.28 , inference loss is 9.87, weight deacy loss is 6.41, training accuracy is 0.437500, time 219.954 samples/sec
+epoch 13, total_step 207400, total loss is 12.75 , inference loss is 6.34, weight deacy loss is 6.41, training accuracy is 0.437500, time 212.349 samples/sec
+epoch 13, total_step 207420, total loss is 13.69 , inference loss is 7.28, weight deacy loss is 6.41, training accuracy is 0.437500, time 218.775 samples/sec
+epoch 13, total_step 207440, total loss is 13.40 , inference loss is 6.99, weight deacy loss is 6.41, training accuracy is 0.593750, time 224.473 samples/sec
+epoch 13, total_step 207460, total loss is 12.24 , inference loss is 5.83, weight deacy loss is 6.41, training accuracy is 0.500000, time 241.490 samples/sec
+epoch 13, total_step 207480, total loss is 11.41 , inference loss is 5.00, weight deacy loss is 6.41, training accuracy is 0.468750, time 237.403 samples/sec
+epoch 13, total_step 207500, total loss is 12.15 , inference loss is 5.73, weight deacy loss is 6.41, training accuracy is 0.406250, time 235.432 samples/sec
+epoch 13, total_step 207520, total loss is 13.16 , inference loss is 6.75, weight deacy loss is 6.41, training accuracy is 0.593750, time 215.650 samples/sec
+epoch 13, total_step 207540, total loss is 10.00 , inference loss is 3.59, weight deacy loss is 6.41, training accuracy is 0.625000, time 202.737 samples/sec
+epoch 13, total_step 207560, total loss is 15.01 , inference loss is 8.59, weight deacy loss is 6.41, training accuracy is 0.531250, time 200.803 samples/sec
+epoch 13, total_step 207580, total loss is 17.56 , inference loss is 11.15, weight deacy loss is 6.41, training accuracy is 0.500000, time 235.887 samples/sec
+epoch 13, total_step 207600, total loss is 15.90 , inference loss is 9.48, weight deacy loss is 6.41, training accuracy is 0.531250, time 12.239 samples/sec
+epoch 13, total_step 207620, total loss is 14.02 , inference loss is 7.60, weight deacy loss is 6.41, training accuracy is 0.593750, time 216.355 samples/sec
+epoch 13, total_step 207640, total loss is 12.69 , inference loss is 6.27, weight deacy loss is 6.41, training accuracy is 0.375000, time 225.450 samples/sec
+epoch 13, total_step 207660, total loss is 14.36 , inference loss is 7.94, weight deacy loss is 6.41, training accuracy is 0.437500, time 219.849 samples/sec
+epoch 13, total_step 207680, total loss is 10.93 , inference loss is 4.52, weight deacy loss is 6.41, training accuracy is 0.593750, time 214.312 samples/sec
+epoch 13, total_step 207700, total loss is 12.56 , inference loss is 6.15, weight deacy loss is 6.41, training accuracy is 0.593750, time 216.454 samples/sec
+epoch 13, total_step 207720, total loss is 12.55 , inference loss is 6.13, weight deacy loss is 6.41, training accuracy is 0.343750, time 234.664 samples/sec
+epoch 13, total_step 207740, total loss is 13.15 , inference loss is 6.74, weight deacy loss is 6.41, training accuracy is 0.500000, time 221.727 samples/sec
+epoch 13, total_step 207760, total loss is 14.94 , inference loss is 8.52, weight deacy loss is 6.41, training accuracy is 0.375000, time 221.146 samples/sec
+epoch 13, total_step 207780, total loss is 15.42 , inference loss is 9.01, weight deacy loss is 6.41, training accuracy is 0.500000, time 230.675 samples/sec
+epoch 13, total_step 207800, total loss is 10.64 , inference loss is 4.23, weight deacy loss is 6.41, training accuracy is 0.562500, time 224.579 samples/sec
+epoch 13, total_step 207820, total loss is 16.15 , inference loss is 9.74, weight deacy loss is 6.41, training accuracy is 0.437500, time 239.207 samples/sec
+epoch 13, total_step 207840, total loss is 15.50 , inference loss is 9.09, weight deacy loss is 6.41, training accuracy is 0.531250, time 213.018 samples/sec
+epoch 13, total_step 207860, total loss is 15.19 , inference loss is 8.78, weight deacy loss is 6.41, training accuracy is 0.468750, time 232.085 samples/sec
+epoch 13, total_step 207880, total loss is 17.77 , inference loss is 11.36, weight deacy loss is 6.41, training accuracy is 0.281250, time 188.339 samples/sec
+epoch 13, total_step 207900, total loss is 14.05 , inference loss is 7.64, weight deacy loss is 6.41, training accuracy is 0.437500, time 12.655 samples/sec
+epoch 13, total_step 207920, total loss is 17.83 , inference loss is 11.41, weight deacy loss is 6.41, training accuracy is 0.468750, time 230.869 samples/sec
+epoch 13, total_step 207940, total loss is 13.64 , inference loss is 7.23, weight deacy loss is 6.41, training accuracy is 0.500000, time 215.028 samples/sec
+epoch 13, total_step 207960, total loss is 12.98 , inference loss is 6.57, weight deacy loss is 6.41, training accuracy is 0.437500, time 182.629 samples/sec
+epoch 13, total_step 207980, total loss is 13.30 , inference loss is 6.89, weight deacy loss is 6.41, training accuracy is 0.343750, time 204.429 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.040967999999975
+best_threshold_index 112 0.9874074074074074
+best_threshold_index 115 0.9879629629629629
+best_threshold_index 116 0.9877777777777778
+best_threshold_index 112 0.9885185185185185
+best_threshold_index 116 0.9881481481481481
+best_threshold_index 117 0.9868518518518519
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 115 0.9868518518518519
+best_threshold_index 115 0.9862962962962963
+best_threshold_index 115 0.9866666666666667
+[lfw][208000]XNorm: 21.910983
+[lfw][208000]Accuracy-Flip: 0.98583+-0.00629
+test accuracy is: 0.9858333333333335
+epoch 13, total_step 208000, total loss is 14.10 , inference loss is 7.68, weight deacy loss is 6.41, training accuracy is 0.437500, time 15.573 samples/sec
+epoch 13, total_step 208020, total loss is 15.46 , inference loss is 9.05, weight deacy loss is 6.41, training accuracy is 0.468750, time 229.926 samples/sec
+epoch 13, total_step 208040, total loss is 16.71 , inference loss is 10.29, weight deacy loss is 6.41, training accuracy is 0.343750, time 222.276 samples/sec
+epoch 13, total_step 208060, total loss is 13.17 , inference loss is 6.76, weight deacy loss is 6.41, training accuracy is 0.468750, time 219.764 samples/sec
+epoch 13, total_step 208080, total loss is 10.46 , inference loss is 4.05, weight deacy loss is 6.41, training accuracy is 0.562500, time 240.472 samples/sec
+epoch 13, total_step 208100, total loss is 10.18 , inference loss is 3.76, weight deacy loss is 6.41, training accuracy is 0.468750, time 219.416 samples/sec
+epoch 13, total_step 208120, total loss is 14.23 , inference loss is 7.82, weight deacy loss is 6.41, training accuracy is 0.375000, time 223.287 samples/sec
+epoch 13, total_step 208140, total loss is 14.55 , inference loss is 8.14, weight deacy loss is 6.41, training accuracy is 0.406250, time 219.209 samples/sec
+epoch 13, total_step 208160, total loss is 12.31 , inference loss is 5.89, weight deacy loss is 6.41, training accuracy is 0.531250, time 218.546 samples/sec
+epoch 13, total_step 208180, total loss is 14.86 , inference loss is 8.44, weight deacy loss is 6.41, training accuracy is 0.437500, time 224.168 samples/sec
+epoch 13, total_step 208200, total loss is 13.16 , inference loss is 6.74, weight deacy loss is 6.41, training accuracy is 0.437500, time 12.387 samples/sec
+epoch 13, total_step 208220, total loss is 15.36 , inference loss is 8.94, weight deacy loss is 6.41, training accuracy is 0.500000, time 220.423 samples/sec
+epoch 13, total_step 208240, total loss is 13.74 , inference loss is 7.33, weight deacy loss is 6.41, training accuracy is 0.406250, time 218.251 samples/sec
+epoch 13, total_step 208260, total loss is 14.68 , inference loss is 8.26, weight deacy loss is 6.41, training accuracy is 0.250000, time 217.826 samples/sec
+epoch 13, total_step 208280, total loss is 15.40 , inference loss is 8.99, weight deacy loss is 6.41, training accuracy is 0.500000, time 220.898 samples/sec
+epoch 13, total_step 208300, total loss is 13.97 , inference loss is 7.55, weight deacy loss is 6.41, training accuracy is 0.468750, time 225.332 samples/sec
+epoch 13, total_step 208320, total loss is 12.08 , inference loss is 5.67, weight deacy loss is 6.41, training accuracy is 0.468750, time 233.894 samples/sec
+epoch 13, total_step 208340, total loss is 11.46 , inference loss is 5.05, weight deacy loss is 6.41, training accuracy is 0.468750, time 222.919 samples/sec
+epoch 13, total_step 208360, total loss is 13.67 , inference loss is 7.26, weight deacy loss is 6.41, training accuracy is 0.437500, time 218.234 samples/sec
+epoch 13, total_step 208380, total loss is 10.85 , inference loss is 4.43, weight deacy loss is 6.41, training accuracy is 0.656250, time 226.194 samples/sec
+epoch 13, total_step 208400, total loss is 15.95 , inference loss is 9.54, weight deacy loss is 6.41, training accuracy is 0.468750, time 222.400 samples/sec
+epoch 13, total_step 208420, total loss is 15.60 , inference loss is 9.19, weight deacy loss is 6.41, training accuracy is 0.500000, time 216.748 samples/sec
+epoch 13, total_step 208440, total loss is 14.59 , inference loss is 8.18, weight deacy loss is 6.41, training accuracy is 0.406250, time 210.387 samples/sec
+epoch 13, total_step 208460, total loss is 14.26 , inference loss is 7.85, weight deacy loss is 6.41, training accuracy is 0.437500, time 224.187 samples/sec
+epoch 13, total_step 208480, total loss is 12.49 , inference loss is 6.08, weight deacy loss is 6.41, training accuracy is 0.562500, time 229.569 samples/sec
+epoch 13, total_step 208500, total loss is 14.39 , inference loss is 7.98, weight deacy loss is 6.41, training accuracy is 0.468750, time 12.467 samples/sec
+epoch 13, total_step 208520, total loss is 13.37 , inference loss is 6.96, weight deacy loss is 6.41, training accuracy is 0.406250, time 221.421 samples/sec
+epoch 13, total_step 208540, total loss is 13.20 , inference loss is 6.79, weight deacy loss is 6.41, training accuracy is 0.437500, time 218.599 samples/sec
+epoch 13, total_step 208560, total loss is 21.19 , inference loss is 14.78, weight deacy loss is 6.41, training accuracy is 0.156250, time 229.593 samples/sec
+epoch 13, total_step 208580, total loss is 15.46 , inference loss is 9.05, weight deacy loss is 6.41, training accuracy is 0.375000, time 222.406 samples/sec
+epoch 13, total_step 208600, total loss is 14.77 , inference loss is 8.36, weight deacy loss is 6.41, training accuracy is 0.375000, time 219.365 samples/sec
+epoch 13, total_step 208620, total loss is 11.76 , inference loss is 5.34, weight deacy loss is 6.41, training accuracy is 0.531250, time 224.090 samples/sec
+epoch 13, total_step 208640, total loss is 13.78 , inference loss is 7.37, weight deacy loss is 6.41, training accuracy is 0.468750, time 222.726 samples/sec
+epoch 13, total_step 208660, total loss is 11.58 , inference loss is 5.17, weight deacy loss is 6.41, training accuracy is 0.593750, time 214.868 samples/sec
+epoch 13, total_step 208680, total loss is 11.79 , inference loss is 5.37, weight deacy loss is 6.41, training accuracy is 0.593750, time 223.301 samples/sec
+epoch 13, total_step 208700, total loss is 12.61 , inference loss is 6.20, weight deacy loss is 6.41, training accuracy is 0.531250, time 217.621 samples/sec
+epoch 13, total_step 208720, total loss is 11.38 , inference loss is 4.97, weight deacy loss is 6.41, training accuracy is 0.593750, time 220.875 samples/sec
+epoch 13, total_step 208740, total loss is 14.21 , inference loss is 7.79, weight deacy loss is 6.41, training accuracy is 0.406250, time 212.324 samples/sec
+epoch 13, total_step 208760, total loss is 10.58 , inference loss is 4.17, weight deacy loss is 6.41, training accuracy is 0.468750, time 221.169 samples/sec
+epoch 13, total_step 208780, total loss is 15.49 , inference loss is 9.08, weight deacy loss is 6.41, training accuracy is 0.593750, time 295.735 samples/sec
+epoch 13, total_step 208800, total loss is 13.90 , inference loss is 7.49, weight deacy loss is 6.41, training accuracy is 0.500000, time 13.426 samples/sec
+epoch 13, total_step 208820, total loss is 11.66 , inference loss is 5.24, weight deacy loss is 6.41, training accuracy is 0.500000, time 218.983 samples/sec
+epoch 13, total_step 208840, total loss is 14.63 , inference loss is 8.22, weight deacy loss is 6.41, training accuracy is 0.593750, time 214.367 samples/sec
+epoch 13, total_step 208860, total loss is 12.12 , inference loss is 5.71, weight deacy loss is 6.41, training accuracy is 0.437500, time 216.206 samples/sec
+epoch 13, total_step 208880, total loss is 12.84 , inference loss is 6.43, weight deacy loss is 6.41, training accuracy is 0.593750, time 214.832 samples/sec
+epoch 13, total_step 208900, total loss is 15.05 , inference loss is 8.64, weight deacy loss is 6.41, training accuracy is 0.437500, time 231.748 samples/sec
+epoch 13, total_step 208920, total loss is 15.14 , inference loss is 8.73, weight deacy loss is 6.41, training accuracy is 0.437500, time 295.810 samples/sec
+epoch 13, total_step 208940, total loss is 10.33 , inference loss is 3.92, weight deacy loss is 6.41, training accuracy is 0.656250, time 299.557 samples/sec
+epoch 13, total_step 208960, total loss is 14.10 , inference loss is 7.69, weight deacy loss is 6.41, training accuracy is 0.468750, time 225.934 samples/sec
+epoch 13, total_step 208980, total loss is 11.88 , inference loss is 5.47, weight deacy loss is 6.41, training accuracy is 0.500000, time 218.926 samples/sec
+epoch 13, total_step 209000, total loss is 12.93 , inference loss is 6.52, weight deacy loss is 6.41, training accuracy is 0.562500, time 219.838 samples/sec
+epoch 13, total_step 209020, total loss is 11.06 , inference loss is 4.65, weight deacy loss is 6.41, training accuracy is 0.562500, time 217.001 samples/sec
+epoch 13, total_step 209040, total loss is 11.49 , inference loss is 5.08, weight deacy loss is 6.41, training accuracy is 0.468750, time 224.613 samples/sec
+epoch 13, total_step 209060, total loss is 13.47 , inference loss is 7.06, weight deacy loss is 6.41, training accuracy is 0.468750, time 245.616 samples/sec
+epoch 13, total_step 209080, total loss is 14.74 , inference loss is 8.33, weight deacy loss is 6.41, training accuracy is 0.531250, time 227.584 samples/sec
+epoch 13, total_step 209100, total loss is 14.54 , inference loss is 8.13, weight deacy loss is 6.41, training accuracy is 0.562500, time 12.289 samples/sec
+epoch 13, total_step 209120, total loss is 12.97 , inference loss is 6.56, weight deacy loss is 6.41, training accuracy is 0.531250, time 225.630 samples/sec
+epoch 13, total_step 209140, total loss is 16.56 , inference loss is 10.15, weight deacy loss is 6.41, training accuracy is 0.406250, time 221.709 samples/sec
+epoch 13, total_step 209160, total loss is 14.12 , inference loss is 7.71, weight deacy loss is 6.41, training accuracy is 0.468750, time 219.136 samples/sec
+epoch 13, total_step 209180, total loss is 13.35 , inference loss is 6.94, weight deacy loss is 6.41, training accuracy is 0.531250, time 223.907 samples/sec
+epoch 13, total_step 209200, total loss is 12.82 , inference loss is 6.41, weight deacy loss is 6.41, training accuracy is 0.562500, time 296.809 samples/sec
+epoch 13, total_step 209220, total loss is 12.56 , inference loss is 6.15, weight deacy loss is 6.41, training accuracy is 0.468750, time 217.711 samples/sec
+epoch 13, total_step 209240, total loss is 11.77 , inference loss is 5.37, weight deacy loss is 6.41, training accuracy is 0.500000, time 215.239 samples/sec
+epoch 13, total_step 209260, total loss is 9.48 , inference loss is 3.07, weight deacy loss is 6.41, training accuracy is 0.593750, time 222.894 samples/sec
+epoch 13, total_step 209280, total loss is 11.55 , inference loss is 5.14, weight deacy loss is 6.41, training accuracy is 0.562500, time 230.255 samples/sec
+epoch 13, total_step 209300, total loss is 9.84 , inference loss is 3.43, weight deacy loss is 6.41, training accuracy is 0.687500, time 239.411 samples/sec
+epoch 13, total_step 209320, total loss is 14.40 , inference loss is 8.00, weight deacy loss is 6.41, training accuracy is 0.593750, time 230.182 samples/sec
+epoch 13, total_step 209340, total loss is 11.52 , inference loss is 5.12, weight deacy loss is 6.41, training accuracy is 0.562500, time 216.421 samples/sec
+epoch 13, total_step 209360, total loss is 11.45 , inference loss is 5.04, weight deacy loss is 6.41, training accuracy is 0.593750, time 218.060 samples/sec
+epoch 13, total_step 209380, total loss is 12.96 , inference loss is 6.55, weight deacy loss is 6.41, training accuracy is 0.468750, time 225.919 samples/sec
+epoch 13, total_step 209400, total loss is 16.97 , inference loss is 10.56, weight deacy loss is 6.41, training accuracy is 0.500000, time 12.259 samples/sec
+epoch 13, total_step 209420, total loss is 11.83 , inference loss is 5.42, weight deacy loss is 6.41, training accuracy is 0.468750, time 217.871 samples/sec
+epoch 13, total_step 209440, total loss is 11.61 , inference loss is 5.21, weight deacy loss is 6.41, training accuracy is 0.437500, time 217.952 samples/sec
+epoch 13, total_step 209460, total loss is 11.82 , inference loss is 5.42, weight deacy loss is 6.41, training accuracy is 0.562500, time 299.243 samples/sec
+epoch 13, total_step 209480, total loss is 11.17 , inference loss is 4.76, weight deacy loss is 6.41, training accuracy is 0.593750, time 219.362 samples/sec
+epoch 13, total_step 209500, total loss is 16.63 , inference loss is 10.23, weight deacy loss is 6.41, training accuracy is 0.500000, time 217.327 samples/sec
+epoch 13, total_step 209520, total loss is 11.32 , inference loss is 4.91, weight deacy loss is 6.41, training accuracy is 0.500000, time 224.392 samples/sec
+epoch 13, total_step 209540, total loss is 11.67 , inference loss is 5.26, weight deacy loss is 6.41, training accuracy is 0.562500, time 240.115 samples/sec
+epoch 13, total_step 209560, total loss is 15.97 , inference loss is 9.57, weight deacy loss is 6.41, training accuracy is 0.312500, time 216.274 samples/sec
+epoch 13, total_step 209580, total loss is 14.99 , inference loss is 8.58, weight deacy loss is 6.41, training accuracy is 0.500000, time 214.310 samples/sec
+epoch 13, total_step 209600, total loss is 11.87 , inference loss is 5.46, weight deacy loss is 6.41, training accuracy is 0.468750, time 220.833 samples/sec
+epoch 13, total_step 209620, total loss is 10.66 , inference loss is 4.25, weight deacy loss is 6.41, training accuracy is 0.531250, time 222.775 samples/sec
+epoch 13, total_step 209640, total loss is 13.11 , inference loss is 6.70, weight deacy loss is 6.41, training accuracy is 0.500000, time 215.104 samples/sec
+epoch 13, total_step 209660, total loss is 12.55 , inference loss is 6.14, weight deacy loss is 6.41, training accuracy is 0.562500, time 222.139 samples/sec
+epoch 13, total_step 209680, total loss is 12.09 , inference loss is 5.68, weight deacy loss is 6.41, training accuracy is 0.625000, time 226.761 samples/sec
+epoch 13, total_step 209700, total loss is 14.03 , inference loss is 7.62, weight deacy loss is 6.41, training accuracy is 0.500000, time 12.788 samples/sec
+epoch 13, total_step 209720, total loss is 12.95 , inference loss is 6.55, weight deacy loss is 6.41, training accuracy is 0.625000, time 295.713 samples/sec
+epoch 13, total_step 209740, total loss is 12.63 , inference loss is 6.23, weight deacy loss is 6.41, training accuracy is 0.562500, time 213.302 samples/sec
+epoch 13, total_step 209760, total loss is 13.33 , inference loss is 6.93, weight deacy loss is 6.41, training accuracy is 0.468750, time 235.145 samples/sec
+epoch 13, total_step 209780, total loss is 13.29 , inference loss is 6.89, weight deacy loss is 6.41, training accuracy is 0.562500, time 236.203 samples/sec
+epoch 13, total_step 209800, total loss is 13.96 , inference loss is 7.55, weight deacy loss is 6.41, training accuracy is 0.625000, time 217.219 samples/sec
+epoch 13, total_step 209820, total loss is 9.40 , inference loss is 2.99, weight deacy loss is 6.41, training accuracy is 0.718750, time 240.653 samples/sec
+epoch 13, total_step 209840, total loss is 14.15 , inference loss is 7.75, weight deacy loss is 6.41, training accuracy is 0.625000, time 216.389 samples/sec
+epoch 13, total_step 209860, total loss is 11.21 , inference loss is 4.80, weight deacy loss is 6.41, training accuracy is 0.500000, time 225.131 samples/sec
+epoch 13, total_step 209880, total loss is 17.66 , inference loss is 11.26, weight deacy loss is 6.41, training accuracy is 0.375000, time 235.051 samples/sec
+epoch 13, total_step 209900, total loss is 11.67 , inference loss is 5.27, weight deacy loss is 6.41, training accuracy is 0.562500, time 228.275 samples/sec
+epoch 13, total_step 209920, total loss is 12.63 , inference loss is 6.22, weight deacy loss is 6.41, training accuracy is 0.406250, time 229.287 samples/sec
+epoch 13, total_step 209940, total loss is 12.38 , inference loss is 5.97, weight deacy loss is 6.41, training accuracy is 0.593750, time 225.660 samples/sec
+epoch 13, total_step 209960, total loss is 13.00 , inference loss is 6.59, weight deacy loss is 6.40, training accuracy is 0.500000, time 222.407 samples/sec
+epoch 13, total_step 209980, total loss is 11.93 , inference loss is 5.52, weight deacy loss is 6.40, training accuracy is 0.531250, time 233.374 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.71724600000003
+best_threshold_index 113 0.9877777777777778
+best_threshold_index 112 0.9864814814814815
+best_threshold_index 107 0.987037037037037
+best_threshold_index 112 0.9877777777777778
+best_threshold_index 110 0.9872222222222222
+best_threshold_index 112 0.9859259259259259
+best_threshold_index 112 0.9866666666666667
+best_threshold_index 107 0.9862962962962963
+best_threshold_index 112 0.9857407407407407
+best_threshold_index 110 0.9862962962962963
+[lfw][210000]XNorm: 21.479118
+[lfw][210000]Accuracy-Flip: 0.98517+-0.00621
+test accuracy is: 0.9851666666666665
+epoch 13, total_step 210000, total loss is 13.02 , inference loss is 6.62, weight deacy loss is 6.40, training accuracy is 0.593750, time 16.320 samples/sec
+epoch 13, total_step 210020, total loss is 13.63 , inference loss is 7.23, weight deacy loss is 6.40, training accuracy is 0.343750, time 218.975 samples/sec
+epoch 13, total_step 210040, total loss is 12.09 , inference loss is 5.69, weight deacy loss is 6.40, training accuracy is 0.500000, time 299.112 samples/sec
+epoch 13, total_step 210060, total loss is 11.93 , inference loss is 5.52, weight deacy loss is 6.40, training accuracy is 0.468750, time 214.429 samples/sec
+epoch 13, total_step 210080, total loss is 14.91 , inference loss is 8.51, weight deacy loss is 6.40, training accuracy is 0.343750, time 225.477 samples/sec
+epoch 13, total_step 210100, total loss is 14.22 , inference loss is 7.81, weight deacy loss is 6.40, training accuracy is 0.406250, time 221.569 samples/sec
+epoch 13, total_step 210120, total loss is 12.79 , inference loss is 6.38, weight deacy loss is 6.40, training accuracy is 0.562500, time 223.811 samples/sec
+epoch 13, total_step 210140, total loss is 14.63 , inference loss is 8.23, weight deacy loss is 6.40, training accuracy is 0.437500, time 217.333 samples/sec
+epoch 13, total_step 210160, total loss is 12.17 , inference loss is 5.77, weight deacy loss is 6.40, training accuracy is 0.531250, time 218.896 samples/sec
+epoch 13, total_step 210180, total loss is 13.10 , inference loss is 6.69, weight deacy loss is 6.40, training accuracy is 0.531250, time 218.554 samples/sec
+epoch 13, total_step 210200, total loss is 12.95 , inference loss is 6.54, weight deacy loss is 6.40, training accuracy is 0.468750, time 225.123 samples/sec
+epoch 13, total_step 210220, total loss is 11.50 , inference loss is 5.10, weight deacy loss is 6.40, training accuracy is 0.531250, time 225.231 samples/sec
+epoch 13, total_step 210240, total loss is 10.74 , inference loss is 4.34, weight deacy loss is 6.40, training accuracy is 0.593750, time 226.129 samples/sec
+epoch 13, total_step 210260, total loss is 13.89 , inference loss is 7.49, weight deacy loss is 6.40, training accuracy is 0.656250, time 221.389 samples/sec
+epoch 13, total_step 210280, total loss is 14.62 , inference loss is 8.22, weight deacy loss is 6.40, training accuracy is 0.531250, time 224.719 samples/sec
+epoch 13, total_step 210300, total loss is 18.54 , inference loss is 12.14, weight deacy loss is 6.40, training accuracy is 0.437500, time 12.470 samples/sec
+epoch 13, total_step 210320, total loss is 13.01 , inference loss is 6.61, weight deacy loss is 6.40, training accuracy is 0.562500, time 221.670 samples/sec
+epoch 13, total_step 210340, total loss is 16.75 , inference loss is 10.34, weight deacy loss is 6.40, training accuracy is 0.437500, time 226.849 samples/sec
+epoch 13, total_step 210360, total loss is 17.62 , inference loss is 11.22, weight deacy loss is 6.40, training accuracy is 0.343750, time 226.253 samples/sec
+epoch 13, total_step 210380, total loss is 14.90 , inference loss is 8.50, weight deacy loss is 6.40, training accuracy is 0.500000, time 226.124 samples/sec
+epoch 13, total_step 210400, total loss is 16.08 , inference loss is 9.68, weight deacy loss is 6.40, training accuracy is 0.375000, time 219.824 samples/sec
+epoch 13, total_step 210420, total loss is 10.92 , inference loss is 4.52, weight deacy loss is 6.40, training accuracy is 0.687500, time 236.024 samples/sec
+epoch 13, total_step 210440, total loss is 12.55 , inference loss is 6.15, weight deacy loss is 6.40, training accuracy is 0.437500, time 221.413 samples/sec
+epoch 13, total_step 210460, total loss is 13.15 , inference loss is 6.75, weight deacy loss is 6.40, training accuracy is 0.500000, time 228.568 samples/sec
+epoch 13, total_step 210480, total loss is 12.37 , inference loss is 5.97, weight deacy loss is 6.40, training accuracy is 0.437500, time 225.327 samples/sec
+epoch 13, total_step 210500, total loss is 13.28 , inference loss is 6.88, weight deacy loss is 6.40, training accuracy is 0.625000, time 223.324 samples/sec
+epoch 13, total_step 210520, total loss is 17.43 , inference loss is 11.03, weight deacy loss is 6.40, training accuracy is 0.437500, time 211.980 samples/sec
+epoch 13, total_step 210540, total loss is 13.53 , inference loss is 7.13, weight deacy loss is 6.40, training accuracy is 0.406250, time 216.626 samples/sec
+epoch 13, total_step 210560, total loss is 12.81 , inference loss is 6.41, weight deacy loss is 6.40, training accuracy is 0.500000, time 219.139 samples/sec
+epoch 13, total_step 210580, total loss is 11.32 , inference loss is 4.92, weight deacy loss is 6.40, training accuracy is 0.593750, time 223.353 samples/sec
+epoch 13, total_step 210600, total loss is 17.94 , inference loss is 11.54, weight deacy loss is 6.40, training accuracy is 0.250000, time 12.519 samples/sec
+epoch 13, total_step 210620, total loss is 15.58 , inference loss is 9.18, weight deacy loss is 6.40, training accuracy is 0.437500, time 297.592 samples/sec
+epoch 13, total_step 210640, total loss is 15.73 , inference loss is 9.33, weight deacy loss is 6.40, training accuracy is 0.406250, time 199.987 samples/sec
+epoch 13, total_step 210660, total loss is 15.43 , inference loss is 9.03, weight deacy loss is 6.40, training accuracy is 0.312500, time 225.362 samples/sec
+epoch 13, total_step 210680, total loss is 13.32 , inference loss is 6.91, weight deacy loss is 6.40, training accuracy is 0.375000, time 226.336 samples/sec
+epoch 13, total_step 210700, total loss is 16.29 , inference loss is 9.89, weight deacy loss is 6.40, training accuracy is 0.343750, time 196.805 samples/sec
+epoch 13, total_step 210720, total loss is 14.13 , inference loss is 7.72, weight deacy loss is 6.40, training accuracy is 0.468750, time 219.985 samples/sec
+epoch 13, total_step 210740, total loss is 12.94 , inference loss is 6.54, weight deacy loss is 6.40, training accuracy is 0.437500, time 297.691 samples/sec
+epoch 13, total_step 210760, total loss is 13.95 , inference loss is 7.55, weight deacy loss is 6.40, training accuracy is 0.468750, time 233.987 samples/sec
+epoch 13, total_step 210780, total loss is 16.12 , inference loss is 9.72, weight deacy loss is 6.40, training accuracy is 0.500000, time 221.023 samples/sec
+epoch 13, total_step 210800, total loss is 15.71 , inference loss is 9.30, weight deacy loss is 6.40, training accuracy is 0.281250, time 218.375 samples/sec
+epoch 13, total_step 210820, total loss is 15.78 , inference loss is 9.38, weight deacy loss is 6.40, training accuracy is 0.375000, time 221.180 samples/sec
+epoch 13, total_step 210840, total loss is 15.41 , inference loss is 9.01, weight deacy loss is 6.40, training accuracy is 0.531250, time 221.323 samples/sec
+epoch 13, total_step 210860, total loss is 13.09 , inference loss is 6.69, weight deacy loss is 6.40, training accuracy is 0.500000, time 225.688 samples/sec
+epoch 13, total_step 210880, total loss is 14.69 , inference loss is 8.29, weight deacy loss is 6.40, training accuracy is 0.531250, time 221.542 samples/sec
+epoch 13, total_step 210900, total loss is 13.42 , inference loss is 7.02, weight deacy loss is 6.40, training accuracy is 0.343750, time 12.146 samples/sec
+epoch 13, total_step 210920, total loss is 15.12 , inference loss is 8.72, weight deacy loss is 6.40, training accuracy is 0.437500, time 227.887 samples/sec
+epoch 13, total_step 210940, total loss is 14.25 , inference loss is 7.85, weight deacy loss is 6.40, training accuracy is 0.437500, time 226.796 samples/sec
+epoch 13, total_step 210960, total loss is 13.28 , inference loss is 6.87, weight deacy loss is 6.40, training accuracy is 0.343750, time 223.042 samples/sec
+epoch 13, total_step 210980, total loss is 15.25 , inference loss is 8.85, weight deacy loss is 6.40, training accuracy is 0.468750, time 198.019 samples/sec
+epoch 13, total_step 211000, total loss is 10.22 , inference loss is 3.82, weight deacy loss is 6.40, training accuracy is 0.531250, time 295.964 samples/sec
+epoch 13, total_step 211020, total loss is 13.17 , inference loss is 6.77, weight deacy loss is 6.40, training accuracy is 0.437500, time 230.019 samples/sec
+epoch 13, total_step 211040, total loss is 14.70 , inference loss is 8.30, weight deacy loss is 6.40, training accuracy is 0.375000, time 215.716 samples/sec
+epoch 13, total_step 211060, total loss is 9.38 , inference loss is 2.98, weight deacy loss is 6.40, training accuracy is 0.718750, time 222.643 samples/sec
+epoch 13, total_step 211080, total loss is 12.04 , inference loss is 5.63, weight deacy loss is 6.40, training accuracy is 0.437500, time 224.284 samples/sec
+epoch 13, total_step 211100, total loss is 15.02 , inference loss is 8.62, weight deacy loss is 6.40, training accuracy is 0.406250, time 226.193 samples/sec
+epoch 13, total_step 211120, total loss is 15.48 , inference loss is 9.08, weight deacy loss is 6.40, training accuracy is 0.343750, time 220.276 samples/sec
+epoch 13, total_step 211140, total loss is 12.72 , inference loss is 6.32, weight deacy loss is 6.40, training accuracy is 0.531250, time 202.641 samples/sec
+epoch 13, total_step 211160, total loss is 13.62 , inference loss is 7.22, weight deacy loss is 6.40, training accuracy is 0.500000, time 235.096 samples/sec
+epoch 13, total_step 211180, total loss is 12.45 , inference loss is 6.05, weight deacy loss is 6.40, training accuracy is 0.437500, time 227.076 samples/sec
+epoch 13, total_step 211200, total loss is 13.16 , inference loss is 6.76, weight deacy loss is 6.40, training accuracy is 0.437500, time 13.322 samples/sec
+epoch 13, total_step 211220, total loss is 16.49 , inference loss is 10.09, weight deacy loss is 6.40, training accuracy is 0.500000, time 202.935 samples/sec
+epoch 13, total_step 211240, total loss is 14.10 , inference loss is 7.70, weight deacy loss is 6.40, training accuracy is 0.500000, time 239.119 samples/sec
+epoch 13, total_step 211260, total loss is 13.33 , inference loss is 6.93, weight deacy loss is 6.40, training accuracy is 0.468750, time 297.291 samples/sec
+epoch 13, total_step 211280, total loss is 12.78 , inference loss is 6.38, weight deacy loss is 6.40, training accuracy is 0.375000, time 175.490 samples/sec
+epoch 13, total_step 211300, total loss is 12.34 , inference loss is 5.94, weight deacy loss is 6.40, training accuracy is 0.500000, time 227.381 samples/sec
+epoch 13, total_step 211320, total loss is 13.80 , inference loss is 7.40, weight deacy loss is 6.40, training accuracy is 0.625000, time 218.545 samples/sec
+epoch 13, total_step 211340, total loss is 13.04 , inference loss is 6.64, weight deacy loss is 6.40, training accuracy is 0.375000, time 227.115 samples/sec
+epoch 13, total_step 211360, total loss is 10.10 , inference loss is 3.70, weight deacy loss is 6.40, training accuracy is 0.500000, time 239.531 samples/sec
+epoch 13, total_step 211380, total loss is 13.38 , inference loss is 6.98, weight deacy loss is 6.40, training accuracy is 0.437500, time 219.244 samples/sec
+epoch 13, total_step 211400, total loss is 16.32 , inference loss is 9.92, weight deacy loss is 6.40, training accuracy is 0.281250, time 222.799 samples/sec
+epoch 13, total_step 211420, total loss is 20.07 , inference loss is 13.67, weight deacy loss is 6.40, training accuracy is 0.218750, time 227.134 samples/sec
+epoch 13, total_step 211440, total loss is 12.45 , inference loss is 6.05, weight deacy loss is 6.40, training accuracy is 0.531250, time 217.924 samples/sec
+epoch 13, total_step 211460, total loss is 12.67 , inference loss is 6.27, weight deacy loss is 6.40, training accuracy is 0.406250, time 219.545 samples/sec
+epoch 13, total_step 211480, total loss is 18.63 , inference loss is 12.23, weight deacy loss is 6.40, training accuracy is 0.406250, time 213.147 samples/sec
+epoch 13, total_step 211500, total loss is 15.57 , inference loss is 9.17, weight deacy loss is 6.40, training accuracy is 0.375000, time 12.807 samples/sec
+epoch 13, total_step 211520, total loss is 16.90 , inference loss is 10.50, weight deacy loss is 6.40, training accuracy is 0.375000, time 298.282 samples/sec
+epoch 13, total_step 211540, total loss is 14.96 , inference loss is 8.56, weight deacy loss is 6.40, training accuracy is 0.375000, time 221.051 samples/sec
+epoch 13, total_step 211560, total loss is 15.29 , inference loss is 8.89, weight deacy loss is 6.40, training accuracy is 0.437500, time 227.037 samples/sec
+epoch 13, total_step 211580, total loss is 13.43 , inference loss is 7.03, weight deacy loss is 6.40, training accuracy is 0.468750, time 220.256 samples/sec
+epoch 13, total_step 211600, total loss is 14.50 , inference loss is 8.10, weight deacy loss is 6.40, training accuracy is 0.468750, time 226.325 samples/sec
+epoch 13, total_step 211620, total loss is 13.68 , inference loss is 7.29, weight deacy loss is 6.40, training accuracy is 0.562500, time 224.594 samples/sec
+epoch 13, total_step 211640, total loss is 13.93 , inference loss is 7.53, weight deacy loss is 6.40, training accuracy is 0.468750, time 226.158 samples/sec
+epoch 13, total_step 211660, total loss is 11.00 , inference loss is 4.60, weight deacy loss is 6.40, training accuracy is 0.437500, time 220.955 samples/sec
+epoch 13, total_step 211680, total loss is 11.92 , inference loss is 5.52, weight deacy loss is 6.40, training accuracy is 0.406250, time 218.583 samples/sec
+epoch 13, total_step 211700, total loss is 12.78 , inference loss is 6.39, weight deacy loss is 6.40, training accuracy is 0.468750, time 225.394 samples/sec
+epoch 13, total_step 211720, total loss is 14.67 , inference loss is 8.27, weight deacy loss is 6.40, training accuracy is 0.437500, time 221.878 samples/sec
+epoch 13, total_step 211740, total loss is 17.64 , inference loss is 11.24, weight deacy loss is 6.40, training accuracy is 0.312500, time 220.364 samples/sec
+epoch 13, total_step 211760, total loss is 12.10 , inference loss is 5.70, weight deacy loss is 6.40, training accuracy is 0.406250, time 222.212 samples/sec
+epoch 13, total_step 211780, total loss is 13.75 , inference loss is 7.35, weight deacy loss is 6.40, training accuracy is 0.437500, time 231.439 samples/sec
+epoch 13, total_step 211800, total loss is 13.68 , inference loss is 7.28, weight deacy loss is 6.40, training accuracy is 0.406250, time 12.362 samples/sec
+epoch 13, total_step 211820, total loss is 12.69 , inference loss is 6.29, weight deacy loss is 6.40, training accuracy is 0.437500, time 226.081 samples/sec
+epoch 13, total_step 211840, total loss is 12.13 , inference loss is 5.73, weight deacy loss is 6.40, training accuracy is 0.562500, time 224.548 samples/sec
+epoch 13, total_step 211860, total loss is 12.00 , inference loss is 5.61, weight deacy loss is 6.40, training accuracy is 0.531250, time 218.303 samples/sec
+epoch 13, total_step 211880, total loss is 11.11 , inference loss is 4.71, weight deacy loss is 6.40, training accuracy is 0.531250, time 228.027 samples/sec
+epoch 13, total_step 211900, total loss is 14.45 , inference loss is 8.05, weight deacy loss is 6.40, training accuracy is 0.468750, time 227.157 samples/sec
+epoch 13, total_step 211920, total loss is 13.68 , inference loss is 7.29, weight deacy loss is 6.40, training accuracy is 0.375000, time 226.818 samples/sec
+epoch 13, total_step 211940, total loss is 14.22 , inference loss is 7.82, weight deacy loss is 6.40, training accuracy is 0.468750, time 224.804 samples/sec
+epoch 13, total_step 211960, total loss is 16.25 , inference loss is 9.85, weight deacy loss is 6.40, training accuracy is 0.312500, time 221.251 samples/sec
+epoch 13, total_step 211980, total loss is 11.68 , inference loss is 5.29, weight deacy loss is 6.40, training accuracy is 0.562500, time 213.888 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.775781000000006
+best_threshold_index 109 0.985
+best_threshold_index 109 0.9840740740740741
+best_threshold_index 109 0.9838888888888889
+best_threshold_index 109 0.9855555555555555
+best_threshold_index 109 0.9848148148148148
+best_threshold_index 109 0.9838888888888889
+best_threshold_index 109 0.9842592592592593
+best_threshold_index 109 0.9840740740740741
+best_threshold_index 109 0.9827777777777778
+best_threshold_index 109 0.9833333333333333
+[lfw][212000]XNorm: 22.116135
+[lfw][212000]Accuracy-Flip: 0.98417+-0.00688
+test accuracy is: 0.9841666666666666
+epoch 13, total_step 212000, total loss is 13.20 , inference loss is 6.80, weight deacy loss is 6.40, training accuracy is 0.437500, time 9.666 samples/sec
+epoch 13, total_step 212020, total loss is 20.49 , inference loss is 14.09, weight deacy loss is 6.40, training accuracy is 0.437500, time 228.791 samples/sec
+epoch 13, total_step 212040, total loss is 11.98 , inference loss is 5.59, weight deacy loss is 6.40, training accuracy is 0.406250, time 221.776 samples/sec
+epoch 13, total_step 212060, total loss is 15.02 , inference loss is 8.62, weight deacy loss is 6.40, training accuracy is 0.406250, time 218.686 samples/sec
+epoch 13, total_step 212080, total loss is 11.27 , inference loss is 4.88, weight deacy loss is 6.40, training accuracy is 0.437500, time 193.801 samples/sec
+epoch 13, total_step 212100, total loss is 13.15 , inference loss is 6.75, weight deacy loss is 6.40, training accuracy is 0.437500, time 12.485 samples/sec
+epoch 13, total_step 212120, total loss is 11.47 , inference loss is 5.07, weight deacy loss is 6.40, training accuracy is 0.468750, time 219.753 samples/sec
+epoch 13, total_step 212140, total loss is 10.95 , inference loss is 4.55, weight deacy loss is 6.40, training accuracy is 0.562500, time 222.253 samples/sec
+epoch 13, total_step 212160, total loss is 11.86 , inference loss is 5.47, weight deacy loss is 6.40, training accuracy is 0.562500, time 295.891 samples/sec
+epoch 13, total_step 212180, total loss is 18.00 , inference loss is 11.61, weight deacy loss is 6.40, training accuracy is 0.437500, time 218.323 samples/sec
+epoch 13, total_step 212200, total loss is 14.13 , inference loss is 7.74, weight deacy loss is 6.40, training accuracy is 0.312500, time 213.846 samples/sec
+epoch 13, total_step 212220, total loss is 13.89 , inference loss is 7.50, weight deacy loss is 6.40, training accuracy is 0.375000, time 229.278 samples/sec
+epoch 13, total_step 212240, total loss is 14.10 , inference loss is 7.70, weight deacy loss is 6.40, training accuracy is 0.593750, time 209.858 samples/sec
+epoch 13, total_step 212260, total loss is 14.12 , inference loss is 7.72, weight deacy loss is 6.40, training accuracy is 0.343750, time 222.787 samples/sec
+epoch 13, total_step 212280, total loss is 17.78 , inference loss is 11.39, weight deacy loss is 6.40, training accuracy is 0.406250, time 213.699 samples/sec
+epoch 13, total_step 212300, total loss is 13.91 , inference loss is 7.52, weight deacy loss is 6.40, training accuracy is 0.312500, time 217.943 samples/sec
+epoch 13, total_step 212320, total loss is 15.34 , inference loss is 8.95, weight deacy loss is 6.39, training accuracy is 0.343750, time 297.422 samples/sec
+epoch 13, total_step 212340, total loss is 13.71 , inference loss is 7.32, weight deacy loss is 6.39, training accuracy is 0.406250, time 298.453 samples/sec
+epoch 13, total_step 212360, total loss is 15.90 , inference loss is 9.50, weight deacy loss is 6.39, training accuracy is 0.312500, time 222.828 samples/sec
+epoch 13, total_step 212380, total loss is 15.70 , inference loss is 9.31, weight deacy loss is 6.39, training accuracy is 0.312500, time 214.383 samples/sec
+epoch 13, total_step 212400, total loss is 13.81 , inference loss is 7.41, weight deacy loss is 6.39, training accuracy is 0.500000, time 11.811 samples/sec
+epoch 13, total_step 212420, total loss is 13.35 , inference loss is 6.95, weight deacy loss is 6.39, training accuracy is 0.406250, time 211.581 samples/sec
+epoch 13, total_step 212440, total loss is 15.63 , inference loss is 9.24, weight deacy loss is 6.39, training accuracy is 0.562500, time 232.103 samples/sec
+epoch 13, total_step 212460, total loss is 13.22 , inference loss is 6.83, weight deacy loss is 6.39, training accuracy is 0.406250, time 243.997 samples/sec
+epoch 13, total_step 212480, total loss is 9.90 , inference loss is 3.50, weight deacy loss is 6.39, training accuracy is 0.531250, time 219.647 samples/sec
+epoch 13, total_step 212500, total loss is 17.20 , inference loss is 10.81, weight deacy loss is 6.39, training accuracy is 0.281250, time 295.885 samples/sec
+epoch 13, total_step 212520, total loss is 12.62 , inference loss is 6.23, weight deacy loss is 6.39, training accuracy is 0.406250, time 223.374 samples/sec
+epoch 13, total_step 212540, total loss is 16.30 , inference loss is 9.91, weight deacy loss is 6.39, training accuracy is 0.343750, time 225.713 samples/sec
+epoch 13, total_step 212560, total loss is 12.63 , inference loss is 6.24, weight deacy loss is 6.39, training accuracy is 0.406250, time 228.665 samples/sec
+epoch 13, total_step 212580, total loss is 12.42 , inference loss is 6.03, weight deacy loss is 6.39, training accuracy is 0.531250, time 225.084 samples/sec
+epoch 13, total_step 212600, total loss is 13.66 , inference loss is 7.26, weight deacy loss is 6.39, training accuracy is 0.375000, time 229.553 samples/sec
+epoch 13, total_step 212620, total loss is 13.99 , inference loss is 7.60, weight deacy loss is 6.39, training accuracy is 0.375000, time 227.552 samples/sec
+epoch 13, total_step 212640, total loss is 10.68 , inference loss is 4.29, weight deacy loss is 6.39, training accuracy is 0.468750, time 221.326 samples/sec
+epoch 13, total_step 212660, total loss is 13.29 , inference loss is 6.90, weight deacy loss is 6.39, training accuracy is 0.406250, time 224.593 samples/sec
+epoch 13, total_step 212680, total loss is 11.30 , inference loss is 4.90, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.110 samples/sec
+epoch 13, total_step 212700, total loss is 18.06 , inference loss is 11.67, weight deacy loss is 6.39, training accuracy is 0.406250, time 12.652 samples/sec
+epoch 13, total_step 212720, total loss is 12.43 , inference loss is 6.04, weight deacy loss is 6.39, training accuracy is 0.562500, time 226.261 samples/sec
+epoch 13, total_step 212740, total loss is 17.74 , inference loss is 11.34, weight deacy loss is 6.39, training accuracy is 0.437500, time 231.110 samples/sec
+epoch 13, total_step 212760, total loss is 11.88 , inference loss is 5.49, weight deacy loss is 6.39, training accuracy is 0.531250, time 298.311 samples/sec
+epoch 13, total_step 212780, total loss is 12.34 , inference loss is 5.94, weight deacy loss is 6.39, training accuracy is 0.468750, time 219.757 samples/sec
+epoch 13, total_step 212800, total loss is 12.64 , inference loss is 6.25, weight deacy loss is 6.39, training accuracy is 0.468750, time 224.427 samples/sec
+epoch 13, total_step 212820, total loss is 11.44 , inference loss is 5.05, weight deacy loss is 6.39, training accuracy is 0.468750, time 223.665 samples/sec
+epoch 13, total_step 212840, total loss is 16.27 , inference loss is 9.87, weight deacy loss is 6.39, training accuracy is 0.375000, time 225.407 samples/sec
+epoch 13, total_step 212860, total loss is 16.81 , inference loss is 10.42, weight deacy loss is 6.39, training accuracy is 0.406250, time 210.386 samples/sec
+epoch 13, total_step 212880, total loss is 11.99 , inference loss is 5.60, weight deacy loss is 6.39, training accuracy is 0.468750, time 218.878 samples/sec
+epoch 13, total_step 212900, total loss is 11.38 , inference loss is 4.98, weight deacy loss is 6.39, training accuracy is 0.468750, time 215.609 samples/sec
+epoch 13, total_step 212920, total loss is 18.89 , inference loss is 12.50, weight deacy loss is 6.39, training accuracy is 0.281250, time 218.722 samples/sec
+epoch 13, total_step 212940, total loss is 10.71 , inference loss is 4.32, weight deacy loss is 6.39, training accuracy is 0.437500, time 249.260 samples/sec
+epoch 13, total_step 212960, total loss is 16.22 , inference loss is 9.83, weight deacy loss is 6.39, training accuracy is 0.281250, time 223.637 samples/sec
+epoch 13, total_step 212980, total loss is 12.41 , inference loss is 6.02, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.639 samples/sec
+epoch 13, total_step 213000, total loss is 16.94 , inference loss is 10.55, weight deacy loss is 6.39, training accuracy is 0.312500, time 12.587 samples/sec
+epoch 13, total_step 213020, total loss is 13.79 , inference loss is 7.40, weight deacy loss is 6.39, training accuracy is 0.281250, time 197.349 samples/sec
+epoch 13, total_step 213040, total loss is 17.71 , inference loss is 11.32, weight deacy loss is 6.39, training accuracy is 0.250000, time 239.222 samples/sec
+epoch 13, total_step 213060, total loss is 15.08 , inference loss is 8.69, weight deacy loss is 6.39, training accuracy is 0.343750, time 225.387 samples/sec
+epoch 13, total_step 213080, total loss is 16.35 , inference loss is 9.96, weight deacy loss is 6.39, training accuracy is 0.312500, time 215.679 samples/sec
+epoch 13, total_step 213100, total loss is 14.39 , inference loss is 8.00, weight deacy loss is 6.39, training accuracy is 0.281250, time 246.818 samples/sec
+epoch 13, total_step 213120, total loss is 12.03 , inference loss is 5.64, weight deacy loss is 6.39, training accuracy is 0.468750, time 219.554 samples/sec
+epoch 13, total_step 213140, total loss is 14.68 , inference loss is 8.29, weight deacy loss is 6.39, training accuracy is 0.406250, time 211.379 samples/sec
+epoch 13, total_step 213160, total loss is 15.34 , inference loss is 8.95, weight deacy loss is 6.39, training accuracy is 0.500000, time 224.708 samples/sec
+epoch 13, total_step 213180, total loss is 18.98 , inference loss is 12.59, weight deacy loss is 6.39, training accuracy is 0.250000, time 213.506 samples/sec
+epoch 13, total_step 213200, total loss is 16.00 , inference loss is 9.61, weight deacy loss is 6.39, training accuracy is 0.437500, time 220.850 samples/sec
+epoch 13, total_step 213220, total loss is 17.49 , inference loss is 11.10, weight deacy loss is 6.39, training accuracy is 0.281250, time 222.743 samples/sec
+epoch 13, total_step 213240, total loss is 15.40 , inference loss is 9.01, weight deacy loss is 6.39, training accuracy is 0.406250, time 218.476 samples/sec
+epoch 13, total_step 213260, total loss is 16.87 , inference loss is 10.48, weight deacy loss is 6.39, training accuracy is 0.281250, time 227.041 samples/sec
+epoch 13, total_step 213280, total loss is 15.67 , inference loss is 9.28, weight deacy loss is 6.39, training accuracy is 0.218750, time 228.715 samples/sec
+epoch 13, total_step 213300, total loss is 18.76 , inference loss is 12.37, weight deacy loss is 6.39, training accuracy is 0.281250, time 12.051 samples/sec
+epoch 13, total_step 213320, total loss is 13.79 , inference loss is 7.40, weight deacy loss is 6.39, training accuracy is 0.375000, time 224.435 samples/sec
+epoch 13, total_step 213340, total loss is 14.20 , inference loss is 7.82, weight deacy loss is 6.39, training accuracy is 0.218750, time 214.533 samples/sec
+epoch 13, total_step 213360, total loss is 11.87 , inference loss is 5.48, weight deacy loss is 6.39, training accuracy is 0.406250, time 228.327 samples/sec
+epoch 13, total_step 213380, total loss is 13.16 , inference loss is 6.77, weight deacy loss is 6.39, training accuracy is 0.437500, time 229.945 samples/sec
+epoch 13, total_step 213400, total loss is 13.41 , inference loss is 7.02, weight deacy loss is 6.39, training accuracy is 0.468750, time 229.848 samples/sec
+epoch 13, total_step 213420, total loss is 13.57 , inference loss is 7.18, weight deacy loss is 6.39, training accuracy is 0.375000, time 220.746 samples/sec
+epoch 13, total_step 213440, total loss is 14.52 , inference loss is 8.13, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.872 samples/sec
+epoch 13, total_step 213460, total loss is 12.48 , inference loss is 6.09, weight deacy loss is 6.39, training accuracy is 0.375000, time 224.756 samples/sec
+epoch 13, total_step 213480, total loss is 14.68 , inference loss is 8.29, weight deacy loss is 6.39, training accuracy is 0.312500, time 224.862 samples/sec
+epoch 13, total_step 213500, total loss is 13.48 , inference loss is 7.09, weight deacy loss is 6.39, training accuracy is 0.343750, time 224.385 samples/sec
+epoch 13, total_step 213520, total loss is 15.18 , inference loss is 8.79, weight deacy loss is 6.39, training accuracy is 0.281250, time 224.948 samples/sec
+epoch 13, total_step 213540, total loss is 13.98 , inference loss is 7.59, weight deacy loss is 6.39, training accuracy is 0.406250, time 221.003 samples/sec
+epoch 13, total_step 213560, total loss is 17.34 , inference loss is 10.95, weight deacy loss is 6.39, training accuracy is 0.468750, time 212.941 samples/sec
+epoch 13, total_step 213580, total loss is 13.85 , inference loss is 7.46, weight deacy loss is 6.39, training accuracy is 0.437500, time 296.723 samples/sec
+epoch 13, total_step 213600, total loss is 15.81 , inference loss is 9.42, weight deacy loss is 6.39, training accuracy is 0.312500, time 12.344 samples/sec
+epoch 13, total_step 213620, total loss is 14.72 , inference loss is 8.33, weight deacy loss is 6.39, training accuracy is 0.312500, time 224.002 samples/sec
+epoch 13, total_step 213640, total loss is 12.26 , inference loss is 5.88, weight deacy loss is 6.39, training accuracy is 0.437500, time 226.172 samples/sec
+epoch 13, total_step 213660, total loss is 14.20 , inference loss is 7.81, weight deacy loss is 6.39, training accuracy is 0.281250, time 176.317 samples/sec
+epoch 13, total_step 213680, total loss is 15.90 , inference loss is 9.51, weight deacy loss is 6.39, training accuracy is 0.406250, time 229.418 samples/sec
+epoch 13, total_step 213700, total loss is 15.35 , inference loss is 8.96, weight deacy loss is 6.39, training accuracy is 0.375000, time 227.125 samples/sec
+epoch 13, total_step 213720, total loss is 11.87 , inference loss is 5.48, weight deacy loss is 6.39, training accuracy is 0.406250, time 219.989 samples/sec
+epoch 13, total_step 213740, total loss is 11.07 , inference loss is 4.68, weight deacy loss is 6.39, training accuracy is 0.531250, time 227.871 samples/sec
+epoch 13, total_step 213760, total loss is 12.36 , inference loss is 5.97, weight deacy loss is 6.39, training accuracy is 0.406250, time 221.299 samples/sec
+epoch 13, total_step 213780, total loss is 13.14 , inference loss is 6.75, weight deacy loss is 6.39, training accuracy is 0.375000, time 226.174 samples/sec
+epoch 13, total_step 213800, total loss is 14.01 , inference loss is 7.62, weight deacy loss is 6.39, training accuracy is 0.406250, time 219.870 samples/sec
+epoch 13, total_step 213820, total loss is 16.76 , inference loss is 10.37, weight deacy loss is 6.39, training accuracy is 0.343750, time 227.168 samples/sec
+epoch 13, total_step 213840, total loss is 12.51 , inference loss is 6.12, weight deacy loss is 6.39, training accuracy is 0.468750, time 297.189 samples/sec
+epoch 13, total_step 213860, total loss is 15.70 , inference loss is 9.32, weight deacy loss is 6.39, training accuracy is 0.312500, time 226.874 samples/sec
+epoch 13, total_step 213880, total loss is 16.75 , inference loss is 10.37, weight deacy loss is 6.39, training accuracy is 0.281250, time 224.513 samples/sec
+epoch 13, total_step 213900, total loss is 13.95 , inference loss is 7.57, weight deacy loss is 6.39, training accuracy is 0.406250, time 12.156 samples/sec
+epoch 13, total_step 213920, total loss is 13.61 , inference loss is 7.22, weight deacy loss is 6.39, training accuracy is 0.437500, time 227.156 samples/sec
+epoch 13, total_step 213940, total loss is 17.87 , inference loss is 11.49, weight deacy loss is 6.39, training accuracy is 0.343750, time 216.035 samples/sec
+epoch 13, total_step 213960, total loss is 14.89 , inference loss is 8.51, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.672 samples/sec
+epoch 13, total_step 213980, total loss is 10.55 , inference loss is 4.17, weight deacy loss is 6.39, training accuracy is 0.500000, time 231.193 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.572648000000008
+best_threshold_index 105 0.9842592592592593
+best_threshold_index 105 0.9835185185185186
+best_threshold_index 103 0.9837037037037037
+best_threshold_index 105 0.9851851851851852
+best_threshold_index 105 0.9837037037037037
+best_threshold_index 106 0.9825925925925926
+best_threshold_index 105 0.9833333333333333
+best_threshold_index 105 0.9829629629629629
+best_threshold_index 105 0.9816666666666667
+best_threshold_index 105 0.9829629629629629
+[lfw][214000]XNorm: 22.200127
+[lfw][214000]Accuracy-Flip: 0.98233+-0.00817
+test accuracy is: 0.9823333333333333
+epoch 13, total_step 214000, total loss is 15.45 , inference loss is 9.06, weight deacy loss is 6.39, training accuracy is 0.437500, time 16.159 samples/sec
+epoch 13, total_step 214020, total loss is 15.62 , inference loss is 9.24, weight deacy loss is 6.39, training accuracy is 0.250000, time 221.592 samples/sec
+epoch 13, total_step 214040, total loss is 14.46 , inference loss is 8.07, weight deacy loss is 6.39, training accuracy is 0.406250, time 300.617 samples/sec
+epoch 13, total_step 214060, total loss is 13.25 , inference loss is 6.87, weight deacy loss is 6.39, training accuracy is 0.406250, time 298.675 samples/sec
+epoch 13, total_step 214080, total loss is 14.19 , inference loss is 7.80, weight deacy loss is 6.39, training accuracy is 0.375000, time 215.995 samples/sec2022-12-01 14:33:40.717779: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 6916 of 10000
+2022-12-01 14:33:45.113660: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 13, total_step 214100, total loss is 12.77 , inference loss is 6.38, weight deacy loss is 6.39, training accuracy is 0.343750, time 217.955 samples/sec
+epoch 13, total_step 214120, total loss is 12.62 , inference loss is 6.24, weight deacy loss is 6.38, training accuracy is 0.406250, time 215.664 samples/sec
+epoch 13, total_step 214140, total loss is 15.07 , inference loss is 8.69, weight deacy loss is 6.38, training accuracy is 0.312500, time 233.458 samples/sec
+epoch 13, total_step 214160, total loss is 14.63 , inference loss is 8.25, weight deacy loss is 6.38, training accuracy is 0.375000, time 227.392 samples/sec
+epoch 13, total_step 214180, total loss is 18.06 , inference loss is 11.67, weight deacy loss is 6.38, training accuracy is 0.250000, time 213.405 samples/sec
+epoch 13, total_step 214200, total loss is 13.26 , inference loss is 6.88, weight deacy loss is 6.38, training accuracy is 0.375000, time 13.115 samples/sec
+epoch 13, total_step 214220, total loss is 13.87 , inference loss is 7.49, weight deacy loss is 6.38, training accuracy is 0.312500, time 235.784 samples/sec
+epoch 13, total_step 214240, total loss is 15.77 , inference loss is 9.38, weight deacy loss is 6.38, training accuracy is 0.281250, time 214.820 samples/sec
+epoch 13, total_step 214260, total loss is 15.27 , inference loss is 8.89, weight deacy loss is 6.38, training accuracy is 0.406250, time 219.136 samples/sec
+epoch 13, total_step 214280, total loss is 14.59 , inference loss is 8.21, weight deacy loss is 6.38, training accuracy is 0.250000, time 226.059 samples/sec
+epoch 13, total_step 214300, total loss is 18.64 , inference loss is 12.26, weight deacy loss is 6.38, training accuracy is 0.250000, time 231.461 samples/sec
+epoch 13, total_step 214320, total loss is 14.66 , inference loss is 8.28, weight deacy loss is 6.38, training accuracy is 0.281250, time 296.379 samples/sec
+epoch 13, total_step 214340, total loss is 12.95 , inference loss is 6.57, weight deacy loss is 6.38, training accuracy is 0.531250, time 202.525 samples/sec
+epoch 13, total_step 214360, total loss is 17.38 , inference loss is 10.99, weight deacy loss is 6.38, training accuracy is 0.281250, time 208.633 samples/sec
+epoch 13, total_step 214380, total loss is 11.02 , inference loss is 4.64, weight deacy loss is 6.38, training accuracy is 0.531250, time 222.972 samples/sec
+epoch 13, total_step 214400, total loss is 12.16 , inference loss is 5.78, weight deacy loss is 6.38, training accuracy is 0.437500, time 194.116 samples/sec
+epoch 13, total_step 214420, total loss is 12.50 , inference loss is 6.12, weight deacy loss is 6.38, training accuracy is 0.500000, time 209.373 samples/sec
+epoch 13, total_step 214440, total loss is 19.14 , inference loss is 12.75, weight deacy loss is 6.38, training accuracy is 0.406250, time 193.340 samples/sec
+epoch 13, total_step 214460, total loss is 11.99 , inference loss is 5.61, weight deacy loss is 6.38, training accuracy is 0.500000, time 202.342 samples/sec
+epoch 13, total_step 214480, total loss is 16.35 , inference loss is 9.97, weight deacy loss is 6.38, training accuracy is 0.312500, time 213.495 samples/sec
+epoch 13, total_step 214500, total loss is 14.22 , inference loss is 7.83, weight deacy loss is 6.38, training accuracy is 0.375000, time 12.872 samples/sec
+epoch 13, total_step 214520, total loss is 14.79 , inference loss is 8.40, weight deacy loss is 6.38, training accuracy is 0.468750, time 228.093 samples/sec
+epoch 13, total_step 214540, total loss is 14.45 , inference loss is 8.07, weight deacy loss is 6.38, training accuracy is 0.500000, time 195.571 samples/sec
+epoch 13, total_step 214560, total loss is 18.89 , inference loss is 12.51, weight deacy loss is 6.38, training accuracy is 0.250000, time 201.427 samples/sec
+epoch 13, total_step 214580, total loss is 14.23 , inference loss is 7.85, weight deacy loss is 6.38, training accuracy is 0.406250, time 200.238 samples/sec
+epoch 13, total_step 214600, total loss is 18.77 , inference loss is 12.39, weight deacy loss is 6.38, training accuracy is 0.312500, time 192.323 samples/sec
+epoch 13, total_step 214620, total loss is 10.92 , inference loss is 4.54, weight deacy loss is 6.38, training accuracy is 0.531250, time 205.844 samples/sec
+End of epoch 13
+epoch 14, total_step 214640, total loss is 20.68 , inference loss is 14.29, weight deacy loss is 6.38, training accuracy is 0.093750, time 216.634 samples/sec
+epoch 14, total_step 214660, total loss is 10.21 , inference loss is 3.83, weight deacy loss is 6.38, training accuracy is 0.656250, time 219.879 samples/sec
+epoch 14, total_step 214680, total loss is 9.34 , inference loss is 2.95, weight deacy loss is 6.38, training accuracy is 0.718750, time 216.656 samples/sec
+epoch 14, total_step 214700, total loss is 9.99 , inference loss is 3.61, weight deacy loss is 6.38, training accuracy is 0.593750, time 237.944 samples/sec
+epoch 14, total_step 214720, total loss is 12.57 , inference loss is 6.19, weight deacy loss is 6.38, training accuracy is 0.656250, time 222.630 samples/sec
+epoch 14, total_step 214740, total loss is 8.70 , inference loss is 2.32, weight deacy loss is 6.38, training accuracy is 0.718750, time 222.137 samples/sec
+epoch 14, total_step 214760, total loss is 10.90 , inference loss is 4.52, weight deacy loss is 6.38, training accuracy is 0.781250, time 227.458 samples/sec
+epoch 14, total_step 214780, total loss is 8.47 , inference loss is 2.09, weight deacy loss is 6.38, training accuracy is 0.937500, time 218.040 samples/sec
+epoch 14, total_step 214800, total loss is 11.21 , inference loss is 4.83, weight deacy loss is 6.38, training accuracy is 0.718750, time 11.972 samples/sec
+epoch 14, total_step 214820, total loss is 12.89 , inference loss is 6.51, weight deacy loss is 6.38, training accuracy is 0.687500, time 298.102 samples/sec
+epoch 14, total_step 214840, total loss is 9.72 , inference loss is 3.34, weight deacy loss is 6.38, training accuracy is 0.625000, time 217.047 samples/sec
+epoch 14, total_step 214860, total loss is 10.41 , inference loss is 4.03, weight deacy loss is 6.38, training accuracy is 0.656250, time 223.988 samples/sec
+epoch 14, total_step 214880, total loss is 8.17 , inference loss is 1.79, weight deacy loss is 6.38, training accuracy is 0.812500, time 216.892 samples/sec
+epoch 14, total_step 214900, total loss is 9.61 , inference loss is 3.23, weight deacy loss is 6.38, training accuracy is 0.750000, time 193.590 samples/sec
+epoch 14, total_step 214920, total loss is 11.44 , inference loss is 5.06, weight deacy loss is 6.38, training accuracy is 0.750000, time 231.807 samples/sec
+epoch 14, total_step 214940, total loss is 10.30 , inference loss is 3.92, weight deacy loss is 6.38, training accuracy is 0.562500, time 219.621 samples/sec
+epoch 14, total_step 214960, total loss is 12.87 , inference loss is 6.49, weight deacy loss is 6.38, training accuracy is 0.625000, time 220.122 samples/sec
+epoch 14, total_step 214980, total loss is 13.74 , inference loss is 7.36, weight deacy loss is 6.38, training accuracy is 0.593750, time 219.011 samples/sec
+epoch 14, total_step 215000, total loss is 11.40 , inference loss is 5.02, weight deacy loss is 6.38, training accuracy is 0.656250, time 225.629 samples/sec
+epoch 14, total_step 215020, total loss is 13.34 , inference loss is 6.96, weight deacy loss is 6.38, training accuracy is 0.718750, time 219.805 samples/sec
+epoch 14, total_step 215040, total loss is 13.22 , inference loss is 6.84, weight deacy loss is 6.38, training accuracy is 0.687500, time 224.119 samples/sec
+epoch 14, total_step 215060, total loss is 9.25 , inference loss is 2.87, weight deacy loss is 6.38, training accuracy is 0.781250, time 225.630 samples/sec
+epoch 14, total_step 215080, total loss is 8.80 , inference loss is 2.42, weight deacy loss is 6.38, training accuracy is 0.718750, time 223.778 samples/sec
+epoch 14, total_step 215100, total loss is 10.92 , inference loss is 4.54, weight deacy loss is 6.38, training accuracy is 0.718750, time 12.562 samples/sec
+epoch 14, total_step 215120, total loss is 11.64 , inference loss is 5.26, weight deacy loss is 6.38, training accuracy is 0.625000, time 222.386 samples/sec
+epoch 14, total_step 215140, total loss is 12.30 , inference loss is 5.92, weight deacy loss is 6.38, training accuracy is 0.656250, time 245.917 samples/sec
+epoch 14, total_step 215160, total loss is 11.52 , inference loss is 5.14, weight deacy loss is 6.38, training accuracy is 0.562500, time 216.791 samples/sec
+epoch 14, total_step 215180, total loss is 12.22 , inference loss is 5.84, weight deacy loss is 6.38, training accuracy is 0.687500, time 221.869 samples/sec
+epoch 14, total_step 215200, total loss is 13.44 , inference loss is 7.06, weight deacy loss is 6.38, training accuracy is 0.781250, time 225.950 samples/sec
+epoch 14, total_step 215220, total loss is 12.74 , inference loss is 6.35, weight deacy loss is 6.38, training accuracy is 0.593750, time 238.854 samples/sec
+epoch 14, total_step 215240, total loss is 11.25 , inference loss is 4.87, weight deacy loss is 6.38, training accuracy is 0.687500, time 224.688 samples/sec
+epoch 14, total_step 215260, total loss is 11.71 , inference loss is 5.33, weight deacy loss is 6.38, training accuracy is 0.562500, time 225.712 samples/sec
+epoch 14, total_step 215280, total loss is 14.02 , inference loss is 7.64, weight deacy loss is 6.38, training accuracy is 0.656250, time 220.481 samples/sec
+epoch 14, total_step 215300, total loss is 14.02 , inference loss is 7.64, weight deacy loss is 6.38, training accuracy is 0.625000, time 225.367 samples/sec
+epoch 14, total_step 215320, total loss is 8.42 , inference loss is 2.04, weight deacy loss is 6.38, training accuracy is 0.718750, time 226.635 samples/sec
+epoch 14, total_step 215340, total loss is 13.25 , inference loss is 6.87, weight deacy loss is 6.38, training accuracy is 0.687500, time 224.933 samples/sec
+epoch 14, total_step 215360, total loss is 9.78 , inference loss is 3.39, weight deacy loss is 6.38, training accuracy is 0.687500, time 223.987 samples/sec
+epoch 14, total_step 215380, total loss is 12.45 , inference loss is 6.07, weight deacy loss is 6.38, training accuracy is 0.656250, time 210.193 samples/sec
+epoch 14, total_step 215400, total loss is 12.22 , inference loss is 5.84, weight deacy loss is 6.38, training accuracy is 0.625000, time 11.552 samples/sec
+epoch 14, total_step 215420, total loss is 13.94 , inference loss is 7.56, weight deacy loss is 6.38, training accuracy is 0.562500, time 226.349 samples/sec
+epoch 14, total_step 215440, total loss is 12.08 , inference loss is 5.70, weight deacy loss is 6.38, training accuracy is 0.500000, time 179.014 samples/sec
+epoch 14, total_step 215460, total loss is 12.81 , inference loss is 6.43, weight deacy loss is 6.38, training accuracy is 0.562500, time 243.257 samples/sec
+epoch 14, total_step 215480, total loss is 14.34 , inference loss is 7.95, weight deacy loss is 6.38, training accuracy is 0.562500, time 190.210 samples/sec
+epoch 14, total_step 215500, total loss is 16.75 , inference loss is 10.37, weight deacy loss is 6.38, training accuracy is 0.531250, time 221.430 samples/sec
+epoch 14, total_step 215520, total loss is 13.61 , inference loss is 7.23, weight deacy loss is 6.39, training accuracy is 0.656250, time 240.275 samples/sec
+epoch 14, total_step 215540, total loss is 12.40 , inference loss is 6.02, weight deacy loss is 6.39, training accuracy is 0.531250, time 231.748 samples/sec
+epoch 14, total_step 215560, total loss is 8.71 , inference loss is 2.33, weight deacy loss is 6.39, training accuracy is 0.625000, time 219.725 samples/sec
+epoch 14, total_step 215580, total loss is 13.66 , inference loss is 7.27, weight deacy loss is 6.39, training accuracy is 0.656250, time 235.614 samples/sec
+epoch 14, total_step 215600, total loss is 11.47 , inference loss is 5.08, weight deacy loss is 6.39, training accuracy is 0.625000, time 226.562 samples/sec
+epoch 14, total_step 215620, total loss is 12.29 , inference loss is 5.90, weight deacy loss is 6.39, training accuracy is 0.593750, time 228.516 samples/sec
+epoch 14, total_step 215640, total loss is 11.16 , inference loss is 4.78, weight deacy loss is 6.39, training accuracy is 0.593750, time 298.700 samples/sec
+epoch 14, total_step 215660, total loss is 13.03 , inference loss is 6.64, weight deacy loss is 6.39, training accuracy is 0.781250, time 184.752 samples/sec
+epoch 14, total_step 215680, total loss is 12.86 , inference loss is 6.47, weight deacy loss is 6.39, training accuracy is 0.562500, time 227.142 samples/sec
+epoch 14, total_step 215700, total loss is 13.36 , inference loss is 6.97, weight deacy loss is 6.39, training accuracy is 0.562500, time 12.628 samples/sec
+epoch 14, total_step 215720, total loss is 9.16 , inference loss is 2.78, weight deacy loss is 6.39, training accuracy is 0.750000, time 230.880 samples/sec
+epoch 14, total_step 215740, total loss is 14.56 , inference loss is 8.17, weight deacy loss is 6.39, training accuracy is 0.531250, time 199.240 samples/sec
+epoch 14, total_step 215760, total loss is 13.61 , inference loss is 7.22, weight deacy loss is 6.39, training accuracy is 0.656250, time 220.743 samples/sec
+epoch 14, total_step 215780, total loss is 12.81 , inference loss is 6.42, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.697 samples/sec
+epoch 14, total_step 215800, total loss is 11.50 , inference loss is 5.11, weight deacy loss is 6.39, training accuracy is 0.531250, time 202.218 samples/sec
+epoch 14, total_step 215820, total loss is 16.71 , inference loss is 10.32, weight deacy loss is 6.39, training accuracy is 0.562500, time 207.374 samples/sec
+epoch 14, total_step 215840, total loss is 15.15 , inference loss is 8.76, weight deacy loss is 6.39, training accuracy is 0.593750, time 224.930 samples/sec
+epoch 14, total_step 215860, total loss is 15.08 , inference loss is 8.69, weight deacy loss is 6.39, training accuracy is 0.437500, time 217.704 samples/sec
+epoch 14, total_step 215880, total loss is 11.40 , inference loss is 5.01, weight deacy loss is 6.39, training accuracy is 0.750000, time 214.075 samples/sec
+epoch 14, total_step 215900, total loss is 13.99 , inference loss is 7.60, weight deacy loss is 6.39, training accuracy is 0.562500, time 223.989 samples/sec
+epoch 14, total_step 215920, total loss is 15.01 , inference loss is 8.62, weight deacy loss is 6.39, training accuracy is 0.531250, time 298.313 samples/sec
+epoch 14, total_step 215940, total loss is 11.70 , inference loss is 5.31, weight deacy loss is 6.39, training accuracy is 0.593750, time 297.323 samples/sec
+epoch 14, total_step 215960, total loss is 14.53 , inference loss is 8.14, weight deacy loss is 6.39, training accuracy is 0.625000, time 225.584 samples/sec
+epoch 14, total_step 215980, total loss is 11.50 , inference loss is 5.10, weight deacy loss is 6.39, training accuracy is 0.656250, time 217.374 samples/sec
+testing verification..
+(12000, 512)
+infer time 36.69806999999997
+best_threshold_index 122 0.9853703703703703
+best_threshold_index 121 0.985
+best_threshold_index 122 0.9840740740740741
+best_threshold_index 122 0.985
+best_threshold_index 121 0.9851851851851852
+best_threshold_index 122 0.9840740740740741
+best_threshold_index 122 0.985
+best_threshold_index 121 0.9851851851851852
+best_threshold_index 122 0.9835185185185186
+best_threshold_index 122 0.9842592592592593
+[lfw][216000]XNorm: 19.926864
+[lfw][216000]Accuracy-Flip: 0.98417+-0.00579
+test accuracy is: 0.9841666666666666
+epoch 14, total_step 216000, total loss is 13.01 , inference loss is 6.62, weight deacy loss is 6.39, training accuracy is 0.593750, time 15.599 samples/sec
+epoch 14, total_step 216020, total loss is 13.46 , inference loss is 7.07, weight deacy loss is 6.39, training accuracy is 0.437500, time 227.250 samples/sec
+epoch 14, total_step 216040, total loss is 15.46 , inference loss is 9.07, weight deacy loss is 6.39, training accuracy is 0.468750, time 297.869 samples/sec
+epoch 14, total_step 216060, total loss is 12.15 , inference loss is 5.76, weight deacy loss is 6.39, training accuracy is 0.718750, time 213.719 samples/sec
+epoch 14, total_step 216080, total loss is 12.68 , inference loss is 6.29, weight deacy loss is 6.39, training accuracy is 0.437500, time 218.929 samples/sec
+epoch 14, total_step 216100, total loss is 12.65 , inference loss is 6.25, weight deacy loss is 6.39, training accuracy is 0.593750, time 220.424 samples/sec
+epoch 14, total_step 216120, total loss is 8.37 , inference loss is 1.98, weight deacy loss is 6.39, training accuracy is 0.750000, time 222.936 samples/sec
+epoch 14, total_step 216140, total loss is 15.54 , inference loss is 9.14, weight deacy loss is 6.39, training accuracy is 0.531250, time 224.102 samples/sec
+epoch 14, total_step 216160, total loss is 12.20 , inference loss is 5.81, weight deacy loss is 6.39, training accuracy is 0.625000, time 224.734 samples/sec
+epoch 14, total_step 216180, total loss is 10.84 , inference loss is 4.45, weight deacy loss is 6.39, training accuracy is 0.656250, time 235.973 samples/sec
+epoch 14, total_step 216200, total loss is 11.60 , inference loss is 5.20, weight deacy loss is 6.39, training accuracy is 0.625000, time 223.856 samples/sec
+epoch 14, total_step 216220, total loss is 10.10 , inference loss is 3.71, weight deacy loss is 6.39, training accuracy is 0.750000, time 215.063 samples/sec
+epoch 14, total_step 216240, total loss is 9.09 , inference loss is 2.70, weight deacy loss is 6.39, training accuracy is 0.625000, time 225.465 samples/sec
+epoch 14, total_step 216260, total loss is 13.17 , inference loss is 6.77, weight deacy loss is 6.39, training accuracy is 0.531250, time 222.045 samples/sec
+epoch 14, total_step 216280, total loss is 11.56 , inference loss is 5.17, weight deacy loss is 6.39, training accuracy is 0.656250, time 217.742 samples/sec
+epoch 14, total_step 216300, total loss is 11.03 , inference loss is 4.64, weight deacy loss is 6.39, training accuracy is 0.750000, time 12.337 samples/sec
+epoch 14, total_step 216320, total loss is 15.44 , inference loss is 9.05, weight deacy loss is 6.39, training accuracy is 0.562500, time 214.562 samples/sec
+epoch 14, total_step 216340, total loss is 13.56 , inference loss is 7.16, weight deacy loss is 6.39, training accuracy is 0.562500, time 227.963 samples/sec
+epoch 14, total_step 216360, total loss is 14.14 , inference loss is 7.75, weight deacy loss is 6.39, training accuracy is 0.531250, time 219.778 samples/sec
+epoch 14, total_step 216380, total loss is 14.81 , inference loss is 8.41, weight deacy loss is 6.40, training accuracy is 0.406250, time 223.543 samples/sec
+epoch 14, total_step 216400, total loss is 11.72 , inference loss is 5.33, weight deacy loss is 6.40, training accuracy is 0.593750, time 212.389 samples/sec
+epoch 14, total_step 216420, total loss is 11.90 , inference loss is 5.50, weight deacy loss is 6.40, training accuracy is 0.625000, time 223.338 samples/sec
+epoch 14, total_step 216440, total loss is 14.74 , inference loss is 8.34, weight deacy loss is 6.40, training accuracy is 0.343750, time 239.401 samples/sec
+epoch 14, total_step 216460, total loss is 15.02 , inference loss is 8.63, weight deacy loss is 6.40, training accuracy is 0.562500, time 227.889 samples/sec
+epoch 14, total_step 216480, total loss is 11.94 , inference loss is 5.55, weight deacy loss is 6.40, training accuracy is 0.656250, time 214.318 samples/sec
+epoch 14, total_step 216500, total loss is 9.21 , inference loss is 2.81, weight deacy loss is 6.40, training accuracy is 0.625000, time 225.665 samples/sec
+epoch 14, total_step 216520, total loss is 14.05 , inference loss is 7.65, weight deacy loss is 6.40, training accuracy is 0.500000, time 230.164 samples/sec
+epoch 14, total_step 216540, total loss is 13.83 , inference loss is 7.43, weight deacy loss is 6.40, training accuracy is 0.562500, time 221.889 samples/sec
+epoch 14, total_step 216560, total loss is 12.53 , inference loss is 6.14, weight deacy loss is 6.40, training accuracy is 0.593750, time 223.255 samples/sec
+epoch 14, total_step 216580, total loss is 10.50 , inference loss is 4.11, weight deacy loss is 6.40, training accuracy is 0.562500, time 224.183 samples/sec
+epoch 14, total_step 216600, total loss is 12.37 , inference loss is 5.98, weight deacy loss is 6.40, training accuracy is 0.531250, time 11.989 samples/sec
+epoch 14, total_step 216620, total loss is 10.80 , inference loss is 4.40, weight deacy loss is 6.40, training accuracy is 0.562500, time 209.255 samples/sec
+epoch 14, total_step 216640, total loss is 17.21 , inference loss is 10.81, weight deacy loss is 6.40, training accuracy is 0.468750, time 233.967 samples/sec
+epoch 14, total_step 216660, total loss is 11.03 , inference loss is 4.64, weight deacy loss is 6.40, training accuracy is 0.718750, time 218.018 samples/sec
+epoch 14, total_step 216680, total loss is 9.54 , inference loss is 3.15, weight deacy loss is 6.40, training accuracy is 0.718750, time 229.918 samples/sec
+epoch 14, total_step 216700, total loss is 10.34 , inference loss is 3.95, weight deacy loss is 6.40, training accuracy is 0.687500, time 222.736 samples/sec
+epoch 14, total_step 216720, total loss is 11.49 , inference loss is 5.09, weight deacy loss is 6.40, training accuracy is 0.593750, time 221.501 samples/sec
+epoch 14, total_step 216740, total loss is 12.49 , inference loss is 6.09, weight deacy loss is 6.40, training accuracy is 0.562500, time 219.956 samples/sec
+epoch 14, total_step 216760, total loss is 9.94 , inference loss is 3.54, weight deacy loss is 6.40, training accuracy is 0.593750, time 226.996 samples/sec
+epoch 14, total_step 216780, total loss is 11.70 , inference loss is 5.30, weight deacy loss is 6.40, training accuracy is 0.687500, time 225.420 samples/sec
+epoch 14, total_step 216800, total loss is 12.92 , inference loss is 6.52, weight deacy loss is 6.40, training accuracy is 0.656250, time 226.123 samples/sec
+epoch 14, total_step 216820, total loss is 9.51 , inference loss is 3.11, weight deacy loss is 6.40, training accuracy is 0.625000, time 215.483 samples/sec
+epoch 14, total_step 216840, total loss is 11.47 , inference loss is 5.08, weight deacy loss is 6.40, training accuracy is 0.593750, time 225.718 samples/sec
+epoch 14, total_step 216860, total loss is 10.95 , inference loss is 4.56, weight deacy loss is 6.40, training accuracy is 0.718750, time 225.631 samples/sec
+epoch 14, total_step 216880, total loss is 11.95 , inference loss is 5.56, weight deacy loss is 6.40, training accuracy is 0.656250, time 298.879 samples/sec
+epoch 14, total_step 216900, total loss is 9.29 , inference loss is 2.90, weight deacy loss is 6.40, training accuracy is 0.687500, time 12.742 samples/sec
+epoch 14, total_step 216920, total loss is 11.54 , inference loss is 5.14, weight deacy loss is 6.40, training accuracy is 0.687500, time 222.615 samples/sec
+epoch 14, total_step 216940, total loss is 15.61 , inference loss is 9.22, weight deacy loss is 6.40, training accuracy is 0.343750, time 219.657 samples/sec
+epoch 14, total_step 216960, total loss is 12.39 , inference loss is 5.99, weight deacy loss is 6.40, training accuracy is 0.593750, time 225.114 samples/sec
+epoch 14, total_step 216980, total loss is 9.17 , inference loss is 2.77, weight deacy loss is 6.40, training accuracy is 0.687500, time 224.701 samples/sec
+epoch 14, total_step 217000, total loss is 7.36 , inference loss is 0.96, weight deacy loss is 6.40, training accuracy is 0.843750, time 229.351 samples/sec
+epoch 14, total_step 217020, total loss is 12.31 , inference loss is 5.92, weight deacy loss is 6.40, training accuracy is 0.687500, time 225.295 samples/sec
+epoch 14, total_step 217040, total loss is 10.86 , inference loss is 4.47, weight deacy loss is 6.40, training accuracy is 0.750000, time 225.569 samples/sec
+epoch 14, total_step 217060, total loss is 9.26 , inference loss is 2.87, weight deacy loss is 6.40, training accuracy is 0.750000, time 217.058 samples/sec
+epoch 14, total_step 217080, total loss is 10.40 , inference loss is 4.01, weight deacy loss is 6.40, training accuracy is 0.718750, time 213.841 samples/sec
+epoch 14, total_step 217100, total loss is 10.29 , inference loss is 3.89, weight deacy loss is 6.39, training accuracy is 0.625000, time 237.320 samples/sec
+epoch 14, total_step 217120, total loss is 10.86 , inference loss is 4.47, weight deacy loss is 6.39, training accuracy is 0.625000, time 215.588 samples/sec
+epoch 14, total_step 217140, total loss is 12.55 , inference loss is 6.15, weight deacy loss is 6.39, training accuracy is 0.593750, time 296.538 samples/sec
+epoch 14, total_step 217160, total loss is 10.16 , inference loss is 3.76, weight deacy loss is 6.39, training accuracy is 0.656250, time 218.935 samples/sec
+epoch 14, total_step 217180, total loss is 11.29 , inference loss is 4.89, weight deacy loss is 6.39, training accuracy is 0.656250, time 222.383 samples/sec
+epoch 14, total_step 217200, total loss is 10.58 , inference loss is 4.19, weight deacy loss is 6.39, training accuracy is 0.562500, time 12.088 samples/sec
+epoch 14, total_step 217220, total loss is 9.95 , inference loss is 3.55, weight deacy loss is 6.39, training accuracy is 0.656250, time 221.872 samples/sec
+epoch 14, total_step 217240, total loss is 10.92 , inference loss is 4.53, weight deacy loss is 6.39, training accuracy is 0.593750, time 216.703 samples/sec
+epoch 14, total_step 217260, total loss is 12.55 , inference loss is 6.16, weight deacy loss is 6.39, training accuracy is 0.687500, time 237.969 samples/sec
+epoch 14, total_step 217280, total loss is 9.13 , inference loss is 2.74, weight deacy loss is 6.39, training accuracy is 0.812500, time 219.197 samples/sec
+epoch 14, total_step 217300, total loss is 11.62 , inference loss is 5.22, weight deacy loss is 6.39, training accuracy is 0.750000, time 235.789 samples/sec
+epoch 14, total_step 217320, total loss is 12.47 , inference loss is 6.07, weight deacy loss is 6.39, training accuracy is 0.625000, time 227.242 samples/sec
+epoch 14, total_step 217340, total loss is 11.73 , inference loss is 5.33, weight deacy loss is 6.39, training accuracy is 0.562500, time 214.301 samples/sec
+epoch 14, total_step 217360, total loss is 14.79 , inference loss is 8.39, weight deacy loss is 6.39, training accuracy is 0.562500, time 228.300 samples/sec
+epoch 14, total_step 217380, total loss is 11.72 , inference loss is 5.32, weight deacy loss is 6.39, training accuracy is 0.687500, time 225.847 samples/sec
+epoch 14, total_step 217400, total loss is 12.50 , inference loss is 6.10, weight deacy loss is 6.39, training accuracy is 0.625000, time 299.403 samples/sec
+epoch 14, total_step 217420, total loss is 13.09 , inference loss is 6.70, weight deacy loss is 6.39, training accuracy is 0.562500, time 216.436 samples/sec
+epoch 14, total_step 217440, total loss is 11.21 , inference loss is 4.82, weight deacy loss is 6.39, training accuracy is 0.562500, time 220.972 samples/sec
+epoch 14, total_step 217460, total loss is 16.54 , inference loss is 10.15, weight deacy loss is 6.39, training accuracy is 0.406250, time 224.476 samples/sec
+epoch 14, total_step 217480, total loss is 10.68 , inference loss is 4.28, weight deacy loss is 6.39, training accuracy is 0.718750, time 178.550 samples/sec
+epoch 14, total_step 217500, total loss is 14.26 , inference loss is 7.87, weight deacy loss is 6.39, training accuracy is 0.468750, time 11.806 samples/sec
+epoch 14, total_step 217520, total loss is 8.48 , inference loss is 2.08, weight deacy loss is 6.39, training accuracy is 0.687500, time 214.855 samples/sec
+epoch 14, total_step 217540, total loss is 10.44 , inference loss is 4.04, weight deacy loss is 6.39, training accuracy is 0.781250, time 216.339 samples/sec
+epoch 14, total_step 217560, total loss is 15.90 , inference loss is 9.50, weight deacy loss is 6.39, training accuracy is 0.593750, time 217.364 samples/sec
+epoch 14, total_step 217580, total loss is 8.72 , inference loss is 2.33, weight deacy loss is 6.39, training accuracy is 0.750000, time 219.551 samples/sec
+epoch 14, total_step 217600, total loss is 10.04 , inference loss is 3.64, weight deacy loss is 6.39, training accuracy is 0.750000, time 214.355 samples/sec
+epoch 14, total_step 217620, total loss is 9.19 , inference loss is 2.80, weight deacy loss is 6.39, training accuracy is 0.781250, time 222.403 samples/sec
+epoch 14, total_step 217640, total loss is 9.02 , inference loss is 2.63, weight deacy loss is 6.39, training accuracy is 0.687500, time 299.845 samples/sec
+epoch 14, total_step 217660, total loss is 11.18 , inference loss is 4.78, weight deacy loss is 6.39, training accuracy is 0.531250, time 208.100 samples/sec
+epoch 14, total_step 217680, total loss is 13.06 , inference loss is 6.67, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.603 samples/sec
+epoch 14, total_step 217700, total loss is 10.58 , inference loss is 4.18, weight deacy loss is 6.39, training accuracy is 0.718750, time 224.696 samples/sec
+epoch 14, total_step 217720, total loss is 12.58 , inference loss is 6.19, weight deacy loss is 6.39, training accuracy is 0.718750, time 225.139 samples/sec
+epoch 14, total_step 217740, total loss is 10.02 , inference loss is 3.63, weight deacy loss is 6.39, training accuracy is 0.656250, time 219.241 samples/sec
+epoch 14, total_step 217760, total loss is 8.36 , inference loss is 1.97, weight deacy loss is 6.39, training accuracy is 0.812500, time 210.712 samples/sec
+epoch 14, total_step 217780, total loss is 13.19 , inference loss is 6.80, weight deacy loss is 6.39, training accuracy is 0.531250, time 228.170 samples/sec
+epoch 14, total_step 217800, total loss is 13.18 , inference loss is 6.78, weight deacy loss is 6.39, training accuracy is 0.593750, time 13.041 samples/sec
+epoch 14, total_step 217820, total loss is 11.30 , inference loss is 4.90, weight deacy loss is 6.39, training accuracy is 0.500000, time 257.603 samples/sec
+epoch 14, total_step 217840, total loss is 10.24 , inference loss is 3.85, weight deacy loss is 6.39, training accuracy is 0.718750, time 221.277 samples/sec
+epoch 14, total_step 217860, total loss is 12.27 , inference loss is 5.88, weight deacy loss is 6.39, training accuracy is 0.625000, time 217.880 samples/sec
+epoch 14, total_step 217880, total loss is 13.78 , inference loss is 7.39, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.714 samples/sec
+epoch 14, total_step 217900, total loss is 10.46 , inference loss is 4.06, weight deacy loss is 6.39, training accuracy is 0.562500, time 227.443 samples/sec
+epoch 14, total_step 217920, total loss is 12.57 , inference loss is 6.17, weight deacy loss is 6.39, training accuracy is 0.406250, time 226.431 samples/sec
+epoch 14, total_step 217940, total loss is 15.29 , inference loss is 8.90, weight deacy loss is 6.39, training accuracy is 0.437500, time 236.828 samples/sec
+epoch 14, total_step 217960, total loss is 16.54 , inference loss is 10.15, weight deacy loss is 6.39, training accuracy is 0.468750, time 226.389 samples/sec
+epoch 14, total_step 217980, total loss is 10.76 , inference loss is 4.37, weight deacy loss is 6.39, training accuracy is 0.687500, time 225.605 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.87530799999998
+best_threshold_index 115 0.9838888888888889
+best_threshold_index 113 0.9840740740740741
+best_threshold_index 114 0.9837037037037037
+best_threshold_index 116 0.985
+best_threshold_index 114 0.9840740740740741
+best_threshold_index 121 0.9831481481481481
+best_threshold_index 114 0.9842592592592593
+best_threshold_index 116 0.9838888888888889
+best_threshold_index 121 0.9822222222222222
+best_threshold_index 121 0.9831481481481481
+[lfw][218000]XNorm: 21.484166
+[lfw][218000]Accuracy-Flip: 0.98117+-0.00654
+test accuracy is: 0.9811666666666665
+epoch 14, total_step 218000, total loss is 10.13 , inference loss is 3.73, weight deacy loss is 6.39, training accuracy is 0.625000, time 16.636 samples/sec
+epoch 14, total_step 218020, total loss is 14.49 , inference loss is 8.09, weight deacy loss is 6.39, training accuracy is 0.531250, time 224.512 samples/sec
+epoch 14, total_step 218040, total loss is 12.57 , inference loss is 6.18, weight deacy loss is 6.39, training accuracy is 0.531250, time 224.407 samples/sec
+epoch 14, total_step 218060, total loss is 13.41 , inference loss is 7.02, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.306 samples/sec
+epoch 14, total_step 218080, total loss is 11.25 , inference loss is 4.86, weight deacy loss is 6.39, training accuracy is 0.656250, time 227.184 samples/sec
+epoch 14, total_step 218100, total loss is 13.27 , inference loss is 6.88, weight deacy loss is 6.39, training accuracy is 0.468750, time 12.249 samples/sec
+epoch 14, total_step 218120, total loss is 11.76 , inference loss is 5.37, weight deacy loss is 6.39, training accuracy is 0.593750, time 225.693 samples/sec
+epoch 14, total_step 218140, total loss is 16.91 , inference loss is 10.51, weight deacy loss is 6.39, training accuracy is 0.500000, time 231.428 samples/sec
+epoch 14, total_step 218160, total loss is 14.97 , inference loss is 8.58, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.583 samples/sec
+epoch 14, total_step 218180, total loss is 11.25 , inference loss is 4.86, weight deacy loss is 6.39, training accuracy is 0.593750, time 298.681 samples/sec
+epoch 14, total_step 218200, total loss is 8.96 , inference loss is 2.57, weight deacy loss is 6.39, training accuracy is 0.687500, time 235.616 samples/sec
+epoch 14, total_step 218220, total loss is 13.06 , inference loss is 6.66, weight deacy loss is 6.39, training accuracy is 0.562500, time 220.864 samples/sec
+epoch 14, total_step 218240, total loss is 13.69 , inference loss is 7.29, weight deacy loss is 6.39, training accuracy is 0.468750, time 221.707 samples/sec
+epoch 14, total_step 218260, total loss is 17.26 , inference loss is 10.86, weight deacy loss is 6.39, training accuracy is 0.500000, time 234.583 samples/sec
+epoch 14, total_step 218280, total loss is 15.71 , inference loss is 9.32, weight deacy loss is 6.39, training accuracy is 0.500000, time 213.890 samples/sec
+epoch 14, total_step 218300, total loss is 17.68 , inference loss is 11.29, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.846 samples/sec
+epoch 14, total_step 218320, total loss is 12.77 , inference loss is 6.38, weight deacy loss is 6.39, training accuracy is 0.562500, time 224.495 samples/sec
+epoch 14, total_step 218340, total loss is 15.56 , inference loss is 9.17, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.406 samples/sec
+epoch 14, total_step 218360, total loss is 13.07 , inference loss is 6.67, weight deacy loss is 6.39, training accuracy is 0.468750, time 224.503 samples/sec
+epoch 14, total_step 218380, total loss is 12.35 , inference loss is 5.96, weight deacy loss is 6.39, training accuracy is 0.593750, time 227.325 samples/sec
+epoch 14, total_step 218400, total loss is 11.39 , inference loss is 5.00, weight deacy loss is 6.39, training accuracy is 0.562500, time 12.383 samples/sec
+epoch 14, total_step 218420, total loss is 12.15 , inference loss is 5.75, weight deacy loss is 6.39, training accuracy is 0.468750, time 226.207 samples/sec
+epoch 14, total_step 218440, total loss is 11.97 , inference loss is 5.57, weight deacy loss is 6.39, training accuracy is 0.656250, time 226.131 samples/sec
+epoch 14, total_step 218460, total loss is 12.08 , inference loss is 5.69, weight deacy loss is 6.39, training accuracy is 0.500000, time 223.342 samples/sec
+epoch 14, total_step 218480, total loss is 11.68 , inference loss is 5.29, weight deacy loss is 6.39, training accuracy is 0.468750, time 223.091 samples/sec
+epoch 14, total_step 218500, total loss is 15.30 , inference loss is 8.90, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.939 samples/sec
+epoch 14, total_step 218520, total loss is 11.75 , inference loss is 5.35, weight deacy loss is 6.39, training accuracy is 0.593750, time 231.203 samples/sec
+epoch 14, total_step 218540, total loss is 12.18 , inference loss is 5.79, weight deacy loss is 6.39, training accuracy is 0.468750, time 217.313 samples/sec
+epoch 14, total_step 218560, total loss is 12.54 , inference loss is 6.14, weight deacy loss is 6.39, training accuracy is 0.406250, time 222.998 samples/sec
+epoch 14, total_step 218580, total loss is 14.53 , inference loss is 8.14, weight deacy loss is 6.39, training accuracy is 0.531250, time 228.380 samples/sec
+epoch 14, total_step 218600, total loss is 16.81 , inference loss is 10.42, weight deacy loss is 6.39, training accuracy is 0.406250, time 216.463 samples/sec
+epoch 14, total_step 218620, total loss is 15.83 , inference loss is 9.43, weight deacy loss is 6.39, training accuracy is 0.500000, time 226.245 samples/sec
+epoch 14, total_step 218640, total loss is 12.37 , inference loss is 5.97, weight deacy loss is 6.39, training accuracy is 0.562500, time 218.730 samples/sec
+epoch 14, total_step 218660, total loss is 10.30 , inference loss is 3.91, weight deacy loss is 6.39, training accuracy is 0.531250, time 220.377 samples/sec
+epoch 14, total_step 218680, total loss is 15.86 , inference loss is 9.46, weight deacy loss is 6.39, training accuracy is 0.343750, time 219.622 samples/sec
+epoch 14, total_step 218700, total loss is 14.10 , inference loss is 7.71, weight deacy loss is 6.39, training accuracy is 0.531250, time 13.328 samples/sec
+epoch 14, total_step 218720, total loss is 13.89 , inference loss is 7.50, weight deacy loss is 6.39, training accuracy is 0.500000, time 214.482 samples/sec
+epoch 14, total_step 218740, total loss is 10.14 , inference loss is 3.75, weight deacy loss is 6.39, training accuracy is 0.625000, time 209.803 samples/sec
+epoch 14, total_step 218760, total loss is 14.66 , inference loss is 8.27, weight deacy loss is 6.39, training accuracy is 0.406250, time 229.000 samples/sec
+epoch 14, total_step 218780, total loss is 17.56 , inference loss is 11.17, weight deacy loss is 6.39, training accuracy is 0.468750, time 228.210 samples/sec
+epoch 14, total_step 218800, total loss is 10.02 , inference loss is 3.63, weight deacy loss is 6.39, training accuracy is 0.531250, time 227.311 samples/sec
+epoch 14, total_step 218820, total loss is 14.78 , inference loss is 8.39, weight deacy loss is 6.39, training accuracy is 0.375000, time 223.365 samples/sec
+epoch 14, total_step 218840, total loss is 11.39 , inference loss is 5.00, weight deacy loss is 6.39, training accuracy is 0.625000, time 220.824 samples/sec
+epoch 14, total_step 218860, total loss is 16.02 , inference loss is 9.62, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.653 samples/sec
+epoch 14, total_step 218880, total loss is 16.04 , inference loss is 9.65, weight deacy loss is 6.39, training accuracy is 0.375000, time 233.264 samples/sec
+epoch 14, total_step 218900, total loss is 16.30 , inference loss is 9.90, weight deacy loss is 6.39, training accuracy is 0.437500, time 228.368 samples/sec
+epoch 14, total_step 218920, total loss is 14.10 , inference loss is 7.70, weight deacy loss is 6.39, training accuracy is 0.468750, time 227.680 samples/sec
+epoch 14, total_step 218940, total loss is 13.99 , inference loss is 7.59, weight deacy loss is 6.39, training accuracy is 0.468750, time 222.886 samples/sec
+epoch 14, total_step 218960, total loss is 15.42 , inference loss is 9.02, weight deacy loss is 6.39, training accuracy is 0.500000, time 211.095 samples/sec
+epoch 14, total_step 218980, total loss is 15.23 , inference loss is 8.83, weight deacy loss is 6.39, training accuracy is 0.375000, time 214.610 samples/sec
+epoch 14, total_step 219000, total loss is 9.79 , inference loss is 3.39, weight deacy loss is 6.39, training accuracy is 0.625000, time 12.011 samples/sec
+epoch 14, total_step 219020, total loss is 10.56 , inference loss is 4.16, weight deacy loss is 6.39, training accuracy is 0.562500, time 221.172 samples/sec
+epoch 14, total_step 219040, total loss is 12.47 , inference loss is 6.07, weight deacy loss is 6.39, training accuracy is 0.562500, time 224.358 samples/sec
+epoch 14, total_step 219060, total loss is 14.85 , inference loss is 8.46, weight deacy loss is 6.39, training accuracy is 0.468750, time 223.153 samples/sec
+epoch 14, total_step 219080, total loss is 13.02 , inference loss is 6.63, weight deacy loss is 6.39, training accuracy is 0.562500, time 215.723 samples/sec
+epoch 14, total_step 219100, total loss is 11.26 , inference loss is 4.87, weight deacy loss is 6.39, training accuracy is 0.687500, time 222.698 samples/sec
+epoch 14, total_step 219120, total loss is 15.88 , inference loss is 9.49, weight deacy loss is 6.39, training accuracy is 0.437500, time 237.989 samples/sec
+epoch 14, total_step 219140, total loss is 11.05 , inference loss is 4.65, weight deacy loss is 6.39, training accuracy is 0.531250, time 242.086 samples/sec
+epoch 14, total_step 219160, total loss is 12.38 , inference loss is 5.99, weight deacy loss is 6.39, training accuracy is 0.562500, time 226.871 samples/sec
+epoch 14, total_step 219180, total loss is 15.24 , inference loss is 8.85, weight deacy loss is 6.39, training accuracy is 0.625000, time 226.425 samples/sec
+epoch 14, total_step 219200, total loss is 15.13 , inference loss is 8.74, weight deacy loss is 6.39, training accuracy is 0.562500, time 219.945 samples/sec
+epoch 14, total_step 219220, total loss is 13.74 , inference loss is 7.35, weight deacy loss is 6.39, training accuracy is 0.468750, time 220.492 samples/sec
+epoch 14, total_step 219240, total loss is 11.06 , inference loss is 4.67, weight deacy loss is 6.39, training accuracy is 0.406250, time 225.169 samples/sec
+epoch 14, total_step 219260, total loss is 13.31 , inference loss is 6.92, weight deacy loss is 6.39, training accuracy is 0.437500, time 240.061 samples/sec
+epoch 14, total_step 219280, total loss is 11.55 , inference loss is 5.16, weight deacy loss is 6.39, training accuracy is 0.656250, time 298.781 samples/sec
+epoch 14, total_step 219300, total loss is 13.25 , inference loss is 6.85, weight deacy loss is 6.39, training accuracy is 0.343750, time 12.795 samples/sec
+epoch 14, total_step 219320, total loss is 14.41 , inference loss is 8.02, weight deacy loss is 6.39, training accuracy is 0.500000, time 214.519 samples/sec
+epoch 14, total_step 219340, total loss is 13.89 , inference loss is 7.50, weight deacy loss is 6.39, training accuracy is 0.250000, time 212.840 samples/sec
+epoch 14, total_step 219360, total loss is 13.35 , inference loss is 6.95, weight deacy loss is 6.39, training accuracy is 0.593750, time 226.235 samples/sec
+epoch 14, total_step 219380, total loss is 15.72 , inference loss is 9.33, weight deacy loss is 6.39, training accuracy is 0.625000, time 300.186 samples/sec
+epoch 14, total_step 219400, total loss is 15.88 , inference loss is 9.48, weight deacy loss is 6.39, training accuracy is 0.437500, time 296.258 samples/sec
+epoch 14, total_step 219420, total loss is 14.58 , inference loss is 8.19, weight deacy loss is 6.39, training accuracy is 0.437500, time 222.621 samples/sec
+epoch 14, total_step 219440, total loss is 13.88 , inference loss is 7.49, weight deacy loss is 6.39, training accuracy is 0.531250, time 225.641 samples/sec
+epoch 14, total_step 219460, total loss is 10.67 , inference loss is 4.28, weight deacy loss is 6.39, training accuracy is 0.593750, time 236.329 samples/sec
+epoch 14, total_step 219480, total loss is 13.54 , inference loss is 7.15, weight deacy loss is 6.39, training accuracy is 0.656250, time 219.471 samples/sec
+epoch 14, total_step 219500, total loss is 11.89 , inference loss is 5.49, weight deacy loss is 6.39, training accuracy is 0.625000, time 228.305 samples/sec
+epoch 14, total_step 219520, total loss is 10.46 , inference loss is 4.06, weight deacy loss is 6.39, training accuracy is 0.593750, time 224.852 samples/sec
+epoch 14, total_step 219540, total loss is 15.13 , inference loss is 8.73, weight deacy loss is 6.39, training accuracy is 0.562500, time 226.125 samples/sec
+epoch 14, total_step 219560, total loss is 16.39 , inference loss is 9.99, weight deacy loss is 6.39, training accuracy is 0.343750, time 226.444 samples/sec
+epoch 14, total_step 219580, total loss is 10.26 , inference loss is 3.86, weight deacy loss is 6.39, training accuracy is 0.562500, time 225.997 samples/sec
+epoch 14, total_step 219600, total loss is 9.31 , inference loss is 2.92, weight deacy loss is 6.39, training accuracy is 0.718750, time 12.954 samples/sec
+epoch 14, total_step 219620, total loss is 17.38 , inference loss is 10.99, weight deacy loss is 6.39, training accuracy is 0.343750, time 211.404 samples/sec
+epoch 14, total_step 219640, total loss is 12.62 , inference loss is 6.23, weight deacy loss is 6.39, training accuracy is 0.531250, time 222.383 samples/sec
+epoch 14, total_step 219660, total loss is 15.30 , inference loss is 8.90, weight deacy loss is 6.39, training accuracy is 0.531250, time 299.532 samples/sec
+epoch 14, total_step 219680, total loss is 13.21 , inference loss is 6.81, weight deacy loss is 6.39, training accuracy is 0.500000, time 219.932 samples/sec
+epoch 14, total_step 219700, total loss is 17.42 , inference loss is 11.03, weight deacy loss is 6.39, training accuracy is 0.468750, time 224.972 samples/sec
+epoch 14, total_step 219720, total loss is 10.19 , inference loss is 3.80, weight deacy loss is 6.39, training accuracy is 0.718750, time 205.631 samples/sec
+epoch 14, total_step 219740, total loss is 13.76 , inference loss is 7.37, weight deacy loss is 6.39, training accuracy is 0.468750, time 224.123 samples/sec
+epoch 14, total_step 219760, total loss is 15.39 , inference loss is 9.00, weight deacy loss is 6.39, training accuracy is 0.593750, time 223.381 samples/sec
+epoch 14, total_step 219780, total loss is 11.36 , inference loss is 4.97, weight deacy loss is 6.39, training accuracy is 0.593750, time 227.455 samples/sec
+epoch 14, total_step 219800, total loss is 10.27 , inference loss is 3.88, weight deacy loss is 6.39, training accuracy is 0.593750, time 227.789 samples/sec
+epoch 14, total_step 219820, total loss is 14.09 , inference loss is 7.70, weight deacy loss is 6.39, training accuracy is 0.406250, time 218.301 samples/sec
+epoch 14, total_step 219840, total loss is 11.95 , inference loss is 5.56, weight deacy loss is 6.39, training accuracy is 0.500000, time 224.791 samples/sec
+epoch 14, total_step 219860, total loss is 11.31 , inference loss is 4.92, weight deacy loss is 6.39, training accuracy is 0.531250, time 194.885 samples/sec
+epoch 14, total_step 219880, total loss is 10.07 , inference loss is 3.68, weight deacy loss is 6.39, training accuracy is 0.593750, time 223.178 samples/sec
+epoch 14, total_step 219900, total loss is 11.10 , inference loss is 4.70, weight deacy loss is 6.39, training accuracy is 0.656250, time 12.205 samples/sec
+epoch 14, total_step 219920, total loss is 10.89 , inference loss is 4.50, weight deacy loss is 6.39, training accuracy is 0.656250, time 298.565 samples/sec
+epoch 14, total_step 219940, total loss is 11.50 , inference loss is 5.11, weight deacy loss is 6.39, training accuracy is 0.468750, time 226.078 samples/sec
+epoch 14, total_step 219960, total loss is 16.32 , inference loss is 9.93, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.664 samples/sec
+epoch 14, total_step 219980, total loss is 13.05 , inference loss is 6.66, weight deacy loss is 6.39, training accuracy is 0.468750, time 226.753 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.759050999999996
+best_threshold_index 123 0.9885185185185185
+best_threshold_index 123 0.9883333333333333
+best_threshold_index 122 0.9875925925925926
+best_threshold_index 118 0.9887037037037038
+best_threshold_index 123 0.9885185185185185
+best_threshold_index 123 0.9874074074074074
+best_threshold_index 118 0.9875925925925926
+best_threshold_index 122 0.9879629629629629
+best_threshold_index 123 0.9868518518518519
+best_threshold_index 123 0.987037037037037
+[lfw][220000]XNorm: 19.728749
+[lfw][220000]Accuracy-Flip: 0.98667+-0.00601
+test accuracy is: 0.9866666666666666
+epoch 14, total_step 220000, total loss is 14.31 , inference loss is 7.92, weight deacy loss is 6.39, training accuracy is 0.468750, time 16.166 samples/sec
+epoch 14, total_step 220020, total loss is 13.08 , inference loss is 6.69, weight deacy loss is 6.39, training accuracy is 0.343750, time 223.952 samples/sec
+epoch 14, total_step 220040, total loss is 12.27 , inference loss is 5.88, weight deacy loss is 6.39, training accuracy is 0.562500, time 227.958 samples/sec
+epoch 14, total_step 220060, total loss is 10.51 , inference loss is 4.12, weight deacy loss is 6.39, training accuracy is 0.625000, time 223.089 samples/sec
+epoch 14, total_step 220080, total loss is 13.26 , inference loss is 6.86, weight deacy loss is 6.39, training accuracy is 0.468750, time 218.322 samples/sec
+epoch 14, total_step 220100, total loss is 10.72 , inference loss is 4.32, weight deacy loss is 6.39, training accuracy is 0.500000, time 241.762 samples/sec
+epoch 14, total_step 220120, total loss is 9.67 , inference loss is 3.27, weight deacy loss is 6.39, training accuracy is 0.531250, time 189.881 samples/sec
+epoch 14, total_step 220140, total loss is 15.63 , inference loss is 9.23, weight deacy loss is 6.39, training accuracy is 0.312500, time 198.611 samples/sec
+epoch 14, total_step 220160, total loss is 15.33 , inference loss is 8.94, weight deacy loss is 6.39, training accuracy is 0.531250, time 191.842 samples/sec
+epoch 14, total_step 220180, total loss is 10.64 , inference loss is 4.25, weight deacy loss is 6.39, training accuracy is 0.593750, time 216.867 samples/sec
+epoch 14, total_step 220200, total loss is 16.27 , inference loss is 9.87, weight deacy loss is 6.39, training accuracy is 0.312500, time 13.438 samples/sec
+epoch 14, total_step 220220, total loss is 12.78 , inference loss is 6.39, weight deacy loss is 6.39, training accuracy is 0.562500, time 182.198 samples/sec
+epoch 14, total_step 220240, total loss is 14.36 , inference loss is 7.97, weight deacy loss is 6.39, training accuracy is 0.562500, time 298.868 samples/sec
+epoch 14, total_step 220260, total loss is 12.73 , inference loss is 6.34, weight deacy loss is 6.39, training accuracy is 0.625000, time 218.981 samples/sec
+epoch 14, total_step 220280, total loss is 12.37 , inference loss is 5.97, weight deacy loss is 6.39, training accuracy is 0.656250, time 227.070 samples/sec
+epoch 14, total_step 220300, total loss is 13.79 , inference loss is 7.40, weight deacy loss is 6.39, training accuracy is 0.593750, time 197.305 samples/sec
+epoch 14, total_step 220320, total loss is 12.43 , inference loss is 6.03, weight deacy loss is 6.39, training accuracy is 0.562500, time 225.347 samples/sec
+epoch 14, total_step 220340, total loss is 11.07 , inference loss is 4.68, weight deacy loss is 6.39, training accuracy is 0.718750, time 202.599 samples/sec
+epoch 14, total_step 220360, total loss is 13.25 , inference loss is 6.85, weight deacy loss is 6.39, training accuracy is 0.500000, time 214.857 samples/sec
+epoch 14, total_step 220380, total loss is 15.14 , inference loss is 8.75, weight deacy loss is 6.39, training accuracy is 0.406250, time 224.006 samples/sec
+epoch 14, total_step 220400, total loss is 15.04 , inference loss is 8.65, weight deacy loss is 6.39, training accuracy is 0.562500, time 218.229 samples/sec
+epoch 14, total_step 220420, total loss is 13.31 , inference loss is 6.92, weight deacy loss is 6.39, training accuracy is 0.500000, time 208.093 samples/sec
+epoch 14, total_step 220440, total loss is 14.98 , inference loss is 8.59, weight deacy loss is 6.39, training accuracy is 0.531250, time 179.010 samples/sec
+epoch 14, total_step 220460, total loss is 8.90 , inference loss is 2.51, weight deacy loss is 6.39, training accuracy is 0.656250, time 223.835 samples/sec
+epoch 14, total_step 220480, total loss is 16.47 , inference loss is 10.08, weight deacy loss is 6.39, training accuracy is 0.437500, time 221.784 samples/sec
+epoch 14, total_step 220500, total loss is 15.08 , inference loss is 8.68, weight deacy loss is 6.39, training accuracy is 0.468750, time 12.977 samples/sec
+epoch 14, total_step 220520, total loss is 15.38 , inference loss is 8.99, weight deacy loss is 6.39, training accuracy is 0.437500, time 220.497 samples/sec
+epoch 14, total_step 220540, total loss is 12.97 , inference loss is 6.58, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.633 samples/sec
+epoch 14, total_step 220560, total loss is 13.22 , inference loss is 6.83, weight deacy loss is 6.39, training accuracy is 0.406250, time 238.234 samples/sec
+epoch 14, total_step 220580, total loss is 16.19 , inference loss is 9.80, weight deacy loss is 6.39, training accuracy is 0.437500, time 228.448 samples/sec
+epoch 14, total_step 220600, total loss is 11.58 , inference loss is 5.18, weight deacy loss is 6.39, training accuracy is 0.562500, time 239.477 samples/sec
+epoch 14, total_step 220620, total loss is 11.73 , inference loss is 5.34, weight deacy loss is 6.39, training accuracy is 0.531250, time 175.093 samples/sec
+epoch 14, total_step 220640, total loss is 13.68 , inference loss is 7.29, weight deacy loss is 6.39, training accuracy is 0.343750, time 218.989 samples/sec
+epoch 14, total_step 220660, total loss is 12.68 , inference loss is 6.28, weight deacy loss is 6.39, training accuracy is 0.687500, time 233.397 samples/sec
+epoch 14, total_step 220680, total loss is 13.23 , inference loss is 6.84, weight deacy loss is 6.39, training accuracy is 0.625000, time 222.281 samples/sec
+epoch 14, total_step 220700, total loss is 14.42 , inference loss is 8.03, weight deacy loss is 6.39, training accuracy is 0.343750, time 204.795 samples/sec
+epoch 14, total_step 220720, total loss is 15.57 , inference loss is 9.18, weight deacy loss is 6.39, training accuracy is 0.437500, time 230.773 samples/sec
+epoch 14, total_step 220740, total loss is 13.50 , inference loss is 7.11, weight deacy loss is 6.39, training accuracy is 0.406250, time 225.121 samples/sec
+epoch 14, total_step 220760, total loss is 12.71 , inference loss is 6.32, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.104 samples/sec
+epoch 14, total_step 220780, total loss is 11.26 , inference loss is 4.87, weight deacy loss is 6.39, training accuracy is 0.531250, time 227.202 samples/sec
+epoch 14, total_step 220800, total loss is 17.11 , inference loss is 10.71, weight deacy loss is 6.39, training accuracy is 0.531250, time 12.546 samples/sec
+epoch 14, total_step 220820, total loss is 11.33 , inference loss is 4.93, weight deacy loss is 6.39, training accuracy is 0.656250, time 228.262 samples/sec
+epoch 14, total_step 220840, total loss is 12.65 , inference loss is 6.26, weight deacy loss is 6.39, training accuracy is 0.500000, time 219.262 samples/sec
+epoch 14, total_step 220860, total loss is 13.26 , inference loss is 6.87, weight deacy loss is 6.39, training accuracy is 0.531250, time 219.538 samples/sec
+epoch 14, total_step 220880, total loss is 14.08 , inference loss is 7.68, weight deacy loss is 6.39, training accuracy is 0.437500, time 224.439 samples/sec
+epoch 14, total_step 220900, total loss is 14.41 , inference loss is 8.02, weight deacy loss is 6.39, training accuracy is 0.468750, time 226.986 samples/sec
+epoch 14, total_step 220920, total loss is 15.10 , inference loss is 8.71, weight deacy loss is 6.39, training accuracy is 0.437500, time 226.521 samples/sec
+epoch 14, total_step 220940, total loss is 16.25 , inference loss is 9.86, weight deacy loss is 6.39, training accuracy is 0.531250, time 225.385 samples/sec
+epoch 14, total_step 220960, total loss is 13.17 , inference loss is 6.78, weight deacy loss is 6.39, training accuracy is 0.406250, time 215.390 samples/sec
+epoch 14, total_step 220980, total loss is 11.18 , inference loss is 4.78, weight deacy loss is 6.39, training accuracy is 0.687500, time 217.791 samples/sec
+epoch 14, total_step 221000, total loss is 12.79 , inference loss is 6.40, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.552 samples/sec
+epoch 14, total_step 221020, total loss is 12.55 , inference loss is 6.16, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.448 samples/sec
+epoch 14, total_step 221040, total loss is 12.38 , inference loss is 5.99, weight deacy loss is 6.39, training accuracy is 0.437500, time 215.537 samples/sec
+epoch 14, total_step 221060, total loss is 14.75 , inference loss is 8.36, weight deacy loss is 6.39, training accuracy is 0.468750, time 215.039 samples/sec
+epoch 14, total_step 221080, total loss is 9.78 , inference loss is 3.38, weight deacy loss is 6.39, training accuracy is 0.687500, time 215.816 samples/sec
+epoch 14, total_step 221100, total loss is 8.07 , inference loss is 1.68, weight deacy loss is 6.39, training accuracy is 0.812500, time 13.600 samples/sec
+epoch 14, total_step 221120, total loss is 15.10 , inference loss is 8.71, weight deacy loss is 6.39, training accuracy is 0.375000, time 216.175 samples/sec
+epoch 14, total_step 221140, total loss is 12.00 , inference loss is 5.61, weight deacy loss is 6.39, training accuracy is 0.593750, time 222.310 samples/sec
+epoch 14, total_step 221160, total loss is 13.78 , inference loss is 7.38, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.875 samples/sec
+epoch 14, total_step 221180, total loss is 13.72 , inference loss is 7.33, weight deacy loss is 6.39, training accuracy is 0.531250, time 234.305 samples/sec
+epoch 14, total_step 221200, total loss is 15.84 , inference loss is 9.44, weight deacy loss is 6.39, training accuracy is 0.343750, time 295.954 samples/sec
+epoch 14, total_step 221220, total loss is 12.17 , inference loss is 5.78, weight deacy loss is 6.39, training accuracy is 0.531250, time 218.634 samples/sec
+epoch 14, total_step 221240, total loss is 13.69 , inference loss is 7.30, weight deacy loss is 6.39, training accuracy is 0.593750, time 223.353 samples/sec
+epoch 14, total_step 221260, total loss is 12.10 , inference loss is 5.71, weight deacy loss is 6.39, training accuracy is 0.531250, time 221.734 samples/sec
+epoch 14, total_step 221280, total loss is 10.74 , inference loss is 4.35, weight deacy loss is 6.39, training accuracy is 0.531250, time 221.236 samples/sec
+epoch 14, total_step 221300, total loss is 12.76 , inference loss is 6.37, weight deacy loss is 6.39, training accuracy is 0.437500, time 222.200 samples/sec
+epoch 14, total_step 221320, total loss is 10.74 , inference loss is 4.35, weight deacy loss is 6.39, training accuracy is 0.468750, time 222.095 samples/sec
+epoch 14, total_step 221340, total loss is 15.47 , inference loss is 9.08, weight deacy loss is 6.39, training accuracy is 0.312500, time 213.151 samples/sec
+epoch 14, total_step 221360, total loss is 13.73 , inference loss is 7.34, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.601 samples/sec
+epoch 14, total_step 221380, total loss is 14.73 , inference loss is 8.34, weight deacy loss is 6.39, training accuracy is 0.531250, time 215.668 samples/sec
+epoch 14, total_step 221400, total loss is 13.08 , inference loss is 6.69, weight deacy loss is 6.39, training accuracy is 0.625000, time 12.186 samples/sec
+epoch 14, total_step 221420, total loss is 11.55 , inference loss is 5.16, weight deacy loss is 6.39, training accuracy is 0.593750, time 222.867 samples/sec
+epoch 14, total_step 221440, total loss is 9.59 , inference loss is 3.20, weight deacy loss is 6.39, training accuracy is 0.687500, time 174.390 samples/sec
+epoch 14, total_step 221460, total loss is 13.46 , inference loss is 7.06, weight deacy loss is 6.39, training accuracy is 0.437500, time 298.511 samples/sec
+epoch 14, total_step 221480, total loss is 14.72 , inference loss is 8.33, weight deacy loss is 6.39, training accuracy is 0.500000, time 224.643 samples/sec
+epoch 14, total_step 221500, total loss is 12.69 , inference loss is 6.30, weight deacy loss is 6.39, training accuracy is 0.468750, time 176.293 samples/sec
+epoch 14, total_step 221520, total loss is 11.51 , inference loss is 5.11, weight deacy loss is 6.39, training accuracy is 0.500000, time 233.010 samples/sec
+epoch 14, total_step 221540, total loss is 11.92 , inference loss is 5.53, weight deacy loss is 6.39, training accuracy is 0.625000, time 221.125 samples/sec
+epoch 14, total_step 221560, total loss is 17.81 , inference loss is 11.42, weight deacy loss is 6.39, training accuracy is 0.343750, time 216.965 samples/sec
+epoch 14, total_step 221580, total loss is 10.57 , inference loss is 4.18, weight deacy loss is 6.39, training accuracy is 0.562500, time 207.475 samples/sec
+epoch 14, total_step 221600, total loss is 13.49 , inference loss is 7.10, weight deacy loss is 6.39, training accuracy is 0.406250, time 223.375 samples/sec
+epoch 14, total_step 221620, total loss is 9.96 , inference loss is 3.57, weight deacy loss is 6.39, training accuracy is 0.687500, time 230.312 samples/sec
+epoch 14, total_step 221640, total loss is 10.01 , inference loss is 3.62, weight deacy loss is 6.39, training accuracy is 0.500000, time 229.823 samples/sec
+epoch 14, total_step 221660, total loss is 12.36 , inference loss is 5.97, weight deacy loss is 6.39, training accuracy is 0.531250, time 175.711 samples/sec
+epoch 14, total_step 221680, total loss is 12.67 , inference loss is 6.28, weight deacy loss is 6.39, training accuracy is 0.406250, time 209.446 samples/sec
+epoch 14, total_step 221700, total loss is 14.96 , inference loss is 8.57, weight deacy loss is 6.39, training accuracy is 0.343750, time 12.323 samples/sec
+epoch 14, total_step 221720, total loss is 11.47 , inference loss is 5.08, weight deacy loss is 6.39, training accuracy is 0.531250, time 226.788 samples/sec
+epoch 14, total_step 221740, total loss is 14.98 , inference loss is 8.59, weight deacy loss is 6.39, training accuracy is 0.375000, time 230.937 samples/sec
+epoch 14, total_step 221760, total loss is 9.49 , inference loss is 3.10, weight deacy loss is 6.39, training accuracy is 0.500000, time 226.561 samples/sec
+epoch 14, total_step 221780, total loss is 9.65 , inference loss is 3.26, weight deacy loss is 6.39, training accuracy is 0.656250, time 223.227 samples/sec
+epoch 14, total_step 221800, total loss is 16.58 , inference loss is 10.19, weight deacy loss is 6.39, training accuracy is 0.468750, time 213.131 samples/sec
+epoch 14, total_step 221820, total loss is 13.24 , inference loss is 6.85, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.511 samples/sec
+epoch 14, total_step 221840, total loss is 12.31 , inference loss is 5.92, weight deacy loss is 6.39, training accuracy is 0.625000, time 204.935 samples/sec
+epoch 14, total_step 221860, total loss is 14.61 , inference loss is 8.22, weight deacy loss is 6.39, training accuracy is 0.437500, time 216.474 samples/sec
+epoch 14, total_step 221880, total loss is 14.03 , inference loss is 7.64, weight deacy loss is 6.39, training accuracy is 0.593750, time 221.995 samples/sec
+epoch 14, total_step 221900, total loss is 16.09 , inference loss is 9.70, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.102 samples/sec
+epoch 14, total_step 221920, total loss is 16.79 , inference loss is 10.40, weight deacy loss is 6.39, training accuracy is 0.500000, time 215.491 samples/sec
+epoch 14, total_step 221940, total loss is 11.61 , inference loss is 5.22, weight deacy loss is 6.39, training accuracy is 0.562500, time 221.712 samples/sec
+epoch 14, total_step 221960, total loss is 18.23 , inference loss is 11.84, weight deacy loss is 6.39, training accuracy is 0.312500, time 231.739 samples/sec
+epoch 14, total_step 221980, total loss is 14.22 , inference loss is 7.83, weight deacy loss is 6.39, training accuracy is 0.468750, time 226.519 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.328447000000022
+best_threshold_index 121 0.99
+best_threshold_index 121 0.9894444444444445
+best_threshold_index 121 0.9890740740740741
+best_threshold_index 121 0.9905555555555555
+best_threshold_index 121 0.9903703703703703
+best_threshold_index 121 0.9894444444444445
+best_threshold_index 121 0.9898148148148148
+best_threshold_index 121 0.9894444444444445
+best_threshold_index 121 0.9890740740740741
+best_threshold_index 121 0.9894444444444445
+[lfw][222000]XNorm: 22.846084
+[lfw][222000]Accuracy-Flip: 0.98967+-0.00433
+test accuracy is: 0.9896666666666667
+epoch 14, total_step 222000, total loss is 11.86 , inference loss is 5.47, weight deacy loss is 6.39, training accuracy is 0.718750, time 16.021 samples/sec
+epoch 14, total_step 222020, total loss is 11.65 , inference loss is 5.26, weight deacy loss is 6.39, training accuracy is 0.750000, time 237.484 samples/sec
+epoch 14, total_step 222040, total loss is 13.42 , inference loss is 7.03, weight deacy loss is 6.39, training accuracy is 0.562500, time 225.110 samples/sec
+epoch 14, total_step 222060, total loss is 16.75 , inference loss is 10.36, weight deacy loss is 6.39, training accuracy is 0.406250, time 224.792 samples/sec
+epoch 14, total_step 222080, total loss is 10.14 , inference loss is 3.75, weight deacy loss is 6.39, training accuracy is 0.687500, time 295.891 samples/sec
+epoch 14, total_step 222100, total loss is 13.28 , inference loss is 6.89, weight deacy loss is 6.39, training accuracy is 0.468750, time 219.764 samples/sec
+epoch 14, total_step 222120, total loss is 12.77 , inference loss is 6.38, weight deacy loss is 6.39, training accuracy is 0.468750, time 224.728 samples/sec
+epoch 14, total_step 222140, total loss is 9.11 , inference loss is 2.72, weight deacy loss is 6.39, training accuracy is 0.593750, time 235.171 samples/sec
+epoch 14, total_step 222160, total loss is 10.88 , inference loss is 4.49, weight deacy loss is 6.39, training accuracy is 0.531250, time 218.547 samples/sec
+epoch 14, total_step 222180, total loss is 11.92 , inference loss is 5.53, weight deacy loss is 6.39, training accuracy is 0.625000, time 231.566 samples/sec
+epoch 14, total_step 222200, total loss is 11.41 , inference loss is 5.02, weight deacy loss is 6.39, training accuracy is 0.593750, time 224.569 samples/sec
+epoch 14, total_step 222220, total loss is 14.05 , inference loss is 7.67, weight deacy loss is 6.39, training accuracy is 0.593750, time 215.469 samples/sec
+epoch 14, total_step 222240, total loss is 15.25 , inference loss is 8.86, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.082 samples/sec
+epoch 14, total_step 222260, total loss is 11.44 , inference loss is 5.05, weight deacy loss is 6.39, training accuracy is 0.593750, time 217.470 samples/sec
+epoch 14, total_step 222280, total loss is 10.56 , inference loss is 4.17, weight deacy loss is 6.39, training accuracy is 0.593750, time 234.137 samples/sec
+epoch 14, total_step 222300, total loss is 11.11 , inference loss is 4.73, weight deacy loss is 6.39, training accuracy is 0.562500, time 11.532 samples/sec
+epoch 14, total_step 222320, total loss is 13.11 , inference loss is 6.72, weight deacy loss is 6.39, training accuracy is 0.468750, time 210.954 samples/sec
+epoch 14, total_step 222340, total loss is 16.37 , inference loss is 9.98, weight deacy loss is 6.39, training accuracy is 0.500000, time 296.899 samples/sec
+epoch 14, total_step 222360, total loss is 9.20 , inference loss is 2.81, weight deacy loss is 6.39, training accuracy is 0.718750, time 225.190 samples/sec
+epoch 14, total_step 222380, total loss is 14.37 , inference loss is 7.99, weight deacy loss is 6.39, training accuracy is 0.437500, time 221.817 samples/sec
+epoch 14, total_step 222400, total loss is 12.71 , inference loss is 6.32, weight deacy loss is 6.39, training accuracy is 0.531250, time 213.969 samples/sec
+epoch 14, total_step 222420, total loss is 13.78 , inference loss is 7.40, weight deacy loss is 6.39, training accuracy is 0.437500, time 224.736 samples/sec
+epoch 14, total_step 222440, total loss is 9.05 , inference loss is 2.66, weight deacy loss is 6.39, training accuracy is 0.656250, time 215.026 samples/sec
+epoch 14, total_step 222460, total loss is 13.98 , inference loss is 7.59, weight deacy loss is 6.39, training accuracy is 0.531250, time 223.962 samples/sec
+epoch 14, total_step 222480, total loss is 16.58 , inference loss is 10.20, weight deacy loss is 6.39, training accuracy is 0.406250, time 221.610 samples/sec
+epoch 14, total_step 222500, total loss is 14.40 , inference loss is 8.01, weight deacy loss is 6.39, training accuracy is 0.562500, time 215.493 samples/sec
+epoch 14, total_step 222520, total loss is 15.96 , inference loss is 9.57, weight deacy loss is 6.39, training accuracy is 0.312500, time 233.241 samples/sec
+epoch 14, total_step 222540, total loss is 11.37 , inference loss is 4.99, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.842 samples/sec
+epoch 14, total_step 222560, total loss is 11.49 , inference loss is 5.10, weight deacy loss is 6.39, training accuracy is 0.531250, time 213.564 samples/sec
+epoch 14, total_step 222580, total loss is 11.98 , inference loss is 5.60, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.406 samples/sec
+epoch 14, total_step 222600, total loss is 11.99 , inference loss is 5.61, weight deacy loss is 6.39, training accuracy is 0.531250, time 12.417 samples/sec
+epoch 14, total_step 222620, total loss is 13.58 , inference loss is 7.19, weight deacy loss is 6.39, training accuracy is 0.500000, time 223.578 samples/sec
+epoch 14, total_step 222640, total loss is 13.15 , inference loss is 6.76, weight deacy loss is 6.39, training accuracy is 0.531250, time 226.301 samples/sec
+epoch 14, total_step 222660, total loss is 12.20 , inference loss is 5.82, weight deacy loss is 6.39, training accuracy is 0.562500, time 214.687 samples/sec
+epoch 14, total_step 222680, total loss is 12.44 , inference loss is 6.06, weight deacy loss is 6.39, training accuracy is 0.531250, time 223.962 samples/sec
+epoch 14, total_step 222700, total loss is 12.30 , inference loss is 5.92, weight deacy loss is 6.39, training accuracy is 0.500000, time 216.440 samples/sec
+epoch 14, total_step 222720, total loss is 12.44 , inference loss is 6.05, weight deacy loss is 6.39, training accuracy is 0.343750, time 218.452 samples/sec
+epoch 14, total_step 222740, total loss is 13.37 , inference loss is 6.99, weight deacy loss is 6.39, training accuracy is 0.562500, time 222.075 samples/sec
+epoch 14, total_step 222760, total loss is 11.21 , inference loss is 4.83, weight deacy loss is 6.39, training accuracy is 0.687500, time 214.307 samples/sec
+epoch 14, total_step 222780, total loss is 17.06 , inference loss is 10.68, weight deacy loss is 6.39, training accuracy is 0.406250, time 217.831 samples/sec
+epoch 14, total_step 222800, total loss is 13.45 , inference loss is 7.06, weight deacy loss is 6.39, training accuracy is 0.406250, time 214.684 samples/sec
+epoch 14, total_step 222820, total loss is 12.07 , inference loss is 5.68, weight deacy loss is 6.39, training accuracy is 0.468750, time 224.617 samples/sec
+epoch 14, total_step 222840, total loss is 11.54 , inference loss is 5.15, weight deacy loss is 6.39, training accuracy is 0.375000, time 223.131 samples/sec
+epoch 14, total_step 222860, total loss is 13.24 , inference loss is 6.86, weight deacy loss is 6.39, training accuracy is 0.468750, time 298.839 samples/sec
+epoch 14, total_step 222880, total loss is 12.73 , inference loss is 6.34, weight deacy loss is 6.39, training accuracy is 0.562500, time 297.214 samples/sec
+epoch 14, total_step 222900, total loss is 15.15 , inference loss is 8.77, weight deacy loss is 6.39, training accuracy is 0.343750, time 12.466 samples/sec
+epoch 14, total_step 222920, total loss is 15.21 , inference loss is 8.83, weight deacy loss is 6.39, training accuracy is 0.531250, time 225.391 samples/sec
+epoch 14, total_step 222940, total loss is 14.64 , inference loss is 8.25, weight deacy loss is 6.39, training accuracy is 0.218750, time 224.079 samples/sec
+epoch 14, total_step 222960, total loss is 13.48 , inference loss is 7.10, weight deacy loss is 6.39, training accuracy is 0.406250, time 232.268 samples/sec
+epoch 14, total_step 222980, total loss is 13.55 , inference loss is 7.16, weight deacy loss is 6.39, training accuracy is 0.531250, time 222.862 samples/sec
+epoch 14, total_step 223000, total loss is 15.02 , inference loss is 8.63, weight deacy loss is 6.39, training accuracy is 0.343750, time 232.932 samples/sec
+epoch 14, total_step 223020, total loss is 13.20 , inference loss is 6.81, weight deacy loss is 6.39, training accuracy is 0.625000, time 218.559 samples/sec
+epoch 14, total_step 223040, total loss is 14.16 , inference loss is 7.77, weight deacy loss is 6.39, training accuracy is 0.656250, time 299.024 samples/sec
+epoch 14, total_step 223060, total loss is 14.01 , inference loss is 7.62, weight deacy loss is 6.39, training accuracy is 0.406250, time 224.695 samples/sec
+epoch 14, total_step 223080, total loss is 11.61 , inference loss is 5.22, weight deacy loss is 6.39, training accuracy is 0.406250, time 223.976 samples/sec
+epoch 14, total_step 223100, total loss is 10.83 , inference loss is 4.44, weight deacy loss is 6.39, training accuracy is 0.625000, time 225.660 samples/sec
+epoch 14, total_step 223120, total loss is 12.49 , inference loss is 6.10, weight deacy loss is 6.39, training accuracy is 0.468750, time 227.337 samples/sec
+epoch 14, total_step 223140, total loss is 16.28 , inference loss is 9.90, weight deacy loss is 6.39, training accuracy is 0.531250, time 218.015 samples/sec
+epoch 14, total_step 223160, total loss is 13.02 , inference loss is 6.63, weight deacy loss is 6.39, training accuracy is 0.531250, time 224.857 samples/sec
+epoch 14, total_step 223180, total loss is 18.91 , inference loss is 12.53, weight deacy loss is 6.39, training accuracy is 0.375000, time 225.699 samples/sec
+epoch 14, total_step 223200, total loss is 11.20 , inference loss is 4.81, weight deacy loss is 6.39, training accuracy is 0.531250, time 13.626 samples/sec
+epoch 14, total_step 223220, total loss is 13.06 , inference loss is 6.68, weight deacy loss is 6.39, training accuracy is 0.656250, time 219.583 samples/sec
+epoch 14, total_step 223240, total loss is 13.08 , inference loss is 6.69, weight deacy loss is 6.39, training accuracy is 0.531250, time 219.150 samples/sec
+epoch 14, total_step 223260, total loss is 13.52 , inference loss is 7.13, weight deacy loss is 6.39, training accuracy is 0.468750, time 224.832 samples/sec
+epoch 14, total_step 223280, total loss is 16.29 , inference loss is 9.90, weight deacy loss is 6.39, training accuracy is 0.218750, time 219.422 samples/sec
+epoch 14, total_step 223300, total loss is 12.52 , inference loss is 6.13, weight deacy loss is 6.39, training accuracy is 0.468750, time 226.828 samples/sec
+epoch 14, total_step 223320, total loss is 14.92 , inference loss is 8.53, weight deacy loss is 6.39, training accuracy is 0.375000, time 179.557 samples/sec
+epoch 14, total_step 223340, total loss is 12.97 , inference loss is 6.58, weight deacy loss is 6.39, training accuracy is 0.531250, time 215.167 samples/sec
+epoch 14, total_step 223360, total loss is 13.66 , inference loss is 7.27, weight deacy loss is 6.39, training accuracy is 0.531250, time 219.313 samples/sec
+epoch 14, total_step 223380, total loss is 11.90 , inference loss is 5.51, weight deacy loss is 6.39, training accuracy is 0.562500, time 220.234 samples/sec
+epoch 14, total_step 223400, total loss is 14.74 , inference loss is 8.36, weight deacy loss is 6.39, training accuracy is 0.406250, time 260.086 samples/sec
+epoch 14, total_step 223420, total loss is 13.79 , inference loss is 7.40, weight deacy loss is 6.39, training accuracy is 0.500000, time 221.480 samples/sec
+epoch 14, total_step 223440, total loss is 15.24 , inference loss is 8.85, weight deacy loss is 6.39, training accuracy is 0.500000, time 239.001 samples/sec
+epoch 14, total_step 223460, total loss is 10.37 , inference loss is 3.98, weight deacy loss is 6.39, training accuracy is 0.593750, time 222.645 samples/sec
+epoch 14, total_step 223480, total loss is 14.61 , inference loss is 8.22, weight deacy loss is 6.39, training accuracy is 0.406250, time 220.462 samples/sec
+epoch 14, total_step 223500, total loss is 10.25 , inference loss is 3.86, weight deacy loss is 6.39, training accuracy is 0.593750, time 12.107 samples/sec
+epoch 14, total_step 223520, total loss is 11.51 , inference loss is 5.13, weight deacy loss is 6.39, training accuracy is 0.500000, time 213.981 samples/sec
+epoch 14, total_step 223540, total loss is 15.21 , inference loss is 8.82, weight deacy loss is 6.39, training accuracy is 0.375000, time 225.625 samples/sec
+epoch 14, total_step 223560, total loss is 12.06 , inference loss is 5.68, weight deacy loss is 6.39, training accuracy is 0.468750, time 227.572 samples/sec
+epoch 14, total_step 223580, total loss is 11.99 , inference loss is 5.61, weight deacy loss is 6.39, training accuracy is 0.500000, time 203.254 samples/sec
+epoch 14, total_step 223600, total loss is 17.31 , inference loss is 10.92, weight deacy loss is 6.39, training accuracy is 0.406250, time 224.934 samples/sec
+epoch 14, total_step 223620, total loss is 11.29 , inference loss is 4.91, weight deacy loss is 6.39, training accuracy is 0.593750, time 224.320 samples/sec
+epoch 14, total_step 223640, total loss is 11.45 , inference loss is 5.07, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.128 samples/sec
+epoch 14, total_step 223660, total loss is 11.29 , inference loss is 4.90, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.197 samples/sec
+epoch 14, total_step 223680, total loss is 14.04 , inference loss is 7.65, weight deacy loss is 6.39, training accuracy is 0.531250, time 215.113 samples/sec
+epoch 14, total_step 223700, total loss is 12.18 , inference loss is 5.79, weight deacy loss is 6.39, training accuracy is 0.500000, time 220.897 samples/sec
+epoch 14, total_step 223720, total loss is 10.61 , inference loss is 4.22, weight deacy loss is 6.39, training accuracy is 0.718750, time 224.690 samples/sec
+epoch 14, total_step 223740, total loss is 11.89 , inference loss is 5.51, weight deacy loss is 6.39, training accuracy is 0.656250, time 237.873 samples/sec
+epoch 14, total_step 223760, total loss is 11.76 , inference loss is 5.38, weight deacy loss is 6.39, training accuracy is 0.687500, time 223.615 samples/sec
+epoch 14, total_step 223780, total loss is 15.07 , inference loss is 8.68, weight deacy loss is 6.39, training accuracy is 0.343750, time 221.660 samples/sec
+epoch 14, total_step 223800, total loss is 13.12 , inference loss is 6.74, weight deacy loss is 6.39, training accuracy is 0.531250, time 12.220 samples/sec
+epoch 14, total_step 223820, total loss is 12.36 , inference loss is 5.97, weight deacy loss is 6.39, training accuracy is 0.562500, time 225.680 samples/sec
+epoch 14, total_step 223840, total loss is 15.33 , inference loss is 8.94, weight deacy loss is 6.39, training accuracy is 0.625000, time 296.523 samples/sec
+epoch 14, total_step 223860, total loss is 13.72 , inference loss is 7.33, weight deacy loss is 6.39, training accuracy is 0.562500, time 223.424 samples/sec
+epoch 14, total_step 223880, total loss is 18.24 , inference loss is 11.85, weight deacy loss is 6.39, training accuracy is 0.375000, time 220.889 samples/sec
+epoch 14, total_step 223900, total loss is 11.37 , inference loss is 4.99, weight deacy loss is 6.39, training accuracy is 0.625000, time 217.302 samples/sec
+epoch 14, total_step 223920, total loss is 10.85 , inference loss is 4.47, weight deacy loss is 6.39, training accuracy is 0.468750, time 224.965 samples/sec
+epoch 14, total_step 223940, total loss is 12.77 , inference loss is 6.38, weight deacy loss is 6.39, training accuracy is 0.531250, time 229.269 samples/sec
+epoch 14, total_step 223960, total loss is 13.09 , inference loss is 6.71, weight deacy loss is 6.39, training accuracy is 0.500000, time 227.578 samples/sec
+epoch 14, total_step 223980, total loss is 16.65 , inference loss is 10.27, weight deacy loss is 6.39, training accuracy is 0.437500, time 219.601 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.396721000000007
+best_threshold_index 116 0.9872222222222222
+best_threshold_index 115 0.9862962962962963
+best_threshold_index 108 0.9864814814814815
+best_threshold_index 116 0.9875925925925926
+best_threshold_index 116 0.9872222222222222
+best_threshold_index 116 0.9866666666666667
+best_threshold_index 108 0.9868518518518519
+best_threshold_index 108 0.9862962962962963
+best_threshold_index 116 0.9853703703703703
+best_threshold_index 116 0.9864814814814815
+[lfw][224000]XNorm: 23.009132
+[lfw][224000]Accuracy-Flip: 0.98400+-0.00528
+test accuracy is: 0.984
+epoch 14, total_step 224000, total loss is 12.64 , inference loss is 6.25, weight deacy loss is 6.38, training accuracy is 0.625000, time 16.727 samples/sec
+epoch 14, total_step 224020, total loss is 9.88 , inference loss is 3.49, weight deacy loss is 6.38, training accuracy is 0.593750, time 211.410 samples/sec
+epoch 14, total_step 224040, total loss is 12.98 , inference loss is 6.60, weight deacy loss is 6.38, training accuracy is 0.531250, time 219.294 samples/sec
+epoch 14, total_step 224060, total loss is 11.38 , inference loss is 4.99, weight deacy loss is 6.38, training accuracy is 0.468750, time 226.838 samples/sec
+epoch 14, total_step 224080, total loss is 10.49 , inference loss is 4.10, weight deacy loss is 6.38, training accuracy is 0.500000, time 225.797 samples/sec
+epoch 14, total_step 224100, total loss is 12.48 , inference loss is 6.10, weight deacy loss is 6.38, training accuracy is 0.562500, time 12.385 samples/sec
+epoch 14, total_step 224120, total loss is 13.24 , inference loss is 6.86, weight deacy loss is 6.38, training accuracy is 0.468750, time 244.200 samples/sec
+epoch 14, total_step 224140, total loss is 12.09 , inference loss is 5.70, weight deacy loss is 6.38, training accuracy is 0.500000, time 220.839 samples/sec
+epoch 14, total_step 224160, total loss is 12.16 , inference loss is 5.77, weight deacy loss is 6.38, training accuracy is 0.500000, time 222.421 samples/sec
+epoch 14, total_step 224180, total loss is 10.92 , inference loss is 4.54, weight deacy loss is 6.38, training accuracy is 0.562500, time 199.509 samples/sec
+epoch 14, total_step 224200, total loss is 14.66 , inference loss is 8.27, weight deacy loss is 6.38, training accuracy is 0.531250, time 297.797 samples/sec
+epoch 14, total_step 224220, total loss is 12.80 , inference loss is 6.42, weight deacy loss is 6.38, training accuracy is 0.500000, time 217.125 samples/sec
+epoch 14, total_step 224240, total loss is 11.28 , inference loss is 4.90, weight deacy loss is 6.38, training accuracy is 0.625000, time 220.565 samples/sec
+epoch 14, total_step 224260, total loss is 15.78 , inference loss is 9.40, weight deacy loss is 6.38, training accuracy is 0.468750, time 217.164 samples/sec
+epoch 14, total_step 224280, total loss is 14.89 , inference loss is 8.51, weight deacy loss is 6.38, training accuracy is 0.468750, time 229.179 samples/sec
+epoch 14, total_step 224300, total loss is 10.68 , inference loss is 4.30, weight deacy loss is 6.38, training accuracy is 0.593750, time 181.218 samples/sec
+epoch 14, total_step 224320, total loss is 13.42 , inference loss is 7.03, weight deacy loss is 6.38, training accuracy is 0.437500, time 220.543 samples/sec
+epoch 14, total_step 224340, total loss is 12.74 , inference loss is 6.36, weight deacy loss is 6.38, training accuracy is 0.531250, time 222.807 samples/sec
+epoch 14, total_step 224360, total loss is 9.45 , inference loss is 3.07, weight deacy loss is 6.38, training accuracy is 0.625000, time 223.029 samples/sec
+epoch 14, total_step 224380, total loss is 12.78 , inference loss is 6.39, weight deacy loss is 6.38, training accuracy is 0.500000, time 225.445 samples/sec
+epoch 14, total_step 224400, total loss is 13.12 , inference loss is 6.74, weight deacy loss is 6.38, training accuracy is 0.625000, time 12.271 samples/sec
+epoch 14, total_step 224420, total loss is 12.45 , inference loss is 6.06, weight deacy loss is 6.38, training accuracy is 0.531250, time 216.014 samples/sec
+epoch 14, total_step 224440, total loss is 12.28 , inference loss is 5.90, weight deacy loss is 6.38, training accuracy is 0.593750, time 222.938 samples/sec
+epoch 14, total_step 224460, total loss is 13.35 , inference loss is 6.97, weight deacy loss is 6.38, training accuracy is 0.437500, time 299.000 samples/sec
+epoch 14, total_step 224480, total loss is 13.28 , inference loss is 6.90, weight deacy loss is 6.38, training accuracy is 0.500000, time 223.876 samples/sec
+epoch 14, total_step 224500, total loss is 13.60 , inference loss is 7.21, weight deacy loss is 6.38, training accuracy is 0.500000, time 218.534 samples/sec
+epoch 14, total_step 224520, total loss is 11.62 , inference loss is 5.23, weight deacy loss is 6.38, training accuracy is 0.562500, time 229.386 samples/sec
+epoch 14, total_step 224540, total loss is 11.06 , inference loss is 4.68, weight deacy loss is 6.38, training accuracy is 0.593750, time 215.245 samples/sec
+epoch 14, total_step 224560, total loss is 10.22 , inference loss is 3.84, weight deacy loss is 6.38, training accuracy is 0.531250, time 216.655 samples/sec
+epoch 14, total_step 224580, total loss is 17.31 , inference loss is 10.93, weight deacy loss is 6.38, training accuracy is 0.468750, time 214.146 samples/sec
+epoch 14, total_step 224600, total loss is 10.88 , inference loss is 4.50, weight deacy loss is 6.38, training accuracy is 0.468750, time 227.584 samples/sec
+epoch 14, total_step 224620, total loss is 13.12 , inference loss is 6.74, weight deacy loss is 6.38, training accuracy is 0.500000, time 296.246 samples/sec
+epoch 14, total_step 224640, total loss is 12.40 , inference loss is 6.01, weight deacy loss is 6.38, training accuracy is 0.593750, time 211.458 samples/sec
+epoch 14, total_step 224660, total loss is 11.41 , inference loss is 5.03, weight deacy loss is 6.38, training accuracy is 0.531250, time 219.208 samples/sec
+epoch 14, total_step 224680, total loss is 12.14 , inference loss is 5.76, weight deacy loss is 6.38, training accuracy is 0.500000, time 227.668 samples/sec
+epoch 14, total_step 224700, total loss is 13.07 , inference loss is 6.69, weight deacy loss is 6.38, training accuracy is 0.406250, time 12.684 samples/sec
+epoch 14, total_step 224720, total loss is 15.38 , inference loss is 9.00, weight deacy loss is 6.38, training accuracy is 0.343750, time 220.148 samples/sec
+epoch 14, total_step 224740, total loss is 10.20 , inference loss is 3.82, weight deacy loss is 6.38, training accuracy is 0.656250, time 225.498 samples/sec
+epoch 14, total_step 224760, total loss is 9.40 , inference loss is 3.02, weight deacy loss is 6.38, training accuracy is 0.562500, time 229.561 samples/sec
+epoch 14, total_step 224780, total loss is 11.10 , inference loss is 4.72, weight deacy loss is 6.38, training accuracy is 0.500000, time 227.235 samples/sec
+epoch 14, total_step 224800, total loss is 12.08 , inference loss is 5.70, weight deacy loss is 6.38, training accuracy is 0.625000, time 228.197 samples/sec
+epoch 14, total_step 224820, total loss is 13.48 , inference loss is 7.10, weight deacy loss is 6.38, training accuracy is 0.500000, time 225.570 samples/sec
+epoch 14, total_step 224840, total loss is 12.88 , inference loss is 6.50, weight deacy loss is 6.38, training accuracy is 0.468750, time 225.286 samples/sec
+epoch 14, total_step 224860, total loss is 9.72 , inference loss is 3.34, weight deacy loss is 6.38, training accuracy is 0.531250, time 235.036 samples/sec
+epoch 14, total_step 224880, total loss is 12.05 , inference loss is 5.67, weight deacy loss is 6.38, training accuracy is 0.625000, time 220.154 samples/sec
+epoch 14, total_step 224900, total loss is 13.56 , inference loss is 7.18, weight deacy loss is 6.38, training accuracy is 0.593750, time 201.598 samples/sec
+epoch 14, total_step 224920, total loss is 12.44 , inference loss is 6.06, weight deacy loss is 6.38, training accuracy is 0.437500, time 240.205 samples/sec
+epoch 14, total_step 224940, total loss is 13.04 , inference loss is 6.65, weight deacy loss is 6.38, training accuracy is 0.437500, time 227.827 samples/sec
+epoch 14, total_step 224960, total loss is 12.26 , inference loss is 5.88, weight deacy loss is 6.38, training accuracy is 0.437500, time 222.509 samples/sec
+epoch 14, total_step 224980, total loss is 11.48 , inference loss is 5.10, weight deacy loss is 6.38, training accuracy is 0.468750, time 221.415 samples/sec
+epoch 14, total_step 225000, total loss is 12.99 , inference loss is 6.61, weight deacy loss is 6.38, training accuracy is 0.500000, time 12.833 samples/sec
+epoch 14, total_step 225020, total loss is 12.83 , inference loss is 6.45, weight deacy loss is 6.38, training accuracy is 0.500000, time 222.713 samples/sec
+epoch 14, total_step 225040, total loss is 12.85 , inference loss is 6.47, weight deacy loss is 6.38, training accuracy is 0.625000, time 236.508 samples/sec
+epoch 14, total_step 225060, total loss is 13.47 , inference loss is 7.09, weight deacy loss is 6.38, training accuracy is 0.531250, time 215.542 samples/sec
+epoch 14, total_step 225080, total loss is 13.36 , inference loss is 6.98, weight deacy loss is 6.38, training accuracy is 0.531250, time 220.131 samples/sec
+epoch 14, total_step 225100, total loss is 12.90 , inference loss is 6.51, weight deacy loss is 6.38, training accuracy is 0.375000, time 213.983 samples/sec
+epoch 14, total_step 225120, total loss is 10.20 , inference loss is 3.82, weight deacy loss is 6.38, training accuracy is 0.593750, time 229.173 samples/sec
+epoch 14, total_step 225140, total loss is 10.76 , inference loss is 4.38, weight deacy loss is 6.38, training accuracy is 0.625000, time 178.879 samples/sec
+epoch 14, total_step 225160, total loss is 13.53 , inference loss is 7.14, weight deacy loss is 6.38, training accuracy is 0.437500, time 211.445 samples/sec
+epoch 14, total_step 225180, total loss is 11.48 , inference loss is 5.09, weight deacy loss is 6.38, training accuracy is 0.531250, time 224.934 samples/sec
+epoch 14, total_step 225200, total loss is 15.20 , inference loss is 8.81, weight deacy loss is 6.38, training accuracy is 0.531250, time 219.357 samples/sec
+epoch 14, total_step 225220, total loss is 13.11 , inference loss is 6.73, weight deacy loss is 6.38, training accuracy is 0.500000, time 198.950 samples/sec
+epoch 14, total_step 225240, total loss is 9.53 , inference loss is 3.14, weight deacy loss is 6.38, training accuracy is 0.562500, time 221.954 samples/sec
+epoch 14, total_step 225260, total loss is 12.51 , inference loss is 6.13, weight deacy loss is 6.38, training accuracy is 0.531250, time 214.537 samples/sec
+epoch 14, total_step 225280, total loss is 12.96 , inference loss is 6.58, weight deacy loss is 6.38, training accuracy is 0.625000, time 179.909 samples/sec
+epoch 14, total_step 225300, total loss is 13.94 , inference loss is 7.56, weight deacy loss is 6.38, training accuracy is 0.406250, time 12.985 samples/sec
+epoch 14, total_step 225320, total loss is 13.45 , inference loss is 7.07, weight deacy loss is 6.38, training accuracy is 0.500000, time 184.821 samples/sec
+epoch 14, total_step 225340, total loss is 16.79 , inference loss is 10.41, weight deacy loss is 6.38, training accuracy is 0.468750, time 225.349 samples/sec
+epoch 14, total_step 225360, total loss is 11.27 , inference loss is 4.89, weight deacy loss is 6.38, training accuracy is 0.500000, time 196.229 samples/sec
+epoch 14, total_step 225380, total loss is 10.35 , inference loss is 3.97, weight deacy loss is 6.38, training accuracy is 0.687500, time 214.817 samples/sec
+epoch 14, total_step 225400, total loss is 16.18 , inference loss is 9.80, weight deacy loss is 6.38, training accuracy is 0.468750, time 295.990 samples/sec
+epoch 14, total_step 225420, total loss is 11.38 , inference loss is 5.00, weight deacy loss is 6.38, training accuracy is 0.625000, time 221.985 samples/sec
+epoch 14, total_step 225440, total loss is 13.28 , inference loss is 6.90, weight deacy loss is 6.38, training accuracy is 0.406250, time 178.028 samples/sec
+epoch 14, total_step 225460, total loss is 11.25 , inference loss is 4.86, weight deacy loss is 6.38, training accuracy is 0.500000, time 218.989 samples/sec
+epoch 14, total_step 225480, total loss is 13.35 , inference loss is 6.97, weight deacy loss is 6.38, training accuracy is 0.562500, time 198.411 samples/sec
+epoch 14, total_step 225500, total loss is 9.64 , inference loss is 3.26, weight deacy loss is 6.38, training accuracy is 0.687500, time 241.412 samples/sec
+epoch 14, total_step 225520, total loss is 11.13 , inference loss is 4.75, weight deacy loss is 6.38, training accuracy is 0.531250, time 190.286 samples/sec
+epoch 14, total_step 225540, total loss is 15.20 , inference loss is 8.82, weight deacy loss is 6.38, training accuracy is 0.406250, time 201.492 samples/sec
+epoch 14, total_step 225560, total loss is 12.78 , inference loss is 6.40, weight deacy loss is 6.38, training accuracy is 0.562500, time 209.842 samples/sec
+epoch 14, total_step 225580, total loss is 12.48 , inference loss is 6.10, weight deacy loss is 6.38, training accuracy is 0.437500, time 228.079 samples/sec
+epoch 14, total_step 225600, total loss is 10.71 , inference loss is 4.33, weight deacy loss is 6.38, training accuracy is 0.437500, time 12.224 samples/sec
+epoch 14, total_step 225620, total loss is 12.96 , inference loss is 6.58, weight deacy loss is 6.38, training accuracy is 0.500000, time 220.500 samples/sec
+epoch 14, total_step 225640, total loss is 14.84 , inference loss is 8.46, weight deacy loss is 6.38, training accuracy is 0.500000, time 212.901 samples/sec
+epoch 14, total_step 225660, total loss is 13.35 , inference loss is 6.97, weight deacy loss is 6.38, training accuracy is 0.437500, time 229.625 samples/sec
+epoch 14, total_step 225680, total loss is 18.10 , inference loss is 11.72, weight deacy loss is 6.38, training accuracy is 0.375000, time 224.230 samples/sec
+epoch 14, total_step 225700, total loss is 14.54 , inference loss is 8.16, weight deacy loss is 6.38, training accuracy is 0.500000, time 228.178 samples/sec
+epoch 14, total_step 225720, total loss is 12.38 , inference loss is 6.00, weight deacy loss is 6.38, training accuracy is 0.500000, time 216.386 samples/sec
+epoch 14, total_step 225740, total loss is 13.94 , inference loss is 7.56, weight deacy loss is 6.38, training accuracy is 0.562500, time 218.351 samples/sec
+epoch 14, total_step 225760, total loss is 11.02 , inference loss is 4.64, weight deacy loss is 6.38, training accuracy is 0.562500, time 237.999 samples/sec
+epoch 14, total_step 225780, total loss is 14.32 , inference loss is 7.93, weight deacy loss is 6.38, training accuracy is 0.531250, time 199.952 samples/sec
+epoch 14, total_step 225800, total loss is 12.03 , inference loss is 5.65, weight deacy loss is 6.38, training accuracy is 0.562500, time 233.066 samples/sec
+epoch 14, total_step 225820, total loss is 16.96 , inference loss is 10.58, weight deacy loss is 6.38, training accuracy is 0.375000, time 218.581 samples/sec
+epoch 14, total_step 225840, total loss is 13.99 , inference loss is 7.61, weight deacy loss is 6.38, training accuracy is 0.531250, time 220.285 samples/sec
+epoch 14, total_step 225860, total loss is 12.38 , inference loss is 6.00, weight deacy loss is 6.38, training accuracy is 0.375000, time 216.723 samples/sec
+epoch 14, total_step 225880, total loss is 15.59 , inference loss is 9.21, weight deacy loss is 6.38, training accuracy is 0.468750, time 242.917 samples/sec
+epoch 14, total_step 225900, total loss is 12.16 , inference loss is 5.78, weight deacy loss is 6.38, training accuracy is 0.468750, time 12.330 samples/sec
+epoch 14, total_step 225920, total loss is 13.78 , inference loss is 7.40, weight deacy loss is 6.38, training accuracy is 0.406250, time 175.359 samples/sec
+epoch 14, total_step 225940, total loss is 13.90 , inference loss is 7.52, weight deacy loss is 6.38, training accuracy is 0.500000, time 225.025 samples/sec
+epoch 14, total_step 225960, total loss is 13.95 , inference loss is 7.57, weight deacy loss is 6.38, training accuracy is 0.531250, time 220.414 samples/sec
+epoch 14, total_step 225980, total loss is 13.66 , inference loss is 7.28, weight deacy loss is 6.38, training accuracy is 0.500000, time 217.229 samples/sec
+testing verification..
+(12000, 512)
+infer time 26.001047000000014
+best_threshold_index 109 0.9853703703703703
+best_threshold_index 109 0.9846296296296296
+best_threshold_index 109 0.9848148148148148
+best_threshold_index 109 0.9855555555555555
+best_threshold_index 109 0.9853703703703703
+best_threshold_index 109 0.9838888888888889
+best_threshold_index 105 0.985
+best_threshold_index 109 0.9844444444444445
+best_threshold_index 105 0.9833333333333333
+best_threshold_index 105 0.9844444444444445
+[lfw][226000]XNorm: 19.447554
+[lfw][226000]Accuracy-Flip: 0.98400+-0.00569
+test accuracy is: 0.984
+epoch 14, total_step 226000, total loss is 14.32 , inference loss is 7.94, weight deacy loss is 6.38, training accuracy is 0.375000, time 16.249 samples/sec
+epoch 14, total_step 226020, total loss is 10.58 , inference loss is 4.20, weight deacy loss is 6.38, training accuracy is 0.656250, time 224.257 samples/sec
+epoch 14, total_step 226040, total loss is 10.93 , inference loss is 4.55, weight deacy loss is 6.38, training accuracy is 0.500000, time 220.867 samples/sec
+epoch 14, total_step 226060, total loss is 15.12 , inference loss is 8.74, weight deacy loss is 6.38, training accuracy is 0.406250, time 225.089 samples/sec
+epoch 14, total_step 226080, total loss is 15.15 , inference loss is 8.77, weight deacy loss is 6.38, training accuracy is 0.500000, time 211.902 samples/sec
+epoch 14, total_step 226100, total loss is 16.15 , inference loss is 9.77, weight deacy loss is 6.38, training accuracy is 0.343750, time 229.876 samples/sec
+epoch 14, total_step 226120, total loss is 9.97 , inference loss is 3.59, weight deacy loss is 6.38, training accuracy is 0.531250, time 220.660 samples/sec
+epoch 14, total_step 226140, total loss is 12.51 , inference loss is 6.13, weight deacy loss is 6.38, training accuracy is 0.625000, time 221.202 samples/sec
+epoch 14, total_step 226160, total loss is 11.91 , inference loss is 5.54, weight deacy loss is 6.38, training accuracy is 0.406250, time 226.164 samples/sec
+epoch 14, total_step 226180, total loss is 16.43 , inference loss is 10.05, weight deacy loss is 6.38, training accuracy is 0.437500, time 221.572 samples/sec
+epoch 14, total_step 226200, total loss is 12.20 , inference loss is 5.82, weight deacy loss is 6.38, training accuracy is 0.531250, time 12.962 samples/sec
+epoch 14, total_step 226220, total loss is 14.94 , inference loss is 8.56, weight deacy loss is 6.38, training accuracy is 0.281250, time 212.740 samples/sec
+epoch 14, total_step 226240, total loss is 17.87 , inference loss is 11.49, weight deacy loss is 6.38, training accuracy is 0.250000, time 218.538 samples/sec
+epoch 14, total_step 226260, total loss is 10.57 , inference loss is 4.19, weight deacy loss is 6.38, training accuracy is 0.625000, time 214.971 samples/sec
+epoch 14, total_step 226280, total loss is 15.62 , inference loss is 9.24, weight deacy loss is 6.38, training accuracy is 0.500000, time 212.639 samples/sec
+epoch 14, total_step 226300, total loss is 14.96 , inference loss is 8.58, weight deacy loss is 6.38, training accuracy is 0.500000, time 208.578 samples/sec
+epoch 14, total_step 226320, total loss is 16.68 , inference loss is 10.30, weight deacy loss is 6.38, training accuracy is 0.375000, time 241.497 samples/sec
+epoch 14, total_step 226340, total loss is 16.04 , inference loss is 9.66, weight deacy loss is 6.38, training accuracy is 0.437500, time 298.327 samples/sec
+epoch 14, total_step 226360, total loss is 12.39 , inference loss is 6.01, weight deacy loss is 6.38, training accuracy is 0.531250, time 297.109 samples/sec
+epoch 14, total_step 226380, total loss is 12.74 , inference loss is 6.36, weight deacy loss is 6.38, training accuracy is 0.500000, time 226.127 samples/sec
+epoch 14, total_step 226400, total loss is 13.80 , inference loss is 7.43, weight deacy loss is 6.38, training accuracy is 0.406250, time 228.358 samples/sec
+epoch 14, total_step 226420, total loss is 14.10 , inference loss is 7.72, weight deacy loss is 6.38, training accuracy is 0.406250, time 243.657 samples/sec
+epoch 14, total_step 226440, total loss is 14.54 , inference loss is 8.16, weight deacy loss is 6.38, training accuracy is 0.531250, time 219.023 samples/sec
+epoch 14, total_step 226460, total loss is 14.17 , inference loss is 7.80, weight deacy loss is 6.38, training accuracy is 0.375000, time 224.249 samples/sec
+epoch 14, total_step 226480, total loss is 11.66 , inference loss is 5.29, weight deacy loss is 6.38, training accuracy is 0.468750, time 299.200 samples/sec
+epoch 14, total_step 226500, total loss is 14.71 , inference loss is 8.34, weight deacy loss is 6.38, training accuracy is 0.468750, time 12.378 samples/sec
+epoch 14, total_step 226520, total loss is 12.80 , inference loss is 6.42, weight deacy loss is 6.38, training accuracy is 0.562500, time 226.177 samples/sec
+epoch 14, total_step 226540, total loss is 16.02 , inference loss is 9.64, weight deacy loss is 6.38, training accuracy is 0.406250, time 220.214 samples/sec
+epoch 14, total_step 226560, total loss is 12.27 , inference loss is 5.89, weight deacy loss is 6.38, training accuracy is 0.437500, time 215.856 samples/sec
+epoch 14, total_step 226580, total loss is 13.26 , inference loss is 6.88, weight deacy loss is 6.38, training accuracy is 0.437500, time 226.090 samples/sec
+epoch 14, total_step 226600, total loss is 9.53 , inference loss is 3.15, weight deacy loss is 6.38, training accuracy is 0.687500, time 224.013 samples/sec
+epoch 14, total_step 226620, total loss is 16.06 , inference loss is 9.68, weight deacy loss is 6.38, training accuracy is 0.343750, time 223.596 samples/sec
+epoch 14, total_step 226640, total loss is 12.73 , inference loss is 6.35, weight deacy loss is 6.38, training accuracy is 0.375000, time 218.141 samples/sec
+epoch 14, total_step 226660, total loss is 15.39 , inference loss is 9.01, weight deacy loss is 6.38, training accuracy is 0.406250, time 227.266 samples/sec
+epoch 14, total_step 226680, total loss is 15.37 , inference loss is 8.99, weight deacy loss is 6.38, training accuracy is 0.437500, time 221.246 samples/sec
+epoch 14, total_step 226700, total loss is 14.74 , inference loss is 8.36, weight deacy loss is 6.38, training accuracy is 0.375000, time 233.942 samples/sec
+epoch 14, total_step 226720, total loss is 15.46 , inference loss is 9.08, weight deacy loss is 6.38, training accuracy is 0.375000, time 227.625 samples/sec
+epoch 14, total_step 226740, total loss is 11.28 , inference loss is 4.90, weight deacy loss is 6.38, training accuracy is 0.437500, time 297.781 samples/sec
+epoch 14, total_step 226760, total loss is 13.27 , inference loss is 6.89, weight deacy loss is 6.38, training accuracy is 0.375000, time 213.754 samples/sec
+epoch 14, total_step 226780, total loss is 15.57 , inference loss is 9.19, weight deacy loss is 6.38, training accuracy is 0.281250, time 220.713 samples/sec
+epoch 14, total_step 226800, total loss is 12.36 , inference loss is 5.99, weight deacy loss is 6.38, training accuracy is 0.531250, time 12.985 samples/sec
+epoch 14, total_step 226820, total loss is 13.22 , inference loss is 6.85, weight deacy loss is 6.38, training accuracy is 0.406250, time 218.626 samples/sec
+epoch 14, total_step 226840, total loss is 14.45 , inference loss is 8.07, weight deacy loss is 6.38, training accuracy is 0.406250, time 222.860 samples/sec
+epoch 14, total_step 226860, total loss is 13.27 , inference loss is 6.90, weight deacy loss is 6.38, training accuracy is 0.375000, time 226.318 samples/sec
+epoch 14, total_step 226880, total loss is 12.83 , inference loss is 6.46, weight deacy loss is 6.38, training accuracy is 0.343750, time 226.806 samples/sec
+epoch 14, total_step 226900, total loss is 13.32 , inference loss is 6.94, weight deacy loss is 6.38, training accuracy is 0.562500, time 228.002 samples/sec
+epoch 14, total_step 226920, total loss is 14.17 , inference loss is 7.79, weight deacy loss is 6.38, training accuracy is 0.406250, time 223.415 samples/sec
+epoch 14, total_step 226940, total loss is 15.24 , inference loss is 8.87, weight deacy loss is 6.38, training accuracy is 0.250000, time 222.975 samples/sec
+epoch 14, total_step 226960, total loss is 12.77 , inference loss is 6.39, weight deacy loss is 6.38, training accuracy is 0.437500, time 224.026 samples/sec
+epoch 14, total_step 226980, total loss is 16.64 , inference loss is 10.27, weight deacy loss is 6.38, training accuracy is 0.250000, time 219.845 samples/sec
+epoch 14, total_step 227000, total loss is 13.76 , inference loss is 7.38, weight deacy loss is 6.38, training accuracy is 0.375000, time 299.372 samples/sec
+epoch 14, total_step 227020, total loss is 15.14 , inference loss is 8.77, weight deacy loss is 6.38, training accuracy is 0.375000, time 223.746 samples/sec
+epoch 14, total_step 227040, total loss is 18.65 , inference loss is 12.28, weight deacy loss is 6.38, training accuracy is 0.406250, time 221.716 samples/sec
+epoch 14, total_step 227060, total loss is 12.77 , inference loss is 6.39, weight deacy loss is 6.38, training accuracy is 0.375000, time 223.152 samples/sec
+epoch 14, total_step 227080, total loss is 12.78 , inference loss is 6.41, weight deacy loss is 6.38, training accuracy is 0.500000, time 213.991 samples/sec
+epoch 14, total_step 227100, total loss is 15.02 , inference loss is 8.64, weight deacy loss is 6.38, training accuracy is 0.375000, time 12.137 samples/sec
+epoch 14, total_step 227120, total loss is 14.95 , inference loss is 8.57, weight deacy loss is 6.38, training accuracy is 0.468750, time 227.684 samples/sec
+epoch 14, total_step 227140, total loss is 13.28 , inference loss is 6.90, weight deacy loss is 6.38, training accuracy is 0.312500, time 205.716 samples/sec
+epoch 14, total_step 227160, total loss is 13.67 , inference loss is 7.29, weight deacy loss is 6.38, training accuracy is 0.500000, time 222.494 samples/sec
+epoch 14, total_step 227180, total loss is 11.96 , inference loss is 5.58, weight deacy loss is 6.38, training accuracy is 0.468750, time 215.366 samples/sec
+epoch 14, total_step 227200, total loss is 16.52 , inference loss is 10.15, weight deacy loss is 6.38, training accuracy is 0.437500, time 215.322 samples/sec
+epoch 14, total_step 227220, total loss is 12.96 , inference loss is 6.58, weight deacy loss is 6.38, training accuracy is 0.562500, time 221.000 samples/sec
+epoch 14, total_step 227240, total loss is 14.43 , inference loss is 8.05, weight deacy loss is 6.38, training accuracy is 0.281250, time 223.977 samples/sec
+epoch 14, total_step 227260, total loss is 11.63 , inference loss is 5.26, weight deacy loss is 6.38, training accuracy is 0.562500, time 295.416 samples/sec
+epoch 14, total_step 227280, total loss is 15.28 , inference loss is 8.91, weight deacy loss is 6.38, training accuracy is 0.406250, time 226.221 samples/sec
+epoch 14, total_step 227300, total loss is 12.63 , inference loss is 6.25, weight deacy loss is 6.38, training accuracy is 0.562500, time 221.044 samples/sec
+epoch 14, total_step 227320, total loss is 16.30 , inference loss is 9.92, weight deacy loss is 6.38, training accuracy is 0.437500, time 220.566 samples/sec
+epoch 14, total_step 227340, total loss is 13.44 , inference loss is 7.07, weight deacy loss is 6.38, training accuracy is 0.406250, time 225.702 samples/sec
+epoch 14, total_step 227360, total loss is 11.72 , inference loss is 5.34, weight deacy loss is 6.38, training accuracy is 0.625000, time 231.805 samples/sec
+epoch 14, total_step 227380, total loss is 14.40 , inference loss is 8.02, weight deacy loss is 6.38, training accuracy is 0.406250, time 222.893 samples/sec
+epoch 14, total_step 227400, total loss is 16.75 , inference loss is 10.37, weight deacy loss is 6.38, training accuracy is 0.437500, time 12.468 samples/sec
+epoch 14, total_step 227420, total loss is 13.04 , inference loss is 6.66, weight deacy loss is 6.38, training accuracy is 0.531250, time 229.902 samples/sec
+epoch 14, total_step 227440, total loss is 9.89 , inference loss is 3.52, weight deacy loss is 6.38, training accuracy is 0.562500, time 225.307 samples/sec
+epoch 14, total_step 227460, total loss is 15.18 , inference loss is 8.81, weight deacy loss is 6.38, training accuracy is 0.343750, time 227.921 samples/sec
+epoch 14, total_step 227480, total loss is 17.45 , inference loss is 11.08, weight deacy loss is 6.38, training accuracy is 0.281250, time 220.375 samples/sec
+epoch 14, total_step 227500, total loss is 17.07 , inference loss is 10.69, weight deacy loss is 6.38, training accuracy is 0.187500, time 232.128 samples/sec
+epoch 14, total_step 227520, total loss is 13.16 , inference loss is 6.78, weight deacy loss is 6.38, training accuracy is 0.406250, time 297.464 samples/sec
+epoch 14, total_step 227540, total loss is 15.30 , inference loss is 8.93, weight deacy loss is 6.37, training accuracy is 0.468750, time 219.469 samples/sec
+epoch 14, total_step 227560, total loss is 16.97 , inference loss is 10.60, weight deacy loss is 6.37, training accuracy is 0.281250, time 213.494 samples/sec
+epoch 14, total_step 227580, total loss is 15.18 , inference loss is 8.81, weight deacy loss is 6.37, training accuracy is 0.468750, time 228.544 samples/sec
+epoch 14, total_step 227600, total loss is 15.05 , inference loss is 8.67, weight deacy loss is 6.37, training accuracy is 0.343750, time 228.124 samples/sec
+epoch 14, total_step 227620, total loss is 13.41 , inference loss is 7.03, weight deacy loss is 6.37, training accuracy is 0.406250, time 217.386 samples/sec
+epoch 14, total_step 227640, total loss is 12.59 , inference loss is 6.22, weight deacy loss is 6.37, training accuracy is 0.343750, time 226.054 samples/sec
+epoch 14, total_step 227660, total loss is 14.54 , inference loss is 8.16, weight deacy loss is 6.37, training accuracy is 0.437500, time 219.743 samples/sec
+epoch 14, total_step 227680, total loss is 12.92 , inference loss is 6.54, weight deacy loss is 6.37, training accuracy is 0.531250, time 223.935 samples/sec
+epoch 14, total_step 227700, total loss is 18.42 , inference loss is 12.04, weight deacy loss is 6.37, training accuracy is 0.281250, time 12.629 samples/sec
+epoch 14, total_step 227720, total loss is 12.89 , inference loss is 6.51, weight deacy loss is 6.37, training accuracy is 0.437500, time 223.123 samples/sec
+epoch 14, total_step 227740, total loss is 14.05 , inference loss is 7.68, weight deacy loss is 6.37, training accuracy is 0.375000, time 222.637 samples/sec
+epoch 14, total_step 227760, total loss is 10.16 , inference loss is 3.78, weight deacy loss is 6.37, training accuracy is 0.468750, time 223.296 samples/sec
+epoch 14, total_step 227780, total loss is 14.58 , inference loss is 8.21, weight deacy loss is 6.37, training accuracy is 0.406250, time 282.682 samples/sec
+epoch 14, total_step 227800, total loss is 13.03 , inference loss is 6.66, weight deacy loss is 6.37, training accuracy is 0.437500, time 225.737 samples/sec
+epoch 14, total_step 227820, total loss is 16.41 , inference loss is 10.04, weight deacy loss is 6.37, training accuracy is 0.343750, time 225.838 samples/sec
+epoch 14, total_step 227840, total loss is 14.30 , inference loss is 7.92, weight deacy loss is 6.37, training accuracy is 0.437500, time 227.720 samples/sec
+epoch 14, total_step 227860, total loss is 15.86 , inference loss is 9.49, weight deacy loss is 6.37, training accuracy is 0.437500, time 220.180 samples/sec
+epoch 14, total_step 227880, total loss is 11.68 , inference loss is 5.30, weight deacy loss is 6.37, training accuracy is 0.468750, time 226.747 samples/sec
+epoch 14, total_step 227900, total loss is 17.16 , inference loss is 10.79, weight deacy loss is 6.37, training accuracy is 0.343750, time 224.973 samples/sec
+epoch 14, total_step 227920, total loss is 17.01 , inference loss is 10.64, weight deacy loss is 6.37, training accuracy is 0.312500, time 246.234 samples/sec
+epoch 14, total_step 227940, total loss is 17.96 , inference loss is 11.59, weight deacy loss is 6.37, training accuracy is 0.343750, time 228.995 samples/sec
+epoch 14, total_step 227960, total loss is 16.16 , inference loss is 9.79, weight deacy loss is 6.37, training accuracy is 0.375000, time 224.445 samples/sec
+epoch 14, total_step 227980, total loss is 16.14 , inference loss is 9.77, weight deacy loss is 6.37, training accuracy is 0.406250, time 222.773 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.74069200000001
+best_threshold_index 108 0.9862962962962963
+best_threshold_index 110 0.9855555555555555
+best_threshold_index 108 0.9848148148148148
+best_threshold_index 107 0.9862962962962963
+best_threshold_index 108 0.9857407407407407
+best_threshold_index 108 0.9848148148148148
+best_threshold_index 110 0.985
+best_threshold_index 110 0.9844444444444445
+best_threshold_index 110 0.9837037037037037
+best_threshold_index 110 0.9842592592592593
+[lfw][228000]XNorm: 22.208574
+[lfw][228000]Accuracy-Flip: 0.98383+-0.00792
+test accuracy is: 0.9838333333333334
+epoch 14, total_step 228000, total loss is 13.11 , inference loss is 6.73, weight deacy loss is 6.37, training accuracy is 0.500000, time 16.087 samples/sec
+epoch 14, total_step 228020, total loss is 10.75 , inference loss is 4.37, weight deacy loss is 6.37, training accuracy is 0.562500, time 212.435 samples/sec
+epoch 14, total_step 228040, total loss is 13.90 , inference loss is 7.53, weight deacy loss is 6.37, training accuracy is 0.406250, time 219.023 samples/sec
+epoch 14, total_step 228060, total loss is 12.06 , inference loss is 5.68, weight deacy loss is 6.37, training accuracy is 0.437500, time 234.263 samples/sec
+epoch 14, total_step 228080, total loss is 9.80 , inference loss is 3.43, weight deacy loss is 6.37, training accuracy is 0.562500, time 227.032 samples/sec
+epoch 14, total_step 228100, total loss is 10.81 , inference loss is 4.44, weight deacy loss is 6.37, training accuracy is 0.500000, time 297.663 samples/sec
+epoch 14, total_step 228120, total loss is 19.10 , inference loss is 12.72, weight deacy loss is 6.37, training accuracy is 0.343750, time 297.468 samples/sec
+epoch 14, total_step 228140, total loss is 17.40 , inference loss is 11.03, weight deacy loss is 6.37, training accuracy is 0.437500, time 213.108 samples/sec
+epoch 14, total_step 228160, total loss is 12.87 , inference loss is 6.50, weight deacy loss is 6.37, training accuracy is 0.500000, time 215.131 samples/sec
+epoch 14, total_step 228180, total loss is 12.55 , inference loss is 6.17, weight deacy loss is 6.37, training accuracy is 0.406250, time 225.803 samples/sec
+epoch 14, total_step 228200, total loss is 16.13 , inference loss is 9.75, weight deacy loss is 6.37, training accuracy is 0.250000, time 220.004 samples/sec
+epoch 14, total_step 228220, total loss is 13.69 , inference loss is 7.31, weight deacy loss is 6.37, training accuracy is 0.343750, time 220.681 samples/sec
+epoch 14, total_step 228240, total loss is 16.25 , inference loss is 9.88, weight deacy loss is 6.37, training accuracy is 0.375000, time 219.765 samples/sec
+epoch 14, total_step 228260, total loss is 13.45 , inference loss is 7.08, weight deacy loss is 6.37, training accuracy is 0.468750, time 205.487 samples/sec
+epoch 14, total_step 228280, total loss is 12.90 , inference loss is 6.52, weight deacy loss is 6.37, training accuracy is 0.531250, time 214.159 samples/sec
+epoch 14, total_step 228300, total loss is 13.29 , inference loss is 6.92, weight deacy loss is 6.37, training accuracy is 0.437500, time 12.766 samples/sec
+epoch 14, total_step 228320, total loss is 18.68 , inference loss is 12.31, weight deacy loss is 6.37, training accuracy is 0.281250, time 218.893 samples/sec
+epoch 14, total_step 228340, total loss is 15.14 , inference loss is 8.77, weight deacy loss is 6.37, training accuracy is 0.250000, time 216.924 samples/sec
+epoch 14, total_step 228360, total loss is 15.38 , inference loss is 9.01, weight deacy loss is 6.37, training accuracy is 0.218750, time 221.639 samples/sec
+epoch 14, total_step 228380, total loss is 10.40 , inference loss is 4.03, weight deacy loss is 6.37, training accuracy is 0.437500, time 228.954 samples/sec
+epoch 14, total_step 228400, total loss is 15.29 , inference loss is 8.92, weight deacy loss is 6.37, training accuracy is 0.312500, time 237.680 samples/sec
+epoch 14, total_step 228420, total loss is 11.28 , inference loss is 4.91, weight deacy loss is 6.37, training accuracy is 0.406250, time 222.992 samples/sec
+epoch 14, total_step 228440, total loss is 13.22 , inference loss is 6.85, weight deacy loss is 6.37, training accuracy is 0.312500, time 227.526 samples/sec
+epoch 14, total_step 228460, total loss is 15.79 , inference loss is 9.42, weight deacy loss is 6.37, training accuracy is 0.312500, time 210.193 samples/sec
+epoch 14, total_step 228480, total loss is 11.77 , inference loss is 5.40, weight deacy loss is 6.37, training accuracy is 0.531250, time 218.515 samples/sec
+epoch 14, total_step 228500, total loss is 14.74 , inference loss is 8.37, weight deacy loss is 6.37, training accuracy is 0.406250, time 206.532 samples/sec
+epoch 14, total_step 228520, total loss is 17.73 , inference loss is 11.36, weight deacy loss is 6.37, training accuracy is 0.250000, time 201.268 samples/sec
+epoch 14, total_step 228540, total loss is 13.64 , inference loss is 7.27, weight deacy loss is 6.37, training accuracy is 0.343750, time 227.889 samples/sec
+epoch 14, total_step 228560, total loss is 17.51 , inference loss is 11.13, weight deacy loss is 6.37, training accuracy is 0.312500, time 217.618 samples/sec
+epoch 14, total_step 228580, total loss is 16.42 , inference loss is 10.05, weight deacy loss is 6.37, training accuracy is 0.406250, time 218.904 samples/sec
+epoch 14, total_step 228600, total loss is 11.73 , inference loss is 5.36, weight deacy loss is 6.37, training accuracy is 0.531250, time 13.370 samples/sec
+epoch 14, total_step 228620, total loss is 14.34 , inference loss is 7.96, weight deacy loss is 6.37, training accuracy is 0.406250, time 220.510 samples/sec
+epoch 14, total_step 228640, total loss is 15.69 , inference loss is 9.32, weight deacy loss is 6.37, training accuracy is 0.281250, time 224.484 samples/sec
+epoch 14, total_step 228660, total loss is 11.56 , inference loss is 5.19, weight deacy loss is 6.37, training accuracy is 0.468750, time 217.663 samples/sec
+epoch 14, total_step 228680, total loss is 20.16 , inference loss is 13.79, weight deacy loss is 6.37, training accuracy is 0.406250, time 230.835 samples/sec
+epoch 14, total_step 228700, total loss is 15.44 , inference loss is 9.07, weight deacy loss is 6.37, training accuracy is 0.218750, time 225.580 samples/sec
+epoch 14, total_step 228720, total loss is 15.97 , inference loss is 9.60, weight deacy loss is 6.37, training accuracy is 0.281250, time 235.241 samples/sec
+epoch 14, total_step 228740, total loss is 13.49 , inference loss is 7.12, weight deacy loss is 6.37, training accuracy is 0.312500, time 222.955 samples/sec
+epoch 14, total_step 228760, total loss is 11.73 , inference loss is 5.36, weight deacy loss is 6.37, training accuracy is 0.406250, time 226.839 samples/sec
+epoch 14, total_step 228780, total loss is 14.75 , inference loss is 8.38, weight deacy loss is 6.37, training accuracy is 0.406250, time 225.543 samples/sec
+epoch 14, total_step 228800, total loss is 14.64 , inference loss is 8.27, weight deacy loss is 6.37, training accuracy is 0.375000, time 226.453 samples/sec
+epoch 14, total_step 228820, total loss is 11.72 , inference loss is 5.35, weight deacy loss is 6.37, training accuracy is 0.437500, time 223.946 samples/sec
+epoch 14, total_step 228840, total loss is 13.00 , inference loss is 6.63, weight deacy loss is 6.37, training accuracy is 0.343750, time 215.694 samples/sec
+epoch 14, total_step 228860, total loss is 19.27 , inference loss is 12.90, weight deacy loss is 6.37, training accuracy is 0.218750, time 295.502 samples/sec
+epoch 14, total_step 228880, total loss is 14.09 , inference loss is 7.72, weight deacy loss is 6.37, training accuracy is 0.468750, time 225.714 samples/sec
+epoch 14, total_step 228900, total loss is 10.97 , inference loss is 4.60, weight deacy loss is 6.37, training accuracy is 0.437500, time 12.047 samples/sec
+epoch 14, total_step 228920, total loss is 13.41 , inference loss is 7.04, weight deacy loss is 6.37, training accuracy is 0.406250, time 223.771 samples/sec
+epoch 14, total_step 228940, total loss is 14.70 , inference loss is 8.33, weight deacy loss is 6.37, training accuracy is 0.500000, time 226.406 samples/sec
+epoch 14, total_step 228960, total loss is 13.01 , inference loss is 6.64, weight deacy loss is 6.37, training accuracy is 0.312500, time 218.776 samples/sec
+epoch 14, total_step 228980, total loss is 18.23 , inference loss is 11.86, weight deacy loss is 6.37, training accuracy is 0.406250, time 233.994 samples/sec
+epoch 14, total_step 229000, total loss is 14.24 , inference loss is 7.87, weight deacy loss is 6.37, training accuracy is 0.500000, time 235.860 samples/sec
+epoch 14, total_step 229020, total loss is 14.37 , inference loss is 8.00, weight deacy loss is 6.37, training accuracy is 0.375000, time 216.794 samples/sec
+epoch 14, total_step 229040, total loss is 16.86 , inference loss is 10.49, weight deacy loss is 6.37, training accuracy is 0.375000, time 197.160 samples/sec
+epoch 14, total_step 229060, total loss is 12.91 , inference loss is 6.54, weight deacy loss is 6.37, training accuracy is 0.468750, time 226.875 samples/sec
+epoch 14, total_step 229080, total loss is 18.26 , inference loss is 11.89, weight deacy loss is 6.37, training accuracy is 0.312500, time 214.341 samples/sec
+epoch 14, total_step 229100, total loss is 15.60 , inference loss is 9.24, weight deacy loss is 6.37, training accuracy is 0.281250, time 181.601 samples/sec
+epoch 14, total_step 229120, total loss is 10.65 , inference loss is 4.28, weight deacy loss is 6.37, training accuracy is 0.625000, time 297.879 samples/sec
+epoch 14, total_step 229140, total loss is 18.20 , inference loss is 11.83, weight deacy loss is 6.37, training accuracy is 0.312500, time 219.219 samples/sec
+epoch 14, total_step 229160, total loss is 16.02 , inference loss is 9.65, weight deacy loss is 6.37, training accuracy is 0.312500, time 224.126 samples/sec
+epoch 14, total_step 229180, total loss is 14.34 , inference loss is 7.97, weight deacy loss is 6.37, training accuracy is 0.500000, time 226.868 samples/sec
+epoch 14, total_step 229200, total loss is 13.89 , inference loss is 7.52, weight deacy loss is 6.37, training accuracy is 0.500000, time 12.080 samples/sec
+epoch 14, total_step 229220, total loss is 14.44 , inference loss is 8.07, weight deacy loss is 6.37, training accuracy is 0.375000, time 223.356 samples/sec2022-12-01 15:34:55.858746: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7057 of 10000
+2022-12-01 15:35:00.004269: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 14, total_step 229240, total loss is 13.86 , inference loss is 7.50, weight deacy loss is 6.37, training accuracy is 0.437500, time 223.880 samples/sec
+epoch 14, total_step 229260, total loss is 16.76 , inference loss is 10.39, weight deacy loss is 6.37, training accuracy is 0.281250, time 232.216 samples/sec
+epoch 14, total_step 229280, total loss is 13.19 , inference loss is 6.82, weight deacy loss is 6.37, training accuracy is 0.312500, time 225.629 samples/sec
+epoch 14, total_step 229300, total loss is 13.73 , inference loss is 7.36, weight deacy loss is 6.37, training accuracy is 0.250000, time 204.136 samples/sec
+epoch 14, total_step 229320, total loss is 16.45 , inference loss is 10.08, weight deacy loss is 6.37, training accuracy is 0.437500, time 218.662 samples/sec
+epoch 14, total_step 229340, total loss is 14.95 , inference loss is 8.58, weight deacy loss is 6.37, training accuracy is 0.312500, time 222.799 samples/sec
+epoch 14, total_step 229360, total loss is 10.99 , inference loss is 4.63, weight deacy loss is 6.37, training accuracy is 0.468750, time 298.605 samples/sec
+epoch 14, total_step 229380, total loss is 12.07 , inference loss is 5.70, weight deacy loss is 6.37, training accuracy is 0.531250, time 210.588 samples/sec
+epoch 14, total_step 229400, total loss is 16.62 , inference loss is 10.25, weight deacy loss is 6.37, training accuracy is 0.343750, time 221.320 samples/sec
+epoch 14, total_step 229420, total loss is 16.26 , inference loss is 9.89, weight deacy loss is 6.37, training accuracy is 0.218750, time 220.397 samples/sec
+epoch 14, total_step 229440, total loss is 12.67 , inference loss is 6.30, weight deacy loss is 6.37, training accuracy is 0.312500, time 215.555 samples/sec
+epoch 14, total_step 229460, total loss is 13.94 , inference loss is 7.57, weight deacy loss is 6.37, training accuracy is 0.406250, time 218.282 samples/sec
+epoch 14, total_step 229480, total loss is 14.48 , inference loss is 8.11, weight deacy loss is 6.37, training accuracy is 0.375000, time 225.792 samples/sec
+epoch 14, total_step 229500, total loss is 13.79 , inference loss is 7.43, weight deacy loss is 6.37, training accuracy is 0.312500, time 13.155 samples/sec
+epoch 14, total_step 229520, total loss is 12.54 , inference loss is 6.17, weight deacy loss is 6.37, training accuracy is 0.437500, time 243.932 samples/sec
+epoch 14, total_step 229540, total loss is 16.11 , inference loss is 9.75, weight deacy loss is 6.37, training accuracy is 0.312500, time 228.530 samples/sec
+epoch 14, total_step 229560, total loss is 15.83 , inference loss is 9.47, weight deacy loss is 6.37, training accuracy is 0.375000, time 227.412 samples/sec
+epoch 14, total_step 229580, total loss is 14.80 , inference loss is 8.43, weight deacy loss is 6.37, training accuracy is 0.281250, time 228.103 samples/sec
+epoch 14, total_step 229600, total loss is 18.52 , inference loss is 12.15, weight deacy loss is 6.37, training accuracy is 0.343750, time 222.177 samples/sec
+epoch 14, total_step 229620, total loss is 14.13 , inference loss is 7.76, weight deacy loss is 6.37, training accuracy is 0.406250, time 298.281 samples/sec
+epoch 14, total_step 229640, total loss is 18.93 , inference loss is 12.56, weight deacy loss is 6.37, training accuracy is 0.218750, time 204.212 samples/sec
+epoch 14, total_step 229660, total loss is 12.08 , inference loss is 5.71, weight deacy loss is 6.37, training accuracy is 0.531250, time 199.651 samples/sec
+epoch 14, total_step 229680, total loss is 13.56 , inference loss is 7.19, weight deacy loss is 6.37, training accuracy is 0.468750, time 225.669 samples/sec
+epoch 14, total_step 229700, total loss is 13.52 , inference loss is 7.16, weight deacy loss is 6.37, training accuracy is 0.343750, time 199.297 samples/sec
+epoch 14, total_step 229720, total loss is 13.93 , inference loss is 7.56, weight deacy loss is 6.37, training accuracy is 0.500000, time 233.945 samples/sec
+epoch 14, total_step 229740, total loss is 12.67 , inference loss is 6.30, weight deacy loss is 6.37, training accuracy is 0.500000, time 214.169 samples/sec
+epoch 14, total_step 229760, total loss is 16.90 , inference loss is 10.53, weight deacy loss is 6.37, training accuracy is 0.312500, time 192.674 samples/sec
+epoch 14, total_step 229780, total loss is 13.38 , inference loss is 7.01, weight deacy loss is 6.37, training accuracy is 0.468750, time 212.636 samples/sec
+epoch 14, total_step 229800, total loss is 15.46 , inference loss is 9.10, weight deacy loss is 6.37, training accuracy is 0.406250, time 12.343 samples/sec
+epoch 14, total_step 229820, total loss is 14.25 , inference loss is 7.89, weight deacy loss is 6.37, training accuracy is 0.343750, time 206.216 samples/sec
+epoch 14, total_step 229840, total loss is 13.23 , inference loss is 6.87, weight deacy loss is 6.37, training accuracy is 0.312500, time 177.358 samples/sec
+epoch 14, total_step 229860, total loss is 13.53 , inference loss is 7.16, weight deacy loss is 6.37, training accuracy is 0.406250, time 207.732 samples/sec
+epoch 14, total_step 229880, total loss is 16.57 , inference loss is 10.21, weight deacy loss is 6.37, training accuracy is 0.468750, time 222.074 samples/sec
+epoch 14, total_step 229900, total loss is 15.61 , inference loss is 9.25, weight deacy loss is 6.37, training accuracy is 0.343750, time 198.496 samples/sec
+epoch 14, total_step 229920, total loss is 14.13 , inference loss is 7.76, weight deacy loss is 6.37, training accuracy is 0.406250, time 191.893 samples/sec
+epoch 14, total_step 229940, total loss is 15.72 , inference loss is 9.35, weight deacy loss is 6.37, training accuracy is 0.437500, time 208.579 samples/sec
+epoch 14, total_step 229960, total loss is 15.66 , inference loss is 9.30, weight deacy loss is 6.37, training accuracy is 0.406250, time 297.621 samples/sec
+End of epoch 14
+epoch 15, total_step 229980, total loss is 10.87 , inference loss is 4.50, weight deacy loss is 6.37, training accuracy is 0.531250, time 214.297 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.085755000000006
+best_threshold_index 96 0.9833333333333333
+best_threshold_index 96 0.9829629629629629
+best_threshold_index 96 0.9827777777777778
+best_threshold_index 96 0.9842592592592593
+best_threshold_index 96 0.9824074074074074
+best_threshold_index 96 0.9824074074074074
+best_threshold_index 96 0.9827777777777778
+best_threshold_index 96 0.9835185185185186
+best_threshold_index 96 0.982037037037037
+best_threshold_index 101 0.982037037037037
+[lfw][230000]XNorm: 22.228475
+[lfw][230000]Accuracy-Flip: 0.98233+-0.00564
+test accuracy is: 0.9823333333333333
+epoch 15, total_step 230000, total loss is 14.02 , inference loss is 7.65, weight deacy loss is 6.37, training accuracy is 0.687500, time 16.028 samples/sec
+epoch 15, total_step 230020, total loss is 9.34 , inference loss is 2.98, weight deacy loss is 6.37, training accuracy is 0.750000, time 221.082 samples/sec
+epoch 15, total_step 230040, total loss is 11.13 , inference loss is 4.77, weight deacy loss is 6.37, training accuracy is 0.812500, time 216.669 samples/sec
+epoch 15, total_step 230060, total loss is 9.37 , inference loss is 3.01, weight deacy loss is 6.37, training accuracy is 0.750000, time 219.809 samples/sec
+epoch 15, total_step 230080, total loss is 11.33 , inference loss is 4.97, weight deacy loss is 6.36, training accuracy is 0.687500, time 225.123 samples/sec
+epoch 15, total_step 230100, total loss is 9.22 , inference loss is 2.86, weight deacy loss is 6.36, training accuracy is 0.781250, time 13.020 samples/sec
+epoch 15, total_step 230120, total loss is 9.79 , inference loss is 3.43, weight deacy loss is 6.36, training accuracy is 0.812500, time 221.568 samples/sec
+epoch 15, total_step 230140, total loss is 13.12 , inference loss is 6.75, weight deacy loss is 6.36, training accuracy is 0.687500, time 224.660 samples/sec
+epoch 15, total_step 230160, total loss is 10.96 , inference loss is 4.59, weight deacy loss is 6.36, training accuracy is 0.593750, time 227.980 samples/sec
+epoch 15, total_step 230180, total loss is 8.96 , inference loss is 2.60, weight deacy loss is 6.36, training accuracy is 0.781250, time 207.005 samples/sec
+epoch 15, total_step 230200, total loss is 10.73 , inference loss is 4.36, weight deacy loss is 6.36, training accuracy is 0.718750, time 230.352 samples/sec
+epoch 15, total_step 230220, total loss is 10.94 , inference loss is 4.58, weight deacy loss is 6.36, training accuracy is 0.687500, time 220.649 samples/sec
+epoch 15, total_step 230240, total loss is 11.48 , inference loss is 5.11, weight deacy loss is 6.36, training accuracy is 0.750000, time 199.610 samples/sec
+epoch 15, total_step 230260, total loss is 10.07 , inference loss is 3.71, weight deacy loss is 6.36, training accuracy is 0.750000, time 224.449 samples/sec
+epoch 15, total_step 230280, total loss is 8.61 , inference loss is 2.25, weight deacy loss is 6.36, training accuracy is 0.781250, time 227.874 samples/sec
+epoch 15, total_step 230300, total loss is 10.25 , inference loss is 3.88, weight deacy loss is 6.36, training accuracy is 0.687500, time 226.289 samples/sec
+epoch 15, total_step 230320, total loss is 13.60 , inference loss is 7.24, weight deacy loss is 6.36, training accuracy is 0.593750, time 213.542 samples/sec
+epoch 15, total_step 230340, total loss is 9.43 , inference loss is 3.07, weight deacy loss is 6.36, training accuracy is 0.843750, time 210.268 samples/sec
+epoch 15, total_step 230360, total loss is 12.64 , inference loss is 6.27, weight deacy loss is 6.36, training accuracy is 0.625000, time 297.681 samples/sec
+epoch 15, total_step 230380, total loss is 9.02 , inference loss is 2.65, weight deacy loss is 6.36, training accuracy is 0.718750, time 224.347 samples/sec
+epoch 15, total_step 230400, total loss is 8.78 , inference loss is 2.41, weight deacy loss is 6.36, training accuracy is 0.781250, time 11.983 samples/sec
+epoch 15, total_step 230420, total loss is 11.00 , inference loss is 4.64, weight deacy loss is 6.36, training accuracy is 0.781250, time 218.899 samples/sec
+epoch 15, total_step 230440, total loss is 13.40 , inference loss is 7.04, weight deacy loss is 6.36, training accuracy is 0.656250, time 214.890 samples/sec
+epoch 15, total_step 230460, total loss is 9.46 , inference loss is 3.10, weight deacy loss is 6.36, training accuracy is 0.781250, time 220.903 samples/sec
+epoch 15, total_step 230480, total loss is 10.07 , inference loss is 3.70, weight deacy loss is 6.36, training accuracy is 0.718750, time 222.618 samples/sec
+epoch 15, total_step 230500, total loss is 8.36 , inference loss is 1.99, weight deacy loss is 6.36, training accuracy is 0.812500, time 224.327 samples/sec
+epoch 15, total_step 230520, total loss is 15.29 , inference loss is 8.93, weight deacy loss is 6.36, training accuracy is 0.562500, time 222.791 samples/sec
+epoch 15, total_step 230540, total loss is 13.11 , inference loss is 6.74, weight deacy loss is 6.36, training accuracy is 0.687500, time 230.329 samples/sec
+epoch 15, total_step 230560, total loss is 10.79 , inference loss is 4.42, weight deacy loss is 6.37, training accuracy is 0.625000, time 223.279 samples/sec
+epoch 15, total_step 230580, total loss is 13.19 , inference loss is 6.82, weight deacy loss is 6.37, training accuracy is 0.593750, time 225.601 samples/sec
+epoch 15, total_step 230600, total loss is 13.13 , inference loss is 6.76, weight deacy loss is 6.37, training accuracy is 0.562500, time 218.130 samples/sec
+epoch 15, total_step 230620, total loss is 11.09 , inference loss is 4.72, weight deacy loss is 6.37, training accuracy is 0.687500, time 298.819 samples/sec
+epoch 15, total_step 230640, total loss is 7.41 , inference loss is 1.04, weight deacy loss is 6.37, training accuracy is 0.843750, time 213.083 samples/sec
+epoch 15, total_step 230660, total loss is 11.96 , inference loss is 5.60, weight deacy loss is 6.37, training accuracy is 0.562500, time 224.464 samples/sec
+epoch 15, total_step 230680, total loss is 11.11 , inference loss is 4.74, weight deacy loss is 6.37, training accuracy is 0.687500, time 216.141 samples/sec
+epoch 15, total_step 230700, total loss is 13.66 , inference loss is 7.29, weight deacy loss is 6.37, training accuracy is 0.468750, time 12.296 samples/sec
+epoch 15, total_step 230720, total loss is 12.18 , inference loss is 5.81, weight deacy loss is 6.37, training accuracy is 0.562500, time 224.980 samples/sec
+epoch 15, total_step 230740, total loss is 10.83 , inference loss is 4.46, weight deacy loss is 6.37, training accuracy is 0.656250, time 221.265 samples/sec
+epoch 15, total_step 230760, total loss is 9.66 , inference loss is 3.29, weight deacy loss is 6.37, training accuracy is 0.687500, time 224.209 samples/sec
+epoch 15, total_step 230780, total loss is 16.16 , inference loss is 9.79, weight deacy loss is 6.37, training accuracy is 0.468750, time 220.669 samples/sec
+epoch 15, total_step 230800, total loss is 11.02 , inference loss is 4.65, weight deacy loss is 6.37, training accuracy is 0.625000, time 220.394 samples/sec
+epoch 15, total_step 230820, total loss is 16.25 , inference loss is 9.88, weight deacy loss is 6.37, training accuracy is 0.531250, time 222.293 samples/sec
+epoch 15, total_step 230840, total loss is 16.91 , inference loss is 10.54, weight deacy loss is 6.37, training accuracy is 0.531250, time 227.098 samples/sec
+epoch 15, total_step 230860, total loss is 15.43 , inference loss is 9.06, weight deacy loss is 6.37, training accuracy is 0.562500, time 224.898 samples/sec
+epoch 15, total_step 230880, total loss is 11.07 , inference loss is 4.70, weight deacy loss is 6.37, training accuracy is 0.750000, time 295.258 samples/sec
+epoch 15, total_step 230900, total loss is 12.93 , inference loss is 6.56, weight deacy loss is 6.37, training accuracy is 0.687500, time 222.068 samples/sec
+epoch 15, total_step 230920, total loss is 17.79 , inference loss is 11.42, weight deacy loss is 6.37, training accuracy is 0.562500, time 225.101 samples/sec
+epoch 15, total_step 230940, total loss is 11.18 , inference loss is 4.81, weight deacy loss is 6.37, training accuracy is 0.625000, time 218.402 samples/sec
+epoch 15, total_step 230960, total loss is 11.87 , inference loss is 5.49, weight deacy loss is 6.37, training accuracy is 0.625000, time 226.166 samples/sec
+epoch 15, total_step 230980, total loss is 9.07 , inference loss is 2.69, weight deacy loss is 6.37, training accuracy is 0.750000, time 228.258 samples/sec
+epoch 15, total_step 231000, total loss is 13.14 , inference loss is 6.77, weight deacy loss is 6.37, training accuracy is 0.625000, time 12.137 samples/sec
+epoch 15, total_step 231020, total loss is 10.22 , inference loss is 3.84, weight deacy loss is 6.37, training accuracy is 0.750000, time 215.859 samples/sec
+epoch 15, total_step 231040, total loss is 15.61 , inference loss is 9.23, weight deacy loss is 6.37, training accuracy is 0.500000, time 220.328 samples/sec
+epoch 15, total_step 231060, total loss is 14.29 , inference loss is 7.91, weight deacy loss is 6.37, training accuracy is 0.531250, time 218.681 samples/sec
+epoch 15, total_step 231080, total loss is 10.48 , inference loss is 4.11, weight deacy loss is 6.37, training accuracy is 0.687500, time 226.347 samples/sec
+epoch 15, total_step 231100, total loss is 11.41 , inference loss is 5.03, weight deacy loss is 6.37, training accuracy is 0.625000, time 224.484 samples/sec
+epoch 15, total_step 231120, total loss is 10.62 , inference loss is 4.24, weight deacy loss is 6.37, training accuracy is 0.625000, time 216.729 samples/sec
+epoch 15, total_step 231140, total loss is 11.72 , inference loss is 5.35, weight deacy loss is 6.37, training accuracy is 0.656250, time 297.103 samples/sec
+epoch 15, total_step 231160, total loss is 12.82 , inference loss is 6.44, weight deacy loss is 6.37, training accuracy is 0.625000, time 217.684 samples/sec
+epoch 15, total_step 231180, total loss is 12.85 , inference loss is 6.47, weight deacy loss is 6.38, training accuracy is 0.500000, time 217.256 samples/sec
+epoch 15, total_step 231200, total loss is 13.34 , inference loss is 6.97, weight deacy loss is 6.38, training accuracy is 0.500000, time 222.415 samples/sec
+epoch 15, total_step 231220, total loss is 9.15 , inference loss is 2.78, weight deacy loss is 6.38, training accuracy is 0.687500, time 220.126 samples/sec
+epoch 15, total_step 231240, total loss is 13.86 , inference loss is 7.48, weight deacy loss is 6.38, training accuracy is 0.593750, time 223.214 samples/sec
+epoch 15, total_step 231260, total loss is 12.68 , inference loss is 6.30, weight deacy loss is 6.38, training accuracy is 0.531250, time 216.091 samples/sec
+epoch 15, total_step 231280, total loss is 11.13 , inference loss is 4.75, weight deacy loss is 6.38, training accuracy is 0.531250, time 219.849 samples/sec
+epoch 15, total_step 231300, total loss is 14.44 , inference loss is 8.06, weight deacy loss is 6.38, training accuracy is 0.531250, time 11.927 samples/sec
+epoch 15, total_step 231320, total loss is 13.56 , inference loss is 7.18, weight deacy loss is 6.38, training accuracy is 0.406250, time 218.641 samples/sec
+epoch 15, total_step 231340, total loss is 14.01 , inference loss is 7.63, weight deacy loss is 6.38, training accuracy is 0.562500, time 226.964 samples/sec
+epoch 15, total_step 231360, total loss is 16.33 , inference loss is 9.96, weight deacy loss is 6.38, training accuracy is 0.531250, time 227.498 samples/sec
+epoch 15, total_step 231380, total loss is 10.51 , inference loss is 4.14, weight deacy loss is 6.38, training accuracy is 0.625000, time 229.532 samples/sec
+epoch 15, total_step 231400, total loss is 12.55 , inference loss is 6.17, weight deacy loss is 6.38, training accuracy is 0.687500, time 295.400 samples/sec
+epoch 15, total_step 231420, total loss is 13.97 , inference loss is 7.59, weight deacy loss is 6.38, training accuracy is 0.468750, time 230.900 samples/sec
+epoch 15, total_step 231440, total loss is 16.69 , inference loss is 10.31, weight deacy loss is 6.38, training accuracy is 0.593750, time 222.594 samples/sec
+epoch 15, total_step 231460, total loss is 12.14 , inference loss is 5.76, weight deacy loss is 6.38, training accuracy is 0.562500, time 218.043 samples/sec
+epoch 15, total_step 231480, total loss is 14.83 , inference loss is 8.45, weight deacy loss is 6.38, training accuracy is 0.500000, time 229.038 samples/sec
+epoch 15, total_step 231500, total loss is 9.72 , inference loss is 3.34, weight deacy loss is 6.38, training accuracy is 0.625000, time 224.548 samples/sec
+epoch 15, total_step 231520, total loss is 14.35 , inference loss is 7.97, weight deacy loss is 6.38, training accuracy is 0.437500, time 219.216 samples/sec
+epoch 15, total_step 231540, total loss is 11.51 , inference loss is 5.13, weight deacy loss is 6.38, training accuracy is 0.718750, time 225.413 samples/sec
+epoch 15, total_step 231560, total loss is 10.84 , inference loss is 4.46, weight deacy loss is 6.38, training accuracy is 0.687500, time 231.801 samples/sec
+epoch 15, total_step 231580, total loss is 9.64 , inference loss is 3.26, weight deacy loss is 6.38, training accuracy is 0.687500, time 211.582 samples/sec
+epoch 15, total_step 231600, total loss is 12.03 , inference loss is 5.65, weight deacy loss is 6.38, training accuracy is 0.625000, time 12.465 samples/sec
+epoch 15, total_step 231620, total loss is 16.52 , inference loss is 10.14, weight deacy loss is 6.38, training accuracy is 0.531250, time 216.597 samples/sec
+epoch 15, total_step 231640, total loss is 12.22 , inference loss is 5.84, weight deacy loss is 6.38, training accuracy is 0.531250, time 214.749 samples/sec
+epoch 15, total_step 231660, total loss is 12.48 , inference loss is 6.10, weight deacy loss is 6.38, training accuracy is 0.562500, time 218.152 samples/sec
+epoch 15, total_step 231680, total loss is 13.34 , inference loss is 6.96, weight deacy loss is 6.38, training accuracy is 0.500000, time 227.940 samples/sec
+epoch 15, total_step 231700, total loss is 14.76 , inference loss is 8.38, weight deacy loss is 6.38, training accuracy is 0.500000, time 299.404 samples/sec
+epoch 15, total_step 231720, total loss is 11.63 , inference loss is 5.24, weight deacy loss is 6.38, training accuracy is 0.625000, time 298.441 samples/sec
+epoch 15, total_step 231740, total loss is 14.12 , inference loss is 7.73, weight deacy loss is 6.38, training accuracy is 0.625000, time 220.958 samples/sec
+epoch 15, total_step 231760, total loss is 11.43 , inference loss is 5.04, weight deacy loss is 6.38, training accuracy is 0.562500, time 219.033 samples/sec
+epoch 15, total_step 231780, total loss is 9.26 , inference loss is 2.88, weight deacy loss is 6.38, training accuracy is 0.718750, time 219.878 samples/sec
+epoch 15, total_step 231800, total loss is 12.72 , inference loss is 6.33, weight deacy loss is 6.38, training accuracy is 0.500000, time 217.266 samples/sec
+epoch 15, total_step 231820, total loss is 13.40 , inference loss is 7.01, weight deacy loss is 6.38, training accuracy is 0.562500, time 221.165 samples/sec
+epoch 15, total_step 231840, total loss is 15.84 , inference loss is 9.45, weight deacy loss is 6.38, training accuracy is 0.375000, time 296.360 samples/sec
+epoch 15, total_step 231860, total loss is 13.08 , inference loss is 6.69, weight deacy loss is 6.38, training accuracy is 0.500000, time 238.300 samples/sec
+epoch 15, total_step 231880, total loss is 11.13 , inference loss is 4.75, weight deacy loss is 6.38, training accuracy is 0.562500, time 214.686 samples/sec
+epoch 15, total_step 231900, total loss is 11.19 , inference loss is 4.81, weight deacy loss is 6.38, training accuracy is 0.625000, time 12.152 samples/sec
+epoch 15, total_step 231920, total loss is 13.97 , inference loss is 7.59, weight deacy loss is 6.38, training accuracy is 0.468750, time 226.655 samples/sec
+epoch 15, total_step 231940, total loss is 9.98 , inference loss is 3.59, weight deacy loss is 6.38, training accuracy is 0.625000, time 220.672 samples/sec
+epoch 15, total_step 231960, total loss is 12.80 , inference loss is 6.42, weight deacy loss is 6.38, training accuracy is 0.531250, time 222.073 samples/sec
+epoch 15, total_step 231980, total loss is 11.91 , inference loss is 5.53, weight deacy loss is 6.38, training accuracy is 0.562500, time 196.528 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.690452000000025
+best_threshold_index 121 0.9855555555555555
+best_threshold_index 119 0.9848148148148148
+best_threshold_index 121 0.9838888888888889
+best_threshold_index 118 0.9853703703703703
+best_threshold_index 121 0.9851851851851852
+best_threshold_index 125 0.9844444444444445
+best_threshold_index 119 0.9846296296296296
+best_threshold_index 119 0.985
+best_threshold_index 120 0.9835185185185186
+best_threshold_index 116 0.9837037037037037
+[lfw][232000]XNorm: 21.724717
+[lfw][232000]Accuracy-Flip: 0.98267+-0.00588
+test accuracy is: 0.9826666666666666
+epoch 15, total_step 232000, total loss is 9.58 , inference loss is 3.20, weight deacy loss is 6.38, training accuracy is 0.718750, time 13.012 samples/sec
+epoch 15, total_step 232020, total loss is 10.47 , inference loss is 4.09, weight deacy loss is 6.38, training accuracy is 0.625000, time 215.885 samples/sec
+epoch 15, total_step 232040, total loss is 10.73 , inference loss is 4.35, weight deacy loss is 6.38, training accuracy is 0.656250, time 217.554 samples/sec
+epoch 15, total_step 232060, total loss is 9.05 , inference loss is 2.67, weight deacy loss is 6.38, training accuracy is 0.718750, time 224.981 samples/sec
+epoch 15, total_step 232080, total loss is 14.75 , inference loss is 8.37, weight deacy loss is 6.38, training accuracy is 0.500000, time 227.062 samples/sec
+epoch 15, total_step 232100, total loss is 10.91 , inference loss is 4.53, weight deacy loss is 6.38, training accuracy is 0.468750, time 295.403 samples/sec
+epoch 15, total_step 232120, total loss is 8.73 , inference loss is 2.35, weight deacy loss is 6.38, training accuracy is 0.750000, time 216.605 samples/sec
+epoch 15, total_step 232140, total loss is 12.05 , inference loss is 5.66, weight deacy loss is 6.38, training accuracy is 0.656250, time 215.136 samples/sec
+epoch 15, total_step 232160, total loss is 12.62 , inference loss is 6.23, weight deacy loss is 6.38, training accuracy is 0.656250, time 227.379 samples/sec
+epoch 15, total_step 232180, total loss is 12.05 , inference loss is 5.67, weight deacy loss is 6.38, training accuracy is 0.593750, time 210.337 samples/sec
+epoch 15, total_step 232200, total loss is 10.30 , inference loss is 3.92, weight deacy loss is 6.38, training accuracy is 0.625000, time 9.240 samples/sec
+epoch 15, total_step 232220, total loss is 10.27 , inference loss is 3.89, weight deacy loss is 6.38, training accuracy is 0.718750, time 210.472 samples/sec
+epoch 15, total_step 232240, total loss is 13.21 , inference loss is 6.82, weight deacy loss is 6.38, training accuracy is 0.468750, time 216.409 samples/sec
+epoch 15, total_step 232260, total loss is 11.37 , inference loss is 4.99, weight deacy loss is 6.38, training accuracy is 0.531250, time 212.390 samples/sec
+epoch 15, total_step 232280, total loss is 9.45 , inference loss is 3.07, weight deacy loss is 6.38, training accuracy is 0.781250, time 210.837 samples/sec
+epoch 15, total_step 232300, total loss is 10.51 , inference loss is 4.12, weight deacy loss is 6.38, training accuracy is 0.718750, time 212.662 samples/sec
+epoch 15, total_step 232320, total loss is 8.14 , inference loss is 1.76, weight deacy loss is 6.38, training accuracy is 0.812500, time 222.016 samples/sec
+epoch 15, total_step 232340, total loss is 14.41 , inference loss is 8.03, weight deacy loss is 6.38, training accuracy is 0.656250, time 299.242 samples/sec
+epoch 15, total_step 232360, total loss is 10.28 , inference loss is 3.89, weight deacy loss is 6.38, training accuracy is 0.656250, time 227.189 samples/sec
+epoch 15, total_step 232380, total loss is 9.10 , inference loss is 2.72, weight deacy loss is 6.38, training accuracy is 0.750000, time 209.989 samples/sec
+epoch 15, total_step 232400, total loss is 10.38 , inference loss is 4.00, weight deacy loss is 6.38, training accuracy is 0.718750, time 220.644 samples/sec
+epoch 15, total_step 232420, total loss is 10.78 , inference loss is 4.39, weight deacy loss is 6.38, training accuracy is 0.562500, time 220.246 samples/sec
+epoch 15, total_step 232440, total loss is 10.31 , inference loss is 3.92, weight deacy loss is 6.38, training accuracy is 0.625000, time 226.469 samples/sec
+epoch 15, total_step 232460, total loss is 10.30 , inference loss is 3.92, weight deacy loss is 6.38, training accuracy is 0.562500, time 238.127 samples/sec
+epoch 15, total_step 232480, total loss is 9.15 , inference loss is 2.77, weight deacy loss is 6.38, training accuracy is 0.750000, time 215.120 samples/sec
+epoch 15, total_step 232500, total loss is 11.25 , inference loss is 4.87, weight deacy loss is 6.38, training accuracy is 0.625000, time 8.536 samples/sec
+epoch 15, total_step 232520, total loss is 14.54 , inference loss is 8.16, weight deacy loss is 6.38, training accuracy is 0.562500, time 216.359 samples/sec
+epoch 15, total_step 232540, total loss is 8.31 , inference loss is 1.93, weight deacy loss is 6.38, training accuracy is 0.812500, time 221.024 samples/sec
+epoch 15, total_step 232560, total loss is 10.22 , inference loss is 3.84, weight deacy loss is 6.38, training accuracy is 0.656250, time 206.874 samples/sec
+epoch 15, total_step 232580, total loss is 9.32 , inference loss is 2.93, weight deacy loss is 6.38, training accuracy is 0.812500, time 297.808 samples/sec
+epoch 15, total_step 232600, total loss is 9.69 , inference loss is 3.31, weight deacy loss is 6.38, training accuracy is 0.718750, time 295.705 samples/sec
+epoch 15, total_step 232620, total loss is 12.37 , inference loss is 5.99, weight deacy loss is 6.38, training accuracy is 0.656250, time 219.759 samples/sec
+epoch 15, total_step 232640, total loss is 13.48 , inference loss is 7.09, weight deacy loss is 6.38, training accuracy is 0.687500, time 206.544 samples/sec
+epoch 15, total_step 232660, total loss is 10.90 , inference loss is 4.52, weight deacy loss is 6.38, training accuracy is 0.656250, time 214.551 samples/sec
+epoch 15, total_step 232680, total loss is 10.97 , inference loss is 4.58, weight deacy loss is 6.38, training accuracy is 0.562500, time 211.432 samples/sec
+epoch 15, total_step 232700, total loss is 11.61 , inference loss is 5.23, weight deacy loss is 6.38, training accuracy is 0.437500, time 224.988 samples/sec
+epoch 15, total_step 232720, total loss is 13.71 , inference loss is 7.33, weight deacy loss is 6.38, training accuracy is 0.687500, time 210.520 samples/sec
+epoch 15, total_step 232740, total loss is 10.26 , inference loss is 3.88, weight deacy loss is 6.38, training accuracy is 0.656250, time 222.591 samples/sec
+epoch 15, total_step 232760, total loss is 10.98 , inference loss is 4.60, weight deacy loss is 6.38, training accuracy is 0.625000, time 223.946 samples/sec
+epoch 15, total_step 232780, total loss is 14.38 , inference loss is 8.00, weight deacy loss is 6.38, training accuracy is 0.625000, time 229.155 samples/sec
+epoch 15, total_step 232800, total loss is 12.01 , inference loss is 5.63, weight deacy loss is 6.38, training accuracy is 0.500000, time 6.510 samples/sec
+epoch 15, total_step 232820, total loss is 13.80 , inference loss is 7.42, weight deacy loss is 6.38, training accuracy is 0.531250, time 220.394 samples/sec
+epoch 15, total_step 232840, total loss is 9.57 , inference loss is 3.19, weight deacy loss is 6.38, training accuracy is 0.593750, time 299.216 samples/sec
+epoch 15, total_step 232860, total loss is 11.09 , inference loss is 4.70, weight deacy loss is 6.38, training accuracy is 0.656250, time 298.867 samples/sec
+epoch 15, total_step 232880, total loss is 12.07 , inference loss is 5.68, weight deacy loss is 6.38, training accuracy is 0.625000, time 228.083 samples/sec
+epoch 15, total_step 232900, total loss is 9.20 , inference loss is 2.82, weight deacy loss is 6.38, training accuracy is 0.687500, time 224.870 samples/sec
+epoch 15, total_step 232920, total loss is 10.27 , inference loss is 3.89, weight deacy loss is 6.38, training accuracy is 0.625000, time 224.364 samples/sec
+epoch 15, total_step 232940, total loss is 10.89 , inference loss is 4.51, weight deacy loss is 6.38, training accuracy is 0.750000, time 215.754 samples/sec
+epoch 15, total_step 232960, total loss is 12.41 , inference loss is 6.03, weight deacy loss is 6.38, training accuracy is 0.625000, time 212.428 samples/sec
+epoch 15, total_step 232980, total loss is 13.48 , inference loss is 7.10, weight deacy loss is 6.38, training accuracy is 0.593750, time 220.212 samples/sec
+epoch 15, total_step 233000, total loss is 12.11 , inference loss is 5.73, weight deacy loss is 6.38, training accuracy is 0.468750, time 214.089 samples/sec
+epoch 15, total_step 233020, total loss is 10.43 , inference loss is 4.05, weight deacy loss is 6.38, training accuracy is 0.656250, time 222.590 samples/sec
+epoch 15, total_step 233040, total loss is 10.27 , inference loss is 3.88, weight deacy loss is 6.38, training accuracy is 0.750000, time 227.470 samples/sec
+epoch 15, total_step 233060, total loss is 12.44 , inference loss is 6.06, weight deacy loss is 6.38, training accuracy is 0.593750, time 222.157 samples/sec
+epoch 15, total_step 233080, total loss is 12.63 , inference loss is 6.25, weight deacy loss is 6.38, training accuracy is 0.656250, time 228.104 samples/sec
+epoch 15, total_step 233100, total loss is 12.75 , inference loss is 6.37, weight deacy loss is 6.38, training accuracy is 0.625000, time 12.660 samples/sec
+epoch 15, total_step 233120, total loss is 10.39 , inference loss is 4.00, weight deacy loss is 6.38, training accuracy is 0.781250, time 295.357 samples/sec
+epoch 15, total_step 233140, total loss is 10.74 , inference loss is 4.36, weight deacy loss is 6.38, training accuracy is 0.625000, time 220.427 samples/sec
+epoch 15, total_step 233160, total loss is 13.13 , inference loss is 6.74, weight deacy loss is 6.38, training accuracy is 0.406250, time 217.443 samples/sec
+epoch 15, total_step 233180, total loss is 15.84 , inference loss is 9.45, weight deacy loss is 6.38, training accuracy is 0.625000, time 224.505 samples/sec
+epoch 15, total_step 233200, total loss is 11.17 , inference loss is 4.79, weight deacy loss is 6.38, training accuracy is 0.593750, time 223.750 samples/sec
+epoch 15, total_step 233220, total loss is 10.96 , inference loss is 4.58, weight deacy loss is 6.38, training accuracy is 0.687500, time 218.847 samples/sec
+epoch 15, total_step 233240, total loss is 12.96 , inference loss is 6.57, weight deacy loss is 6.38, training accuracy is 0.437500, time 214.953 samples/sec
+epoch 15, total_step 233260, total loss is 16.09 , inference loss is 9.71, weight deacy loss is 6.38, training accuracy is 0.531250, time 212.297 samples/sec
+epoch 15, total_step 233280, total loss is 8.29 , inference loss is 1.91, weight deacy loss is 6.38, training accuracy is 0.687500, time 219.769 samples/sec
+epoch 15, total_step 233300, total loss is 15.00 , inference loss is 8.62, weight deacy loss is 6.38, training accuracy is 0.437500, time 215.344 samples/sec
+epoch 15, total_step 233320, total loss is 11.04 , inference loss is 4.65, weight deacy loss is 6.38, training accuracy is 0.625000, time 211.718 samples/sec
+epoch 15, total_step 233340, total loss is 17.21 , inference loss is 10.82, weight deacy loss is 6.38, training accuracy is 0.343750, time 216.086 samples/sec
+epoch 15, total_step 233360, total loss is 13.44 , inference loss is 7.06, weight deacy loss is 6.38, training accuracy is 0.531250, time 217.171 samples/sec
+epoch 15, total_step 233380, total loss is 11.79 , inference loss is 5.40, weight deacy loss is 6.38, training accuracy is 0.625000, time 296.827 samples/sec
+epoch 15, total_step 233400, total loss is 12.27 , inference loss is 5.89, weight deacy loss is 6.38, training accuracy is 0.500000, time 12.348 samples/sec
+epoch 15, total_step 233420, total loss is 10.57 , inference loss is 4.18, weight deacy loss is 6.38, training accuracy is 0.593750, time 214.712 samples/sec
+epoch 15, total_step 233440, total loss is 12.14 , inference loss is 5.76, weight deacy loss is 6.38, training accuracy is 0.562500, time 234.041 samples/sec
+epoch 15, total_step 233460, total loss is 12.39 , inference loss is 6.00, weight deacy loss is 6.38, training accuracy is 0.500000, time 219.199 samples/sec
+epoch 15, total_step 233480, total loss is 11.69 , inference loss is 5.30, weight deacy loss is 6.38, training accuracy is 0.625000, time 243.077 samples/sec
+epoch 15, total_step 233500, total loss is 11.09 , inference loss is 4.71, weight deacy loss is 6.38, training accuracy is 0.656250, time 227.957 samples/sec
+epoch 15, total_step 233520, total loss is 14.01 , inference loss is 7.62, weight deacy loss is 6.38, training accuracy is 0.531250, time 215.321 samples/sec
+epoch 15, total_step 233540, total loss is 13.71 , inference loss is 7.33, weight deacy loss is 6.38, training accuracy is 0.562500, time 219.685 samples/sec
+epoch 15, total_step 233560, total loss is 12.21 , inference loss is 5.83, weight deacy loss is 6.38, training accuracy is 0.562500, time 299.973 samples/sec
+epoch 15, total_step 233580, total loss is 13.85 , inference loss is 7.47, weight deacy loss is 6.38, training accuracy is 0.406250, time 230.723 samples/sec
+epoch 15, total_step 233600, total loss is 14.18 , inference loss is 7.80, weight deacy loss is 6.38, training accuracy is 0.562500, time 231.730 samples/sec
+epoch 15, total_step 233620, total loss is 9.04 , inference loss is 2.66, weight deacy loss is 6.38, training accuracy is 0.687500, time 218.964 samples/sec
+epoch 15, total_step 233640, total loss is 10.78 , inference loss is 4.40, weight deacy loss is 6.38, training accuracy is 0.406250, time 211.543 samples/sec
+epoch 15, total_step 233660, total loss is 15.05 , inference loss is 8.67, weight deacy loss is 6.38, training accuracy is 0.562500, time 227.715 samples/sec
+epoch 15, total_step 233680, total loss is 14.27 , inference loss is 7.88, weight deacy loss is 6.38, training accuracy is 0.531250, time 224.710 samples/sec
+epoch 15, total_step 233700, total loss is 12.54 , inference loss is 6.16, weight deacy loss is 6.38, training accuracy is 0.531250, time 13.722 samples/sec
+epoch 15, total_step 233720, total loss is 10.43 , inference loss is 4.05, weight deacy loss is 6.38, training accuracy is 0.593750, time 230.642 samples/sec
+epoch 15, total_step 233740, total loss is 11.21 , inference loss is 4.83, weight deacy loss is 6.38, training accuracy is 0.500000, time 223.794 samples/sec
+epoch 15, total_step 233760, total loss is 13.52 , inference loss is 7.14, weight deacy loss is 6.38, training accuracy is 0.468750, time 227.029 samples/sec
+epoch 15, total_step 233780, total loss is 15.68 , inference loss is 9.30, weight deacy loss is 6.38, training accuracy is 0.562500, time 242.717 samples/sec
+epoch 15, total_step 233800, total loss is 12.30 , inference loss is 5.92, weight deacy loss is 6.38, training accuracy is 0.593750, time 219.071 samples/sec
+epoch 15, total_step 233820, total loss is 16.39 , inference loss is 10.01, weight deacy loss is 6.38, training accuracy is 0.437500, time 299.120 samples/sec
+epoch 15, total_step 233840, total loss is 15.33 , inference loss is 8.94, weight deacy loss is 6.38, training accuracy is 0.531250, time 225.778 samples/sec
+epoch 15, total_step 233860, total loss is 13.90 , inference loss is 7.52, weight deacy loss is 6.38, training accuracy is 0.531250, time 233.155 samples/sec
+epoch 15, total_step 233880, total loss is 12.16 , inference loss is 5.78, weight deacy loss is 6.38, training accuracy is 0.468750, time 226.941 samples/sec
+epoch 15, total_step 233900, total loss is 15.53 , inference loss is 9.15, weight deacy loss is 6.38, training accuracy is 0.500000, time 223.245 samples/sec
+epoch 15, total_step 233920, total loss is 11.90 , inference loss is 5.52, weight deacy loss is 6.38, training accuracy is 0.593750, time 236.303 samples/sec
+epoch 15, total_step 233940, total loss is 13.13 , inference loss is 6.75, weight deacy loss is 6.38, training accuracy is 0.437500, time 222.689 samples/sec
+epoch 15, total_step 233960, total loss is 15.32 , inference loss is 8.94, weight deacy loss is 6.38, training accuracy is 0.593750, time 219.097 samples/sec
+epoch 15, total_step 233980, total loss is 11.68 , inference loss is 5.30, weight deacy loss is 6.38, training accuracy is 0.593750, time 221.347 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.275073000000035
+best_threshold_index 122 0.9890740740740741
+best_threshold_index 122 0.9879629629629629
+best_threshold_index 122 0.9888888888888889
+best_threshold_index 119 0.9890740740740741
+best_threshold_index 122 0.9883333333333333
+best_threshold_index 122 0.9881481481481481
+best_threshold_index 119 0.9875925925925926
+best_threshold_index 122 0.9879629629629629
+best_threshold_index 122 0.9868518518518519
+best_threshold_index 122 0.9877777777777778
+[lfw][234000]XNorm: 19.795196
+[lfw][234000]Accuracy-Flip: 0.98750+-0.00634
+test accuracy is: 0.9875
+epoch 15, total_step 234000, total loss is 9.93 , inference loss is 3.54, weight deacy loss is 6.38, training accuracy is 0.656250, time 16.428 samples/sec
+epoch 15, total_step 234020, total loss is 11.36 , inference loss is 4.98, weight deacy loss is 6.38, training accuracy is 0.500000, time 219.906 samples/sec
+epoch 15, total_step 234040, total loss is 17.38 , inference loss is 11.00, weight deacy loss is 6.38, training accuracy is 0.406250, time 217.659 samples/sec
+epoch 15, total_step 234060, total loss is 11.13 , inference loss is 4.74, weight deacy loss is 6.38, training accuracy is 0.562500, time 234.232 samples/sec
+epoch 15, total_step 234080, total loss is 11.55 , inference loss is 5.16, weight deacy loss is 6.38, training accuracy is 0.562500, time 223.785 samples/sec
+epoch 15, total_step 234100, total loss is 12.97 , inference loss is 6.59, weight deacy loss is 6.38, training accuracy is 0.500000, time 224.075 samples/sec
+epoch 15, total_step 234120, total loss is 12.68 , inference loss is 6.29, weight deacy loss is 6.38, training accuracy is 0.531250, time 224.727 samples/sec
+epoch 15, total_step 234140, total loss is 11.53 , inference loss is 5.15, weight deacy loss is 6.38, training accuracy is 0.625000, time 224.530 samples/sec
+epoch 15, total_step 234160, total loss is 11.90 , inference loss is 5.51, weight deacy loss is 6.38, training accuracy is 0.500000, time 220.729 samples/sec
+epoch 15, total_step 234180, total loss is 11.66 , inference loss is 5.28, weight deacy loss is 6.38, training accuracy is 0.562500, time 217.062 samples/sec
+epoch 15, total_step 234200, total loss is 11.42 , inference loss is 5.04, weight deacy loss is 6.38, training accuracy is 0.656250, time 299.829 samples/sec
+epoch 15, total_step 234220, total loss is 17.25 , inference loss is 10.86, weight deacy loss is 6.38, training accuracy is 0.468750, time 213.411 samples/sec
+epoch 15, total_step 234240, total loss is 13.95 , inference loss is 7.57, weight deacy loss is 6.38, training accuracy is 0.500000, time 225.490 samples/sec
+epoch 15, total_step 234260, total loss is 13.71 , inference loss is 7.32, weight deacy loss is 6.38, training accuracy is 0.531250, time 218.842 samples/sec
+epoch 15, total_step 234280, total loss is 13.23 , inference loss is 6.85, weight deacy loss is 6.38, training accuracy is 0.593750, time 223.053 samples/sec
+epoch 15, total_step 234300, total loss is 10.64 , inference loss is 4.25, weight deacy loss is 6.38, training accuracy is 0.625000, time 12.133 samples/sec
+epoch 15, total_step 234320, total loss is 12.82 , inference loss is 6.43, weight deacy loss is 6.38, training accuracy is 0.562500, time 232.595 samples/sec
+epoch 15, total_step 234340, total loss is 12.61 , inference loss is 6.23, weight deacy loss is 6.38, training accuracy is 0.500000, time 211.224 samples/sec
+epoch 15, total_step 234360, total loss is 14.09 , inference loss is 7.70, weight deacy loss is 6.38, training accuracy is 0.375000, time 225.807 samples/sec
+epoch 15, total_step 234380, total loss is 12.72 , inference loss is 6.33, weight deacy loss is 6.38, training accuracy is 0.593750, time 226.075 samples/sec
+epoch 15, total_step 234400, total loss is 12.67 , inference loss is 6.28, weight deacy loss is 6.38, training accuracy is 0.593750, time 225.897 samples/sec
+epoch 15, total_step 234420, total loss is 15.45 , inference loss is 9.07, weight deacy loss is 6.38, training accuracy is 0.312500, time 206.411 samples/sec
+epoch 15, total_step 234440, total loss is 12.76 , inference loss is 6.38, weight deacy loss is 6.38, training accuracy is 0.562500, time 212.719 samples/sec
+epoch 15, total_step 234460, total loss is 11.25 , inference loss is 4.87, weight deacy loss is 6.38, training accuracy is 0.531250, time 297.833 samples/sec
+epoch 15, total_step 234480, total loss is 12.09 , inference loss is 5.71, weight deacy loss is 6.38, training accuracy is 0.406250, time 228.284 samples/sec
+epoch 15, total_step 234500, total loss is 12.36 , inference loss is 5.97, weight deacy loss is 6.38, training accuracy is 0.593750, time 222.676 samples/sec
+epoch 15, total_step 234520, total loss is 12.86 , inference loss is 6.47, weight deacy loss is 6.38, training accuracy is 0.593750, time 240.050 samples/sec
+epoch 15, total_step 234540, total loss is 15.14 , inference loss is 8.75, weight deacy loss is 6.38, training accuracy is 0.468750, time 243.750 samples/sec
+epoch 15, total_step 234560, total loss is 13.41 , inference loss is 7.02, weight deacy loss is 6.38, training accuracy is 0.312500, time 212.325 samples/sec
+epoch 15, total_step 234580, total loss is 9.92 , inference loss is 3.53, weight deacy loss is 6.38, training accuracy is 0.687500, time 246.018 samples/sec
+epoch 15, total_step 234600, total loss is 11.23 , inference loss is 4.85, weight deacy loss is 6.38, training accuracy is 0.468750, time 12.630 samples/sec
+epoch 15, total_step 234620, total loss is 14.56 , inference loss is 8.17, weight deacy loss is 6.38, training accuracy is 0.375000, time 224.221 samples/sec
+epoch 15, total_step 234640, total loss is 13.72 , inference loss is 7.33, weight deacy loss is 6.39, training accuracy is 0.375000, time 213.917 samples/sec
+epoch 15, total_step 234660, total loss is 13.12 , inference loss is 6.74, weight deacy loss is 6.39, training accuracy is 0.500000, time 213.059 samples/sec
+epoch 15, total_step 234680, total loss is 11.21 , inference loss is 4.82, weight deacy loss is 6.39, training accuracy is 0.625000, time 219.411 samples/sec
+epoch 15, total_step 234700, total loss is 12.98 , inference loss is 6.59, weight deacy loss is 6.39, training accuracy is 0.500000, time 214.715 samples/sec
+epoch 15, total_step 234720, total loss is 14.30 , inference loss is 7.92, weight deacy loss is 6.39, training accuracy is 0.500000, time 299.995 samples/sec
+epoch 15, total_step 234740, total loss is 13.47 , inference loss is 7.08, weight deacy loss is 6.39, training accuracy is 0.437500, time 217.860 samples/sec
+epoch 15, total_step 234760, total loss is 14.31 , inference loss is 7.92, weight deacy loss is 6.39, training accuracy is 0.500000, time 219.270 samples/sec
+epoch 15, total_step 234780, total loss is 12.69 , inference loss is 6.30, weight deacy loss is 6.39, training accuracy is 0.437500, time 215.745 samples/sec
+epoch 15, total_step 234800, total loss is 12.47 , inference loss is 6.08, weight deacy loss is 6.39, training accuracy is 0.531250, time 214.670 samples/sec
+epoch 15, total_step 234820, total loss is 13.00 , inference loss is 6.62, weight deacy loss is 6.39, training accuracy is 0.531250, time 215.760 samples/sec
+epoch 15, total_step 234840, total loss is 10.88 , inference loss is 4.50, weight deacy loss is 6.39, training accuracy is 0.531250, time 217.372 samples/sec
+epoch 15, total_step 234860, total loss is 14.37 , inference loss is 7.99, weight deacy loss is 6.39, training accuracy is 0.562500, time 222.302 samples/sec
+epoch 15, total_step 234880, total loss is 11.20 , inference loss is 4.81, weight deacy loss is 6.39, training accuracy is 0.593750, time 229.450 samples/sec
+epoch 15, total_step 234900, total loss is 18.53 , inference loss is 12.14, weight deacy loss is 6.39, training accuracy is 0.437500, time 12.610 samples/sec
+epoch 15, total_step 234920, total loss is 20.53 , inference loss is 14.14, weight deacy loss is 6.39, training accuracy is 0.312500, time 221.850 samples/sec
+epoch 15, total_step 234940, total loss is 16.68 , inference loss is 10.29, weight deacy loss is 6.39, training accuracy is 0.406250, time 215.693 samples/sec
+epoch 15, total_step 234960, total loss is 8.93 , inference loss is 2.54, weight deacy loss is 6.39, training accuracy is 0.593750, time 219.842 samples/sec
+epoch 15, total_step 234980, total loss is 13.25 , inference loss is 6.87, weight deacy loss is 6.39, training accuracy is 0.562500, time 297.115 samples/sec
+epoch 15, total_step 235000, total loss is 9.39 , inference loss is 3.00, weight deacy loss is 6.39, training accuracy is 0.656250, time 300.867 samples/sec
+epoch 15, total_step 235020, total loss is 11.27 , inference loss is 4.89, weight deacy loss is 6.39, training accuracy is 0.531250, time 215.353 samples/sec
+epoch 15, total_step 235040, total loss is 14.58 , inference loss is 8.19, weight deacy loss is 6.39, training accuracy is 0.531250, time 214.296 samples/sec
+epoch 15, total_step 235060, total loss is 11.14 , inference loss is 4.75, weight deacy loss is 6.39, training accuracy is 0.625000, time 218.285 samples/sec
+epoch 15, total_step 235080, total loss is 12.03 , inference loss is 5.65, weight deacy loss is 6.39, training accuracy is 0.718750, time 219.978 samples/sec
+epoch 15, total_step 235100, total loss is 8.76 , inference loss is 2.37, weight deacy loss is 6.39, training accuracy is 0.718750, time 218.666 samples/sec
+epoch 15, total_step 235120, total loss is 12.66 , inference loss is 6.28, weight deacy loss is 6.39, training accuracy is 0.593750, time 220.308 samples/sec
+epoch 15, total_step 235140, total loss is 15.86 , inference loss is 9.47, weight deacy loss is 6.39, training accuracy is 0.437500, time 226.231 samples/sec
+epoch 15, total_step 235160, total loss is 12.31 , inference loss is 5.93, weight deacy loss is 6.39, training accuracy is 0.562500, time 297.472 samples/sec
+epoch 15, total_step 235180, total loss is 8.40 , inference loss is 2.01, weight deacy loss is 6.39, training accuracy is 0.718750, time 298.252 samples/sec
+epoch 15, total_step 235200, total loss is 11.24 , inference loss is 4.86, weight deacy loss is 6.39, training accuracy is 0.562500, time 12.626 samples/sec
+epoch 15, total_step 235220, total loss is 13.83 , inference loss is 7.44, weight deacy loss is 6.39, training accuracy is 0.562500, time 224.574 samples/sec
+epoch 15, total_step 235240, total loss is 13.21 , inference loss is 6.83, weight deacy loss is 6.38, training accuracy is 0.593750, time 223.350 samples/sec
+epoch 15, total_step 235260, total loss is 14.24 , inference loss is 7.86, weight deacy loss is 6.38, training accuracy is 0.500000, time 226.641 samples/sec
+epoch 15, total_step 235280, total loss is 17.01 , inference loss is 10.63, weight deacy loss is 6.38, training accuracy is 0.468750, time 230.871 samples/sec
+epoch 15, total_step 235300, total loss is 16.62 , inference loss is 10.23, weight deacy loss is 6.38, training accuracy is 0.406250, time 215.648 samples/sec
+epoch 15, total_step 235320, total loss is 13.64 , inference loss is 7.26, weight deacy loss is 6.38, training accuracy is 0.593750, time 217.892 samples/sec
+epoch 15, total_step 235340, total loss is 11.92 , inference loss is 5.53, weight deacy loss is 6.39, training accuracy is 0.437500, time 219.689 samples/sec
+epoch 15, total_step 235360, total loss is 13.08 , inference loss is 6.69, weight deacy loss is 6.38, training accuracy is 0.593750, time 224.171 samples/sec
+epoch 15, total_step 235380, total loss is 11.37 , inference loss is 4.98, weight deacy loss is 6.38, training accuracy is 0.531250, time 225.234 samples/sec
+epoch 15, total_step 235400, total loss is 9.88 , inference loss is 3.50, weight deacy loss is 6.38, training accuracy is 0.656250, time 225.932 samples/sec
+epoch 15, total_step 235420, total loss is 16.83 , inference loss is 10.45, weight deacy loss is 6.38, training accuracy is 0.343750, time 220.781 samples/sec
+epoch 15, total_step 235440, total loss is 13.97 , inference loss is 7.59, weight deacy loss is 6.38, training accuracy is 0.500000, time 229.431 samples/sec
+epoch 15, total_step 235460, total loss is 13.38 , inference loss is 6.99, weight deacy loss is 6.38, training accuracy is 0.468750, time 297.225 samples/sec
+epoch 15, total_step 235480, total loss is 11.04 , inference loss is 4.65, weight deacy loss is 6.38, training accuracy is 0.625000, time 217.016 samples/sec
+epoch 15, total_step 235500, total loss is 12.06 , inference loss is 5.67, weight deacy loss is 6.38, training accuracy is 0.500000, time 12.398 samples/sec
+epoch 15, total_step 235520, total loss is 11.56 , inference loss is 5.18, weight deacy loss is 6.38, training accuracy is 0.468750, time 219.609 samples/sec
+epoch 15, total_step 235540, total loss is 11.44 , inference loss is 5.06, weight deacy loss is 6.38, training accuracy is 0.656250, time 213.187 samples/sec
+epoch 15, total_step 235560, total loss is 10.50 , inference loss is 4.12, weight deacy loss is 6.38, training accuracy is 0.562500, time 225.675 samples/sec
+epoch 15, total_step 235580, total loss is 12.15 , inference loss is 5.77, weight deacy loss is 6.38, training accuracy is 0.437500, time 237.477 samples/sec
+epoch 15, total_step 235600, total loss is 10.63 , inference loss is 4.24, weight deacy loss is 6.38, training accuracy is 0.593750, time 225.530 samples/sec
+epoch 15, total_step 235620, total loss is 16.98 , inference loss is 10.59, weight deacy loss is 6.38, training accuracy is 0.437500, time 225.225 samples/sec
+epoch 15, total_step 235640, total loss is 12.69 , inference loss is 6.30, weight deacy loss is 6.38, training accuracy is 0.625000, time 225.513 samples/sec
+epoch 15, total_step 235660, total loss is 12.98 , inference loss is 6.59, weight deacy loss is 6.38, training accuracy is 0.500000, time 219.218 samples/sec
+epoch 15, total_step 235680, total loss is 16.12 , inference loss is 9.74, weight deacy loss is 6.38, training accuracy is 0.531250, time 223.232 samples/sec
+epoch 15, total_step 235700, total loss is 14.06 , inference loss is 7.68, weight deacy loss is 6.38, training accuracy is 0.406250, time 223.196 samples/sec
+epoch 15, total_step 235720, total loss is 13.54 , inference loss is 7.16, weight deacy loss is 6.38, training accuracy is 0.468750, time 296.160 samples/sec
+epoch 15, total_step 235740, total loss is 14.51 , inference loss is 8.13, weight deacy loss is 6.38, training accuracy is 0.531250, time 234.183 samples/sec
+epoch 15, total_step 235760, total loss is 13.37 , inference loss is 6.99, weight deacy loss is 6.38, training accuracy is 0.406250, time 176.203 samples/sec
+epoch 15, total_step 235780, total loss is 10.52 , inference loss is 4.14, weight deacy loss is 6.38, training accuracy is 0.500000, time 243.154 samples/sec
+epoch 15, total_step 235800, total loss is 13.19 , inference loss is 6.81, weight deacy loss is 6.38, training accuracy is 0.625000, time 11.339 samples/sec
+epoch 15, total_step 235820, total loss is 10.93 , inference loss is 4.55, weight deacy loss is 6.38, training accuracy is 0.531250, time 213.416 samples/sec
+epoch 15, total_step 235840, total loss is 12.61 , inference loss is 6.23, weight deacy loss is 6.38, training accuracy is 0.468750, time 220.530 samples/sec
+epoch 15, total_step 235860, total loss is 12.34 , inference loss is 5.96, weight deacy loss is 6.38, training accuracy is 0.562500, time 219.416 samples/sec
+epoch 15, total_step 235880, total loss is 14.77 , inference loss is 8.38, weight deacy loss is 6.38, training accuracy is 0.500000, time 225.483 samples/sec
+epoch 15, total_step 235900, total loss is 10.31 , inference loss is 3.93, weight deacy loss is 6.38, training accuracy is 0.625000, time 225.184 samples/sec
+epoch 15, total_step 235920, total loss is 10.46 , inference loss is 4.08, weight deacy loss is 6.38, training accuracy is 0.500000, time 221.602 samples/sec
+epoch 15, total_step 235940, total loss is 11.20 , inference loss is 4.82, weight deacy loss is 6.38, training accuracy is 0.468750, time 217.719 samples/sec
+epoch 15, total_step 235960, total loss is 14.75 , inference loss is 8.37, weight deacy loss is 6.38, training accuracy is 0.312500, time 225.783 samples/sec
+epoch 15, total_step 235980, total loss is 9.89 , inference loss is 3.50, weight deacy loss is 6.38, training accuracy is 0.625000, time 299.068 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.494266999999994
+best_threshold_index 118 0.9892592592592593
+best_threshold_index 112 0.9885185185185185
+best_threshold_index 112 0.9888888888888889
+best_threshold_index 112 0.9892592592592593
+best_threshold_index 122 0.9888888888888889
+best_threshold_index 119 0.9875925925925926
+best_threshold_index 112 0.9888888888888889
+best_threshold_index 112 0.9881481481481481
+best_threshold_index 119 0.9874074074074074
+best_threshold_index 112 0.9877777777777778
+[lfw][236000]XNorm: 21.149444
+[lfw][236000]Accuracy-Flip: 0.98683+-0.00621
+test accuracy is: 0.9868333333333335
+epoch 15, total_step 236000, total loss is 15.47 , inference loss is 9.09, weight deacy loss is 6.38, training accuracy is 0.406250, time 17.231 samples/sec
+epoch 15, total_step 236020, total loss is 14.03 , inference loss is 7.65, weight deacy loss is 6.38, training accuracy is 0.406250, time 219.137 samples/sec
+epoch 15, total_step 236040, total loss is 13.31 , inference loss is 6.93, weight deacy loss is 6.38, training accuracy is 0.531250, time 298.300 samples/sec
+epoch 15, total_step 236060, total loss is 12.36 , inference loss is 5.98, weight deacy loss is 6.38, training accuracy is 0.562500, time 221.436 samples/sec
+epoch 15, total_step 236080, total loss is 12.09 , inference loss is 5.71, weight deacy loss is 6.38, training accuracy is 0.531250, time 214.785 samples/sec
+epoch 15, total_step 236100, total loss is 10.78 , inference loss is 4.40, weight deacy loss is 6.38, training accuracy is 0.593750, time 12.462 samples/sec
+epoch 15, total_step 236120, total loss is 13.23 , inference loss is 6.85, weight deacy loss is 6.38, training accuracy is 0.656250, time 224.253 samples/sec
+epoch 15, total_step 236140, total loss is 15.42 , inference loss is 9.04, weight deacy loss is 6.38, training accuracy is 0.625000, time 217.594 samples/sec
+epoch 15, total_step 236160, total loss is 14.38 , inference loss is 8.00, weight deacy loss is 6.38, training accuracy is 0.562500, time 225.256 samples/sec
+epoch 15, total_step 236180, total loss is 11.57 , inference loss is 5.18, weight deacy loss is 6.38, training accuracy is 0.625000, time 225.908 samples/sec
+epoch 15, total_step 236200, total loss is 11.36 , inference loss is 4.97, weight deacy loss is 6.38, training accuracy is 0.531250, time 220.399 samples/sec
+epoch 15, total_step 236220, total loss is 13.39 , inference loss is 7.00, weight deacy loss is 6.38, training accuracy is 0.562500, time 224.948 samples/sec
+epoch 15, total_step 236240, total loss is 14.45 , inference loss is 8.07, weight deacy loss is 6.38, training accuracy is 0.531250, time 223.857 samples/sec
+epoch 15, total_step 236260, total loss is 12.01 , inference loss is 5.63, weight deacy loss is 6.38, training accuracy is 0.500000, time 235.243 samples/sec
+epoch 15, total_step 236280, total loss is 10.75 , inference loss is 4.37, weight deacy loss is 6.38, training accuracy is 0.562500, time 222.385 samples/sec
+epoch 15, total_step 236300, total loss is 13.38 , inference loss is 7.00, weight deacy loss is 6.38, training accuracy is 0.500000, time 299.348 samples/sec
+epoch 15, total_step 236320, total loss is 11.62 , inference loss is 5.24, weight deacy loss is 6.38, training accuracy is 0.562500, time 222.910 samples/sec
+epoch 15, total_step 236340, total loss is 12.94 , inference loss is 6.55, weight deacy loss is 6.38, training accuracy is 0.593750, time 222.110 samples/sec
+epoch 15, total_step 236360, total loss is 12.57 , inference loss is 6.19, weight deacy loss is 6.38, training accuracy is 0.531250, time 227.529 samples/sec
+epoch 15, total_step 236380, total loss is 10.73 , inference loss is 4.35, weight deacy loss is 6.38, training accuracy is 0.500000, time 222.500 samples/sec
+epoch 15, total_step 236400, total loss is 12.13 , inference loss is 5.75, weight deacy loss is 6.38, training accuracy is 0.625000, time 12.065 samples/sec
+epoch 15, total_step 236420, total loss is 12.01 , inference loss is 5.63, weight deacy loss is 6.38, training accuracy is 0.593750, time 231.428 samples/sec
+epoch 15, total_step 236440, total loss is 12.39 , inference loss is 6.01, weight deacy loss is 6.38, training accuracy is 0.562500, time 227.463 samples/sec
+epoch 15, total_step 236460, total loss is 10.87 , inference loss is 4.49, weight deacy loss is 6.38, training accuracy is 0.593750, time 219.411 samples/sec
+epoch 15, total_step 236480, total loss is 15.85 , inference loss is 9.47, weight deacy loss is 6.38, training accuracy is 0.437500, time 226.916 samples/sec
+epoch 15, total_step 236500, total loss is 11.47 , inference loss is 5.08, weight deacy loss is 6.38, training accuracy is 0.343750, time 214.681 samples/sec
+epoch 15, total_step 236520, total loss is 13.19 , inference loss is 6.80, weight deacy loss is 6.38, training accuracy is 0.375000, time 222.978 samples/sec
+epoch 15, total_step 236540, total loss is 12.31 , inference loss is 5.93, weight deacy loss is 6.38, training accuracy is 0.437500, time 224.639 samples/sec
+epoch 15, total_step 236560, total loss is 15.54 , inference loss is 9.15, weight deacy loss is 6.38, training accuracy is 0.437500, time 185.002 samples/sec
+epoch 15, total_step 236580, total loss is 12.03 , inference loss is 5.64, weight deacy loss is 6.38, training accuracy is 0.437500, time 218.042 samples/sec
+epoch 15, total_step 236600, total loss is 10.55 , inference loss is 4.17, weight deacy loss is 6.38, training accuracy is 0.531250, time 225.837 samples/sec
+epoch 15, total_step 236620, total loss is 14.91 , inference loss is 8.53, weight deacy loss is 6.38, training accuracy is 0.468750, time 226.556 samples/sec
+epoch 15, total_step 236640, total loss is 11.34 , inference loss is 4.96, weight deacy loss is 6.38, training accuracy is 0.531250, time 224.291 samples/sec
+epoch 15, total_step 236660, total loss is 15.89 , inference loss is 9.51, weight deacy loss is 6.38, training accuracy is 0.406250, time 224.170 samples/sec
+epoch 15, total_step 236680, total loss is 9.58 , inference loss is 3.19, weight deacy loss is 6.38, training accuracy is 0.656250, time 224.283 samples/sec
+epoch 15, total_step 236700, total loss is 17.39 , inference loss is 11.01, weight deacy loss is 6.38, training accuracy is 0.625000, time 9.033 samples/sec
+epoch 15, total_step 236720, total loss is 15.36 , inference loss is 8.98, weight deacy loss is 6.38, training accuracy is 0.437500, time 216.446 samples/sec
+epoch 15, total_step 236740, total loss is 17.40 , inference loss is 11.02, weight deacy loss is 6.38, training accuracy is 0.500000, time 225.199 samples/sec
+epoch 15, total_step 236760, total loss is 11.71 , inference loss is 5.33, weight deacy loss is 6.38, training accuracy is 0.500000, time 254.451 samples/sec
+epoch 15, total_step 236780, total loss is 13.91 , inference loss is 7.53, weight deacy loss is 6.38, training accuracy is 0.593750, time 240.416 samples/sec
+epoch 15, total_step 236800, total loss is 11.53 , inference loss is 5.15, weight deacy loss is 6.38, training accuracy is 0.500000, time 205.075 samples/sec
+epoch 15, total_step 236820, total loss is 12.18 , inference loss is 5.80, weight deacy loss is 6.38, training accuracy is 0.562500, time 212.548 samples/sec
+epoch 15, total_step 236840, total loss is 12.80 , inference loss is 6.42, weight deacy loss is 6.38, training accuracy is 0.625000, time 301.113 samples/sec
+epoch 15, total_step 236860, total loss is 13.39 , inference loss is 7.01, weight deacy loss is 6.38, training accuracy is 0.500000, time 200.631 samples/sec
+epoch 15, total_step 236880, total loss is 10.54 , inference loss is 4.16, weight deacy loss is 6.38, training accuracy is 0.656250, time 204.980 samples/sec
+epoch 15, total_step 236900, total loss is 20.23 , inference loss is 13.85, weight deacy loss is 6.38, training accuracy is 0.281250, time 215.346 samples/sec
+epoch 15, total_step 236920, total loss is 18.85 , inference loss is 12.47, weight deacy loss is 6.38, training accuracy is 0.375000, time 195.886 samples/sec
+epoch 15, total_step 236940, total loss is 10.45 , inference loss is 4.07, weight deacy loss is 6.38, training accuracy is 0.500000, time 199.681 samples/sec
+epoch 15, total_step 236960, total loss is 11.10 , inference loss is 4.72, weight deacy loss is 6.38, training accuracy is 0.562500, time 299.864 samples/sec
+epoch 15, total_step 236980, total loss is 13.38 , inference loss is 7.00, weight deacy loss is 6.38, training accuracy is 0.593750, time 300.082 samples/sec
+epoch 15, total_step 237000, total loss is 16.43 , inference loss is 10.05, weight deacy loss is 6.38, training accuracy is 0.406250, time 12.063 samples/sec
+epoch 15, total_step 237020, total loss is 15.48 , inference loss is 9.10, weight deacy loss is 6.38, training accuracy is 0.343750, time 207.877 samples/sec
+epoch 15, total_step 237040, total loss is 13.43 , inference loss is 7.05, weight deacy loss is 6.38, training accuracy is 0.406250, time 223.764 samples/sec
+epoch 15, total_step 237060, total loss is 14.04 , inference loss is 7.66, weight deacy loss is 6.38, training accuracy is 0.531250, time 234.256 samples/sec
+epoch 15, total_step 237080, total loss is 13.99 , inference loss is 7.61, weight deacy loss is 6.38, training accuracy is 0.437500, time 227.358 samples/sec
+epoch 15, total_step 237100, total loss is 13.39 , inference loss is 7.01, weight deacy loss is 6.38, training accuracy is 0.531250, time 227.188 samples/sec
+epoch 15, total_step 237120, total loss is 14.87 , inference loss is 8.49, weight deacy loss is 6.38, training accuracy is 0.625000, time 214.319 samples/sec
+epoch 15, total_step 237140, total loss is 16.10 , inference loss is 9.71, weight deacy loss is 6.38, training accuracy is 0.437500, time 199.624 samples/sec
+epoch 15, total_step 237160, total loss is 17.65 , inference loss is 11.27, weight deacy loss is 6.38, training accuracy is 0.468750, time 202.972 samples/sec
+epoch 15, total_step 237180, total loss is 12.14 , inference loss is 5.76, weight deacy loss is 6.38, training accuracy is 0.656250, time 207.454 samples/sec
+epoch 15, total_step 237200, total loss is 8.86 , inference loss is 2.48, weight deacy loss is 6.38, training accuracy is 0.718750, time 298.496 samples/sec
+epoch 15, total_step 237220, total loss is 12.10 , inference loss is 5.72, weight deacy loss is 6.38, training accuracy is 0.500000, time 297.735 samples/sec
+epoch 15, total_step 237240, total loss is 11.97 , inference loss is 5.59, weight deacy loss is 6.38, training accuracy is 0.406250, time 205.760 samples/sec
+epoch 15, total_step 237260, total loss is 11.57 , inference loss is 5.19, weight deacy loss is 6.38, training accuracy is 0.718750, time 203.836 samples/sec
+epoch 15, total_step 237280, total loss is 16.77 , inference loss is 10.39, weight deacy loss is 6.38, training accuracy is 0.531250, time 210.606 samples/sec
+epoch 15, total_step 237300, total loss is 13.06 , inference loss is 6.68, weight deacy loss is 6.38, training accuracy is 0.625000, time 8.259 samples/sec
+epoch 15, total_step 237320, total loss is 13.23 , inference loss is 6.85, weight deacy loss is 6.38, training accuracy is 0.562500, time 208.532 samples/sec
+epoch 15, total_step 237340, total loss is 12.93 , inference loss is 6.55, weight deacy loss is 6.38, training accuracy is 0.500000, time 215.483 samples/sec
+epoch 15, total_step 237360, total loss is 12.25 , inference loss is 5.87, weight deacy loss is 6.38, training accuracy is 0.656250, time 211.411 samples/sec
+epoch 15, total_step 237380, total loss is 14.52 , inference loss is 8.14, weight deacy loss is 6.38, training accuracy is 0.437500, time 217.153 samples/sec
+epoch 15, total_step 237400, total loss is 12.16 , inference loss is 5.78, weight deacy loss is 6.38, training accuracy is 0.562500, time 197.491 samples/sec
+epoch 15, total_step 237420, total loss is 14.30 , inference loss is 7.92, weight deacy loss is 6.38, training accuracy is 0.468750, time 178.975 samples/sec
+epoch 15, total_step 237440, total loss is 10.19 , inference loss is 3.80, weight deacy loss is 6.38, training accuracy is 0.656250, time 299.365 samples/sec
+epoch 15, total_step 237460, total loss is 14.14 , inference loss is 7.76, weight deacy loss is 6.38, training accuracy is 0.437500, time 200.288 samples/sec
+epoch 15, total_step 237480, total loss is 13.14 , inference loss is 6.76, weight deacy loss is 6.38, training accuracy is 0.531250, time 198.767 samples/sec
+epoch 15, total_step 237500, total loss is 13.47 , inference loss is 7.09, weight deacy loss is 6.38, training accuracy is 0.500000, time 205.598 samples/sec
+epoch 15, total_step 237520, total loss is 12.99 , inference loss is 6.61, weight deacy loss is 6.38, training accuracy is 0.531250, time 201.207 samples/sec
+epoch 15, total_step 237540, total loss is 9.87 , inference loss is 3.49, weight deacy loss is 6.38, training accuracy is 0.468750, time 211.201 samples/sec
+epoch 15, total_step 237560, total loss is 12.12 , inference loss is 5.74, weight deacy loss is 6.38, training accuracy is 0.593750, time 223.909 samples/sec
+epoch 15, total_step 237580, total loss is 13.66 , inference loss is 7.28, weight deacy loss is 6.38, training accuracy is 0.531250, time 222.087 samples/sec
+epoch 15, total_step 237600, total loss is 9.03 , inference loss is 2.65, weight deacy loss is 6.38, training accuracy is 0.500000, time 7.539 samples/sec
+epoch 15, total_step 237620, total loss is 11.39 , inference loss is 5.01, weight deacy loss is 6.38, training accuracy is 0.593750, time 205.775 samples/sec
+epoch 15, total_step 237640, total loss is 9.00 , inference loss is 2.62, weight deacy loss is 6.38, training accuracy is 0.750000, time 203.309 samples/sec
+epoch 15, total_step 237660, total loss is 15.69 , inference loss is 9.31, weight deacy loss is 6.38, training accuracy is 0.468750, time 297.061 samples/sec
+epoch 15, total_step 237680, total loss is 12.61 , inference loss is 6.23, weight deacy loss is 6.38, training accuracy is 0.656250, time 213.402 samples/sec
+epoch 15, total_step 237700, total loss is 12.82 , inference loss is 6.44, weight deacy loss is 6.38, training accuracy is 0.531250, time 213.843 samples/sec
+epoch 15, total_step 237720, total loss is 12.73 , inference loss is 6.35, weight deacy loss is 6.38, training accuracy is 0.593750, time 213.328 samples/sec
+epoch 15, total_step 237740, total loss is 13.09 , inference loss is 6.71, weight deacy loss is 6.38, training accuracy is 0.625000, time 227.569 samples/sec
+epoch 15, total_step 237760, total loss is 13.34 , inference loss is 6.96, weight deacy loss is 6.38, training accuracy is 0.468750, time 221.322 samples/sec
+epoch 15, total_step 237780, total loss is 10.60 , inference loss is 4.22, weight deacy loss is 6.38, training accuracy is 0.562500, time 218.351 samples/sec
+epoch 15, total_step 237800, total loss is 11.04 , inference loss is 4.66, weight deacy loss is 6.38, training accuracy is 0.531250, time 204.236 samples/sec
+epoch 15, total_step 237820, total loss is 12.77 , inference loss is 6.39, weight deacy loss is 6.38, training accuracy is 0.500000, time 201.674 samples/sec
+epoch 15, total_step 237840, total loss is 15.29 , inference loss is 8.91, weight deacy loss is 6.38, training accuracy is 0.343750, time 204.583 samples/sec
+epoch 15, total_step 237860, total loss is 14.50 , inference loss is 8.12, weight deacy loss is 6.38, training accuracy is 0.468750, time 212.341 samples/sec
+epoch 15, total_step 237880, total loss is 12.84 , inference loss is 6.46, weight deacy loss is 6.38, training accuracy is 0.593750, time 212.896 samples/sec
+epoch 15, total_step 237900, total loss is 9.96 , inference loss is 3.58, weight deacy loss is 6.38, training accuracy is 0.625000, time 12.005 samples/sec
+epoch 15, total_step 237920, total loss is 10.65 , inference loss is 4.27, weight deacy loss is 6.38, training accuracy is 0.625000, time 179.494 samples/sec
+epoch 15, total_step 237940, total loss is 10.16 , inference loss is 3.78, weight deacy loss is 6.38, training accuracy is 0.593750, time 179.471 samples/sec
+epoch 15, total_step 237960, total loss is 14.14 , inference loss is 7.77, weight deacy loss is 6.38, training accuracy is 0.593750, time 224.161 samples/sec
+epoch 15, total_step 237980, total loss is 13.48 , inference loss is 7.10, weight deacy loss is 6.38, training accuracy is 0.562500, time 210.945 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.269894000000022
+best_threshold_index 118 0.9894444444444445
+best_threshold_index 118 0.9887037037037038
+best_threshold_index 119 0.9887037037037038
+best_threshold_index 118 0.9901851851851852
+best_threshold_index 118 0.9887037037037038
+best_threshold_index 119 0.9879629629629629
+best_threshold_index 119 0.9896296296296296
+best_threshold_index 118 0.9879629629629629
+best_threshold_index 118 0.9879629629629629
+best_threshold_index 118 0.9881481481481481
+[lfw][238000]XNorm: 21.477620
+[lfw][238000]Accuracy-Flip: 0.98800+-0.00670
+test accuracy is: 0.9880000000000001
+epoch 15, total_step 238000, total loss is 11.25 , inference loss is 4.87, weight deacy loss is 6.38, training accuracy is 0.656250, time 16.138 samples/sec
+epoch 15, total_step 238020, total loss is 12.28 , inference loss is 5.90, weight deacy loss is 6.38, training accuracy is 0.562500, time 298.968 samples/sec
+epoch 15, total_step 238040, total loss is 11.20 , inference loss is 4.82, weight deacy loss is 6.38, training accuracy is 0.625000, time 224.339 samples/sec
+epoch 15, total_step 238060, total loss is 9.33 , inference loss is 2.95, weight deacy loss is 6.38, training accuracy is 0.531250, time 208.997 samples/sec
+epoch 15, total_step 238080, total loss is 15.69 , inference loss is 9.31, weight deacy loss is 6.38, training accuracy is 0.343750, time 213.521 samples/sec
+epoch 15, total_step 238100, total loss is 11.05 , inference loss is 4.67, weight deacy loss is 6.38, training accuracy is 0.468750, time 214.376 samples/sec
+epoch 15, total_step 238120, total loss is 15.78 , inference loss is 9.40, weight deacy loss is 6.38, training accuracy is 0.593750, time 227.925 samples/sec
+epoch 15, total_step 238140, total loss is 14.40 , inference loss is 8.02, weight deacy loss is 6.38, training accuracy is 0.500000, time 223.983 samples/sec
+epoch 15, total_step 238160, total loss is 11.74 , inference loss is 5.37, weight deacy loss is 6.38, training accuracy is 0.531250, time 213.005 samples/sec
+epoch 15, total_step 238180, total loss is 16.85 , inference loss is 10.47, weight deacy loss is 6.38, training accuracy is 0.375000, time 215.987 samples/sec
+epoch 15, total_step 238200, total loss is 16.02 , inference loss is 9.64, weight deacy loss is 6.38, training accuracy is 0.500000, time 14.305 samples/sec
+epoch 15, total_step 238220, total loss is 15.44 , inference loss is 9.06, weight deacy loss is 6.38, training accuracy is 0.468750, time 219.512 samples/sec
+epoch 15, total_step 238240, total loss is 15.81 , inference loss is 9.43, weight deacy loss is 6.38, training accuracy is 0.468750, time 222.588 samples/sec
+epoch 15, total_step 238260, total loss is 16.36 , inference loss is 9.99, weight deacy loss is 6.38, training accuracy is 0.437500, time 218.478 samples/sec
+epoch 15, total_step 238280, total loss is 13.16 , inference loss is 6.78, weight deacy loss is 6.38, training accuracy is 0.531250, time 213.135 samples/sec
+epoch 15, total_step 238300, total loss is 12.02 , inference loss is 5.65, weight deacy loss is 6.38, training accuracy is 0.625000, time 298.268 samples/sec
+epoch 15, total_step 238320, total loss is 14.47 , inference loss is 8.09, weight deacy loss is 6.38, training accuracy is 0.531250, time 218.633 samples/sec
+epoch 15, total_step 238340, total loss is 9.01 , inference loss is 2.64, weight deacy loss is 6.38, training accuracy is 0.593750, time 225.107 samples/sec
+epoch 15, total_step 238360, total loss is 12.25 , inference loss is 5.87, weight deacy loss is 6.38, training accuracy is 0.562500, time 218.959 samples/sec
+epoch 15, total_step 238380, total loss is 11.54 , inference loss is 5.16, weight deacy loss is 6.38, training accuracy is 0.718750, time 212.126 samples/sec
+epoch 15, total_step 238400, total loss is 15.76 , inference loss is 9.39, weight deacy loss is 6.38, training accuracy is 0.468750, time 222.023 samples/sec
+epoch 15, total_step 238420, total loss is 14.59 , inference loss is 8.21, weight deacy loss is 6.38, training accuracy is 0.406250, time 213.644 samples/sec
+epoch 15, total_step 238440, total loss is 12.48 , inference loss is 6.10, weight deacy loss is 6.38, training accuracy is 0.468750, time 211.303 samples/sec
+epoch 15, total_step 238460, total loss is 10.92 , inference loss is 4.54, weight deacy loss is 6.38, training accuracy is 0.625000, time 221.612 samples/sec
+epoch 15, total_step 238480, total loss is 12.23 , inference loss is 5.86, weight deacy loss is 6.38, training accuracy is 0.531250, time 221.804 samples/sec
+epoch 15, total_step 238500, total loss is 14.86 , inference loss is 8.49, weight deacy loss is 6.38, training accuracy is 0.437500, time 14.310 samples/sec
+epoch 15, total_step 238520, total loss is 11.48 , inference loss is 5.10, weight deacy loss is 6.38, training accuracy is 0.500000, time 299.254 samples/sec
+epoch 15, total_step 238540, total loss is 13.72 , inference loss is 7.34, weight deacy loss is 6.38, training accuracy is 0.406250, time 296.767 samples/sec
+epoch 15, total_step 238560, total loss is 12.05 , inference loss is 5.67, weight deacy loss is 6.38, training accuracy is 0.500000, time 298.503 samples/sec
+epoch 15, total_step 238580, total loss is 12.22 , inference loss is 5.84, weight deacy loss is 6.38, training accuracy is 0.625000, time 208.049 samples/sec
+epoch 15, total_step 238600, total loss is 11.99 , inference loss is 5.61, weight deacy loss is 6.38, training accuracy is 0.625000, time 217.173 samples/sec
+epoch 15, total_step 238620, total loss is 9.53 , inference loss is 3.15, weight deacy loss is 6.38, training accuracy is 0.656250, time 212.662 samples/sec
+epoch 15, total_step 238640, total loss is 12.04 , inference loss is 5.66, weight deacy loss is 6.38, training accuracy is 0.531250, time 230.784 samples/sec
+epoch 15, total_step 238660, total loss is 12.35 , inference loss is 5.97, weight deacy loss is 6.38, training accuracy is 0.593750, time 218.273 samples/sec
+epoch 15, total_step 238680, total loss is 13.78 , inference loss is 7.40, weight deacy loss is 6.38, training accuracy is 0.531250, time 210.420 samples/sec
+epoch 15, total_step 238700, total loss is 16.04 , inference loss is 9.66, weight deacy loss is 6.38, training accuracy is 0.656250, time 214.052 samples/sec
+epoch 15, total_step 238720, total loss is 9.70 , inference loss is 3.33, weight deacy loss is 6.38, training accuracy is 0.687500, time 221.766 samples/sec
+epoch 15, total_step 238740, total loss is 13.32 , inference loss is 6.94, weight deacy loss is 6.38, training accuracy is 0.375000, time 217.605 samples/sec
+epoch 15, total_step 238760, total loss is 11.92 , inference loss is 5.54, weight deacy loss is 6.38, training accuracy is 0.593750, time 208.817 samples/sec
+epoch 15, total_step 238780, total loss is 12.75 , inference loss is 6.37, weight deacy loss is 6.38, training accuracy is 0.437500, time 200.618 samples/sec
+epoch 15, total_step 238800, total loss is 13.44 , inference loss is 7.06, weight deacy loss is 6.38, training accuracy is 0.625000, time 14.030 samples/sec
+epoch 15, total_step 238820, total loss is 12.63 , inference loss is 6.25, weight deacy loss is 6.38, training accuracy is 0.437500, time 214.906 samples/sec
+epoch 15, total_step 238840, total loss is 10.45 , inference loss is 4.07, weight deacy loss is 6.38, training accuracy is 0.562500, time 220.067 samples/sec
+epoch 15, total_step 238860, total loss is 14.05 , inference loss is 7.67, weight deacy loss is 6.38, training accuracy is 0.406250, time 224.538 samples/sec
+epoch 15, total_step 238880, total loss is 15.38 , inference loss is 9.00, weight deacy loss is 6.38, training accuracy is 0.375000, time 219.450 samples/sec
+epoch 15, total_step 238900, total loss is 12.96 , inference loss is 6.58, weight deacy loss is 6.38, training accuracy is 0.500000, time 213.400 samples/sec
+epoch 15, total_step 238920, total loss is 10.80 , inference loss is 4.42, weight deacy loss is 6.38, training accuracy is 0.468750, time 210.672 samples/sec
+epoch 15, total_step 238940, total loss is 15.19 , inference loss is 8.81, weight deacy loss is 6.38, training accuracy is 0.500000, time 218.674 samples/sec
+epoch 15, total_step 238960, total loss is 10.98 , inference loss is 4.60, weight deacy loss is 6.38, training accuracy is 0.593750, time 222.011 samples/sec
+epoch 15, total_step 238980, total loss is 12.41 , inference loss is 6.04, weight deacy loss is 6.38, training accuracy is 0.687500, time 217.100 samples/sec
+epoch 15, total_step 239000, total loss is 15.22 , inference loss is 8.84, weight deacy loss is 6.38, training accuracy is 0.437500, time 221.920 samples/sec
+epoch 15, total_step 239020, total loss is 16.88 , inference loss is 10.50, weight deacy loss is 6.38, training accuracy is 0.312500, time 213.443 samples/sec
+epoch 15, total_step 239040, total loss is 11.54 , inference loss is 5.16, weight deacy loss is 6.38, training accuracy is 0.531250, time 297.863 samples/sec
+epoch 15, total_step 239060, total loss is 10.73 , inference loss is 4.35, weight deacy loss is 6.38, training accuracy is 0.593750, time 211.877 samples/sec
+epoch 15, total_step 239080, total loss is 14.88 , inference loss is 8.50, weight deacy loss is 6.38, training accuracy is 0.437500, time 213.333 samples/sec
+epoch 15, total_step 239100, total loss is 13.97 , inference loss is 7.59, weight deacy loss is 6.38, training accuracy is 0.437500, time 12.723 samples/sec
+epoch 15, total_step 239120, total loss is 10.62 , inference loss is 4.25, weight deacy loss is 6.38, training accuracy is 0.593750, time 211.138 samples/sec
+epoch 15, total_step 239140, total loss is 12.71 , inference loss is 6.33, weight deacy loss is 6.38, training accuracy is 0.500000, time 213.325 samples/sec
+epoch 15, total_step 239160, total loss is 15.40 , inference loss is 9.02, weight deacy loss is 6.38, training accuracy is 0.593750, time 219.297 samples/sec
+epoch 15, total_step 239180, total loss is 12.90 , inference loss is 6.52, weight deacy loss is 6.38, training accuracy is 0.500000, time 221.688 samples/sec
+epoch 15, total_step 239200, total loss is 11.47 , inference loss is 5.09, weight deacy loss is 6.38, training accuracy is 0.593750, time 221.984 samples/sec
+epoch 15, total_step 239220, total loss is 16.00 , inference loss is 9.62, weight deacy loss is 6.38, training accuracy is 0.343750, time 208.592 samples/sec
+epoch 15, total_step 239240, total loss is 13.72 , inference loss is 7.34, weight deacy loss is 6.38, training accuracy is 0.468750, time 219.632 samples/sec
+epoch 15, total_step 239260, total loss is 13.31 , inference loss is 6.93, weight deacy loss is 6.38, training accuracy is 0.593750, time 214.189 samples/sec
+epoch 15, total_step 239280, total loss is 11.26 , inference loss is 4.88, weight deacy loss is 6.38, training accuracy is 0.531250, time 213.073 samples/sec
+epoch 15, total_step 239300, total loss is 12.97 , inference loss is 6.59, weight deacy loss is 6.38, training accuracy is 0.500000, time 298.506 samples/sec
+epoch 15, total_step 239320, total loss is 14.90 , inference loss is 8.52, weight deacy loss is 6.38, training accuracy is 0.437500, time 213.650 samples/sec
+epoch 15, total_step 239340, total loss is 9.47 , inference loss is 3.09, weight deacy loss is 6.38, training accuracy is 0.625000, time 212.687 samples/sec
+epoch 15, total_step 239360, total loss is 12.16 , inference loss is 5.78, weight deacy loss is 6.38, training accuracy is 0.531250, time 224.723 samples/sec
+epoch 15, total_step 239380, total loss is 16.58 , inference loss is 10.20, weight deacy loss is 6.38, training accuracy is 0.468750, time 219.549 samples/sec
+epoch 15, total_step 239400, total loss is 12.99 , inference loss is 6.61, weight deacy loss is 6.38, training accuracy is 0.562500, time 12.598 samples/sec
+epoch 15, total_step 239420, total loss is 12.47 , inference loss is 6.09, weight deacy loss is 6.38, training accuracy is 0.531250, time 217.087 samples/sec
+epoch 15, total_step 239440, total loss is 12.92 , inference loss is 6.54, weight deacy loss is 6.38, training accuracy is 0.468750, time 215.901 samples/sec
+epoch 15, total_step 239460, total loss is 10.77 , inference loss is 4.39, weight deacy loss is 6.38, training accuracy is 0.687500, time 212.868 samples/sec
+epoch 15, total_step 239480, total loss is 16.69 , inference loss is 10.31, weight deacy loss is 6.38, training accuracy is 0.375000, time 222.561 samples/sec
+epoch 15, total_step 239500, total loss is 14.26 , inference loss is 7.88, weight deacy loss is 6.38, training accuracy is 0.468750, time 214.754 samples/sec
+epoch 15, total_step 239520, total loss is 11.94 , inference loss is 5.56, weight deacy loss is 6.38, training accuracy is 0.562500, time 212.558 samples/sec
+epoch 15, total_step 239540, total loss is 12.03 , inference loss is 5.65, weight deacy loss is 6.38, training accuracy is 0.593750, time 214.475 samples/sec
+epoch 15, total_step 239560, total loss is 15.54 , inference loss is 9.16, weight deacy loss is 6.38, training accuracy is 0.437500, time 298.164 samples/sec
+epoch 15, total_step 239580, total loss is 15.69 , inference loss is 9.31, weight deacy loss is 6.38, training accuracy is 0.437500, time 223.331 samples/sec
+epoch 15, total_step 239600, total loss is 8.88 , inference loss is 2.50, weight deacy loss is 6.38, training accuracy is 0.656250, time 222.037 samples/sec
+epoch 15, total_step 239620, total loss is 10.26 , inference loss is 3.88, weight deacy loss is 6.38, training accuracy is 0.500000, time 215.807 samples/sec
+epoch 15, total_step 239640, total loss is 10.34 , inference loss is 3.96, weight deacy loss is 6.38, training accuracy is 0.562500, time 222.551 samples/sec
+epoch 15, total_step 239660, total loss is 10.77 , inference loss is 4.39, weight deacy loss is 6.38, training accuracy is 0.625000, time 214.645 samples/sec
+epoch 15, total_step 239680, total loss is 11.69 , inference loss is 5.31, weight deacy loss is 6.38, training accuracy is 0.562500, time 240.121 samples/sec
+epoch 15, total_step 239700, total loss is 13.14 , inference loss is 6.76, weight deacy loss is 6.38, training accuracy is 0.500000, time 12.421 samples/sec
+epoch 15, total_step 239720, total loss is 9.50 , inference loss is 3.12, weight deacy loss is 6.38, training accuracy is 0.562500, time 215.292 samples/sec
+epoch 15, total_step 239740, total loss is 13.78 , inference loss is 7.40, weight deacy loss is 6.38, training accuracy is 0.375000, time 215.031 samples/sec
+epoch 15, total_step 239760, total loss is 14.00 , inference loss is 7.62, weight deacy loss is 6.38, training accuracy is 0.562500, time 210.578 samples/sec
+epoch 15, total_step 239780, total loss is 10.52 , inference loss is 4.14, weight deacy loss is 6.38, training accuracy is 0.687500, time 215.106 samples/sec
+epoch 15, total_step 239800, total loss is 11.86 , inference loss is 5.48, weight deacy loss is 6.38, training accuracy is 0.625000, time 215.938 samples/sec
+epoch 15, total_step 239820, total loss is 13.82 , inference loss is 7.44, weight deacy loss is 6.38, training accuracy is 0.437500, time 298.759 samples/sec
+epoch 15, total_step 239840, total loss is 10.96 , inference loss is 4.58, weight deacy loss is 6.38, training accuracy is 0.562500, time 218.833 samples/sec
+epoch 15, total_step 239860, total loss is 10.64 , inference loss is 4.27, weight deacy loss is 6.38, training accuracy is 0.593750, time 220.723 samples/sec
+epoch 15, total_step 239880, total loss is 10.80 , inference loss is 4.42, weight deacy loss is 6.38, training accuracy is 0.562500, time 235.106 samples/sec
+epoch 15, total_step 239900, total loss is 12.62 , inference loss is 6.24, weight deacy loss is 6.38, training accuracy is 0.562500, time 227.129 samples/sec
+epoch 15, total_step 239920, total loss is 13.23 , inference loss is 6.85, weight deacy loss is 6.38, training accuracy is 0.406250, time 214.420 samples/sec
+epoch 15, total_step 239940, total loss is 11.84 , inference loss is 5.46, weight deacy loss is 6.38, training accuracy is 0.718750, time 213.980 samples/sec
+epoch 15, total_step 239960, total loss is 12.47 , inference loss is 6.09, weight deacy loss is 6.38, training accuracy is 0.562500, time 219.383 samples/sec
+epoch 15, total_step 239980, total loss is 17.94 , inference loss is 11.56, weight deacy loss is 6.38, training accuracy is 0.468750, time 215.660 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.888999999999985
+best_threshold_index 121 0.9866666666666667
+best_threshold_index 118 0.9866666666666667
+best_threshold_index 121 0.9855555555555555
+best_threshold_index 118 0.9866666666666667
+best_threshold_index 116 0.9859259259259259
+best_threshold_index 121 0.9853703703703703
+best_threshold_index 116 0.9862962962962963
+best_threshold_index 118 0.9859259259259259
+best_threshold_index 121 0.9848148148148148
+best_threshold_index 118 0.9853703703703703
+[lfw][240000]XNorm: 22.500482
+[lfw][240000]Accuracy-Flip: 0.98467+-0.00567
+test accuracy is: 0.9846666666666666
+epoch 15, total_step 240000, total loss is 15.03 , inference loss is 8.66, weight deacy loss is 6.38, training accuracy is 0.593750, time 16.321 samples/sec
+epoch 15, total_step 240020, total loss is 11.12 , inference loss is 4.75, weight deacy loss is 6.38, training accuracy is 0.593750, time 219.774 samples/sec
+epoch 15, total_step 240040, total loss is 12.58 , inference loss is 6.20, weight deacy loss is 6.38, training accuracy is 0.593750, time 211.965 samples/sec
+epoch 15, total_step 240060, total loss is 10.21 , inference loss is 3.83, weight deacy loss is 6.38, training accuracy is 0.625000, time 219.814 samples/sec
+epoch 15, total_step 240080, total loss is 12.54 , inference loss is 6.16, weight deacy loss is 6.38, training accuracy is 0.406250, time 221.923 samples/sec
+epoch 15, total_step 240100, total loss is 15.52 , inference loss is 9.14, weight deacy loss is 6.38, training accuracy is 0.468750, time 214.059 samples/sec
+epoch 15, total_step 240120, total loss is 12.79 , inference loss is 6.41, weight deacy loss is 6.38, training accuracy is 0.625000, time 216.146 samples/sec
+epoch 15, total_step 240140, total loss is 13.82 , inference loss is 7.44, weight deacy loss is 6.38, training accuracy is 0.500000, time 222.604 samples/sec
+epoch 15, total_step 240160, total loss is 11.08 , inference loss is 4.70, weight deacy loss is 6.38, training accuracy is 0.406250, time 223.199 samples/sec
+epoch 15, total_step 240180, total loss is 15.30 , inference loss is 8.92, weight deacy loss is 6.38, training accuracy is 0.531250, time 213.451 samples/sec
+epoch 15, total_step 240200, total loss is 15.00 , inference loss is 8.62, weight deacy loss is 6.38, training accuracy is 0.343750, time 213.807 samples/sec
+epoch 15, total_step 240220, total loss is 12.59 , inference loss is 6.22, weight deacy loss is 6.38, training accuracy is 0.531250, time 213.243 samples/sec
+epoch 15, total_step 240240, total loss is 12.98 , inference loss is 6.60, weight deacy loss is 6.38, training accuracy is 0.625000, time 219.343 samples/sec
+epoch 15, total_step 240260, total loss is 11.19 , inference loss is 4.81, weight deacy loss is 6.38, training accuracy is 0.562500, time 216.536 samples/sec
+epoch 15, total_step 240280, total loss is 15.41 , inference loss is 9.03, weight deacy loss is 6.38, training accuracy is 0.593750, time 216.864 samples/sec
+epoch 15, total_step 240300, total loss is 13.39 , inference loss is 7.01, weight deacy loss is 6.38, training accuracy is 0.625000, time 13.356 samples/sec
+epoch 15, total_step 240320, total loss is 12.16 , inference loss is 5.78, weight deacy loss is 6.38, training accuracy is 0.531250, time 218.305 samples/sec
+epoch 15, total_step 240340, total loss is 12.21 , inference loss is 5.83, weight deacy loss is 6.38, training accuracy is 0.593750, time 209.926 samples/sec
+epoch 15, total_step 240360, total loss is 9.80 , inference loss is 3.42, weight deacy loss is 6.38, training accuracy is 0.687500, time 210.594 samples/sec
+epoch 15, total_step 240380, total loss is 13.66 , inference loss is 7.28, weight deacy loss is 6.38, training accuracy is 0.437500, time 222.412 samples/sec
+epoch 15, total_step 240400, total loss is 11.67 , inference loss is 5.29, weight deacy loss is 6.38, training accuracy is 0.500000, time 220.731 samples/sec
+epoch 15, total_step 240420, total loss is 11.76 , inference loss is 5.38, weight deacy loss is 6.38, training accuracy is 0.468750, time 213.211 samples/sec
+epoch 15, total_step 240440, total loss is 12.44 , inference loss is 6.07, weight deacy loss is 6.38, training accuracy is 0.500000, time 216.179 samples/sec
+epoch 15, total_step 240460, total loss is 11.00 , inference loss is 4.62, weight deacy loss is 6.38, training accuracy is 0.562500, time 218.756 samples/sec
+epoch 15, total_step 240480, total loss is 15.19 , inference loss is 8.81, weight deacy loss is 6.38, training accuracy is 0.406250, time 217.382 samples/sec
+epoch 15, total_step 240500, total loss is 13.89 , inference loss is 7.51, weight deacy loss is 6.38, training accuracy is 0.531250, time 224.282 samples/sec
+epoch 15, total_step 240520, total loss is 10.75 , inference loss is 4.37, weight deacy loss is 6.38, training accuracy is 0.500000, time 222.934 samples/sec
+epoch 15, total_step 240540, total loss is 10.60 , inference loss is 4.22, weight deacy loss is 6.38, training accuracy is 0.500000, time 223.233 samples/sec
+epoch 15, total_step 240560, total loss is 13.20 , inference loss is 6.83, weight deacy loss is 6.38, training accuracy is 0.500000, time 224.702 samples/sec
+epoch 15, total_step 240580, total loss is 13.69 , inference loss is 7.31, weight deacy loss is 6.38, training accuracy is 0.437500, time 229.429 samples/sec
+epoch 15, total_step 240600, total loss is 12.43 , inference loss is 6.06, weight deacy loss is 6.38, training accuracy is 0.500000, time 12.718 samples/sec
+epoch 15, total_step 240620, total loss is 11.28 , inference loss is 4.91, weight deacy loss is 6.38, training accuracy is 0.531250, time 216.685 samples/sec
+epoch 15, total_step 240640, total loss is 12.34 , inference loss is 5.96, weight deacy loss is 6.38, training accuracy is 0.468750, time 219.463 samples/sec
+epoch 15, total_step 240660, total loss is 14.47 , inference loss is 8.10, weight deacy loss is 6.38, training accuracy is 0.468750, time 209.513 samples/sec
+epoch 15, total_step 240680, total loss is 15.93 , inference loss is 9.55, weight deacy loss is 6.38, training accuracy is 0.406250, time 220.292 samples/sec
+epoch 15, total_step 240700, total loss is 11.80 , inference loss is 5.43, weight deacy loss is 6.38, training accuracy is 0.531250, time 214.849 samples/sec
+epoch 15, total_step 240720, total loss is 10.20 , inference loss is 3.82, weight deacy loss is 6.38, training accuracy is 0.656250, time 226.816 samples/sec
+epoch 15, total_step 240740, total loss is 9.84 , inference loss is 3.46, weight deacy loss is 6.38, training accuracy is 0.593750, time 214.416 samples/sec
+epoch 15, total_step 240760, total loss is 12.68 , inference loss is 6.30, weight deacy loss is 6.38, training accuracy is 0.500000, time 224.959 samples/sec
+epoch 15, total_step 240780, total loss is 14.97 , inference loss is 8.60, weight deacy loss is 6.38, training accuracy is 0.500000, time 220.819 samples/sec
+epoch 15, total_step 240800, total loss is 11.57 , inference loss is 5.19, weight deacy loss is 6.38, training accuracy is 0.531250, time 216.205 samples/sec
+epoch 15, total_step 240820, total loss is 11.31 , inference loss is 4.94, weight deacy loss is 6.38, training accuracy is 0.562500, time 215.920 samples/sec
+epoch 15, total_step 240840, total loss is 11.41 , inference loss is 5.03, weight deacy loss is 6.38, training accuracy is 0.375000, time 234.174 samples/sec
+epoch 15, total_step 240860, total loss is 16.21 , inference loss is 9.84, weight deacy loss is 6.38, training accuracy is 0.500000, time 299.906 samples/sec
+epoch 15, total_step 240880, total loss is 13.96 , inference loss is 7.58, weight deacy loss is 6.38, training accuracy is 0.437500, time 202.971 samples/sec
+epoch 15, total_step 240900, total loss is 10.57 , inference loss is 4.20, weight deacy loss is 6.38, training accuracy is 0.500000, time 12.377 samples/sec
+epoch 15, total_step 240920, total loss is 15.36 , inference loss is 8.99, weight deacy loss is 6.38, training accuracy is 0.562500, time 217.876 samples/sec
+epoch 15, total_step 240940, total loss is 12.27 , inference loss is 5.89, weight deacy loss is 6.38, training accuracy is 0.343750, time 226.808 samples/sec
+epoch 15, total_step 240960, total loss is 10.77 , inference loss is 4.39, weight deacy loss is 6.38, training accuracy is 0.656250, time 213.532 samples/sec
+epoch 15, total_step 240980, total loss is 11.01 , inference loss is 4.64, weight deacy loss is 6.38, training accuracy is 0.468750, time 225.368 samples/sec
+epoch 15, total_step 241000, total loss is 10.59 , inference loss is 4.21, weight deacy loss is 6.38, training accuracy is 0.625000, time 243.672 samples/sec
+epoch 15, total_step 241020, total loss is 13.96 , inference loss is 7.58, weight deacy loss is 6.38, training accuracy is 0.562500, time 222.038 samples/sec
+epoch 15, total_step 241040, total loss is 14.96 , inference loss is 8.58, weight deacy loss is 6.38, training accuracy is 0.406250, time 227.731 samples/sec
+epoch 15, total_step 241060, total loss is 13.26 , inference loss is 6.89, weight deacy loss is 6.38, training accuracy is 0.531250, time 220.207 samples/sec
+epoch 15, total_step 241080, total loss is 12.02 , inference loss is 5.64, weight deacy loss is 6.38, training accuracy is 0.468750, time 223.795 samples/sec
+epoch 15, total_step 241100, total loss is 15.20 , inference loss is 8.82, weight deacy loss is 6.38, training accuracy is 0.406250, time 222.209 samples/sec
+epoch 15, total_step 241120, total loss is 14.62 , inference loss is 8.24, weight deacy loss is 6.38, training accuracy is 0.500000, time 271.818 samples/sec
+epoch 15, total_step 241140, total loss is 10.16 , inference loss is 3.78, weight deacy loss is 6.38, training accuracy is 0.718750, time 216.004 samples/sec
+epoch 15, total_step 241160, total loss is 16.37 , inference loss is 10.00, weight deacy loss is 6.38, training accuracy is 0.312500, time 210.835 samples/sec
+epoch 15, total_step 241180, total loss is 11.12 , inference loss is 4.74, weight deacy loss is 6.38, training accuracy is 0.562500, time 211.837 samples/sec
+epoch 15, total_step 241200, total loss is 16.22 , inference loss is 9.84, weight deacy loss is 6.38, training accuracy is 0.375000, time 13.279 samples/sec
+epoch 15, total_step 241220, total loss is 12.02 , inference loss is 5.65, weight deacy loss is 6.38, training accuracy is 0.562500, time 215.434 samples/sec
+epoch 15, total_step 241240, total loss is 14.23 , inference loss is 7.86, weight deacy loss is 6.38, training accuracy is 0.531250, time 223.986 samples/sec
+epoch 15, total_step 241260, total loss is 10.68 , inference loss is 4.30, weight deacy loss is 6.38, training accuracy is 0.593750, time 226.711 samples/sec
+epoch 15, total_step 241280, total loss is 11.63 , inference loss is 5.26, weight deacy loss is 6.38, training accuracy is 0.468750, time 226.483 samples/sec
+epoch 15, total_step 241300, total loss is 12.71 , inference loss is 6.33, weight deacy loss is 6.38, training accuracy is 0.375000, time 213.882 samples/sec
+epoch 15, total_step 241320, total loss is 12.96 , inference loss is 6.59, weight deacy loss is 6.38, training accuracy is 0.406250, time 220.628 samples/sec
+epoch 15, total_step 241340, total loss is 14.28 , inference loss is 7.91, weight deacy loss is 6.38, training accuracy is 0.437500, time 217.704 samples/sec
+epoch 15, total_step 241360, total loss is 11.30 , inference loss is 4.92, weight deacy loss is 6.38, training accuracy is 0.500000, time 233.408 samples/sec
+epoch 15, total_step 241380, total loss is 11.75 , inference loss is 5.37, weight deacy loss is 6.38, training accuracy is 0.437500, time 218.107 samples/sec
+epoch 15, total_step 241400, total loss is 14.08 , inference loss is 7.70, weight deacy loss is 6.38, training accuracy is 0.406250, time 254.259 samples/sec
+epoch 15, total_step 241420, total loss is 14.83 , inference loss is 8.45, weight deacy loss is 6.38, training accuracy is 0.437500, time 221.579 samples/sec
+epoch 15, total_step 241440, total loss is 14.34 , inference loss is 7.97, weight deacy loss is 6.38, training accuracy is 0.500000, time 223.005 samples/sec
+epoch 15, total_step 241460, total loss is 13.86 , inference loss is 7.48, weight deacy loss is 6.38, training accuracy is 0.625000, time 217.926 samples/sec
+epoch 15, total_step 241480, total loss is 16.42 , inference loss is 10.04, weight deacy loss is 6.38, training accuracy is 0.406250, time 218.438 samples/sec
+epoch 15, total_step 241500, total loss is 11.22 , inference loss is 4.84, weight deacy loss is 6.38, training accuracy is 0.531250, time 12.509 samples/sec
+epoch 15, total_step 241520, total loss is 19.13 , inference loss is 12.75, weight deacy loss is 6.38, training accuracy is 0.312500, time 213.720 samples/sec
+epoch 15, total_step 241540, total loss is 11.57 , inference loss is 5.19, weight deacy loss is 6.38, training accuracy is 0.375000, time 207.726 samples/sec
+epoch 15, total_step 241560, total loss is 13.21 , inference loss is 6.83, weight deacy loss is 6.38, training accuracy is 0.531250, time 221.311 samples/sec
+epoch 15, total_step 241580, total loss is 12.86 , inference loss is 6.48, weight deacy loss is 6.38, training accuracy is 0.437500, time 217.009 samples/sec
+epoch 15, total_step 241600, total loss is 15.93 , inference loss is 9.56, weight deacy loss is 6.38, training accuracy is 0.531250, time 213.105 samples/sec
+epoch 15, total_step 241620, total loss is 11.62 , inference loss is 5.25, weight deacy loss is 6.38, training accuracy is 0.562500, time 223.757 samples/sec
+epoch 15, total_step 241640, total loss is 12.31 , inference loss is 5.93, weight deacy loss is 6.38, training accuracy is 0.625000, time 221.856 samples/sec
+epoch 15, total_step 241660, total loss is 14.03 , inference loss is 7.65, weight deacy loss is 6.38, training accuracy is 0.375000, time 299.325 samples/sec
+epoch 15, total_step 241680, total loss is 13.63 , inference loss is 7.25, weight deacy loss is 6.38, training accuracy is 0.437500, time 226.106 samples/sec
+epoch 15, total_step 241700, total loss is 12.37 , inference loss is 6.00, weight deacy loss is 6.38, training accuracy is 0.406250, time 221.341 samples/sec
+epoch 15, total_step 241720, total loss is 12.02 , inference loss is 5.65, weight deacy loss is 6.38, training accuracy is 0.593750, time 222.959 samples/sec
+epoch 15, total_step 241740, total loss is 10.81 , inference loss is 4.43, weight deacy loss is 6.38, training accuracy is 0.437500, time 215.609 samples/sec
+epoch 15, total_step 241760, total loss is 13.58 , inference loss is 7.20, weight deacy loss is 6.38, training accuracy is 0.406250, time 217.682 samples/sec
+epoch 15, total_step 241780, total loss is 13.80 , inference loss is 7.42, weight deacy loss is 6.38, training accuracy is 0.500000, time 219.016 samples/sec
+epoch 15, total_step 241800, total loss is 11.49 , inference loss is 5.11, weight deacy loss is 6.38, training accuracy is 0.500000, time 12.810 samples/sec
+epoch 15, total_step 241820, total loss is 11.88 , inference loss is 5.50, weight deacy loss is 6.38, training accuracy is 0.468750, time 224.643 samples/sec
+epoch 15, total_step 241840, total loss is 16.88 , inference loss is 10.50, weight deacy loss is 6.38, training accuracy is 0.437500, time 224.992 samples/sec
+epoch 15, total_step 241860, total loss is 14.39 , inference loss is 8.01, weight deacy loss is 6.38, training accuracy is 0.281250, time 217.054 samples/sec
+epoch 15, total_step 241880, total loss is 14.50 , inference loss is 8.12, weight deacy loss is 6.38, training accuracy is 0.343750, time 226.521 samples/sec
+epoch 15, total_step 241900, total loss is 11.50 , inference loss is 5.12, weight deacy loss is 6.38, training accuracy is 0.562500, time 221.726 samples/sec
+epoch 15, total_step 241920, total loss is 10.80 , inference loss is 4.42, weight deacy loss is 6.38, training accuracy is 0.531250, time 296.782 samples/sec
+epoch 15, total_step 241940, total loss is 13.67 , inference loss is 7.29, weight deacy loss is 6.38, training accuracy is 0.312500, time 212.257 samples/sec
+epoch 15, total_step 241960, total loss is 17.31 , inference loss is 10.93, weight deacy loss is 6.38, training accuracy is 0.375000, time 222.787 samples/sec
+epoch 15, total_step 241980, total loss is 12.63 , inference loss is 6.25, weight deacy loss is 6.38, training accuracy is 0.406250, time 220.373 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.56913900000002
+best_threshold_index 109 0.9855555555555555
+best_threshold_index 109 0.985
+best_threshold_index 109 0.9846296296296296
+best_threshold_index 109 0.9864814814814815
+best_threshold_index 108 0.9855555555555555
+best_threshold_index 109 0.9842592592592593
+best_threshold_index 109 0.985
+best_threshold_index 109 0.985
+best_threshold_index 109 0.9838888888888889
+best_threshold_index 109 0.9846296296296296
+[lfw][242000]XNorm: 21.562455
+[lfw][242000]Accuracy-Flip: 0.98467+-0.00662
+test accuracy is: 0.9846666666666666
+epoch 15, total_step 242000, total loss is 16.59 , inference loss is 10.21, weight deacy loss is 6.38, training accuracy is 0.468750, time 16.473 samples/sec
+epoch 15, total_step 242020, total loss is 17.97 , inference loss is 11.59, weight deacy loss is 6.38, training accuracy is 0.375000, time 217.762 samples/sec
+epoch 15, total_step 242040, total loss is 13.28 , inference loss is 6.90, weight deacy loss is 6.38, training accuracy is 0.500000, time 222.483 samples/sec
+epoch 15, total_step 242060, total loss is 12.26 , inference loss is 5.88, weight deacy loss is 6.38, training accuracy is 0.625000, time 219.582 samples/sec
+epoch 15, total_step 242080, total loss is 14.62 , inference loss is 8.24, weight deacy loss is 6.38, training accuracy is 0.406250, time 296.896 samples/sec
+epoch 15, total_step 242100, total loss is 16.32 , inference loss is 9.94, weight deacy loss is 6.38, training accuracy is 0.468750, time 12.457 samples/sec
+epoch 15, total_step 242120, total loss is 14.19 , inference loss is 7.81, weight deacy loss is 6.38, training accuracy is 0.375000, time 209.944 samples/sec
+epoch 15, total_step 242140, total loss is 11.52 , inference loss is 5.14, weight deacy loss is 6.38, training accuracy is 0.343750, time 218.218 samples/sec
+epoch 15, total_step 242160, total loss is 13.45 , inference loss is 7.07, weight deacy loss is 6.38, training accuracy is 0.375000, time 296.585 samples/sec
+epoch 15, total_step 242180, total loss is 14.21 , inference loss is 7.84, weight deacy loss is 6.38, training accuracy is 0.312500, time 199.028 samples/sec
+epoch 15, total_step 242200, total loss is 10.06 , inference loss is 3.69, weight deacy loss is 6.38, training accuracy is 0.593750, time 209.217 samples/sec
+epoch 15, total_step 242220, total loss is 11.12 , inference loss is 4.75, weight deacy loss is 6.38, training accuracy is 0.531250, time 214.163 samples/sec
+epoch 15, total_step 242240, total loss is 12.56 , inference loss is 6.19, weight deacy loss is 6.38, training accuracy is 0.531250, time 232.706 samples/sec
+epoch 15, total_step 242260, total loss is 15.39 , inference loss is 9.01, weight deacy loss is 6.38, training accuracy is 0.593750, time 207.762 samples/sec
+epoch 15, total_step 242280, total loss is 14.59 , inference loss is 8.22, weight deacy loss is 6.38, training accuracy is 0.312500, time 215.322 samples/sec
+epoch 15, total_step 242300, total loss is 14.83 , inference loss is 8.45, weight deacy loss is 6.38, training accuracy is 0.468750, time 209.686 samples/sec
+epoch 15, total_step 242320, total loss is 14.81 , inference loss is 8.43, weight deacy loss is 6.38, training accuracy is 0.468750, time 224.393 samples/sec
+epoch 15, total_step 242340, total loss is 13.13 , inference loss is 6.76, weight deacy loss is 6.38, training accuracy is 0.437500, time 222.145 samples/sec
+epoch 15, total_step 242360, total loss is 16.19 , inference loss is 9.82, weight deacy loss is 6.38, training accuracy is 0.281250, time 227.569 samples/sec
+epoch 15, total_step 242380, total loss is 11.57 , inference loss is 5.19, weight deacy loss is 6.38, training accuracy is 0.406250, time 217.713 samples/sec
+epoch 15, total_step 242400, total loss is 16.69 , inference loss is 10.31, weight deacy loss is 6.38, training accuracy is 0.406250, time 12.421 samples/sec
+epoch 15, total_step 242420, total loss is 11.61 , inference loss is 5.23, weight deacy loss is 6.38, training accuracy is 0.562500, time 300.195 samples/sec
+epoch 15, total_step 242440, total loss is 15.42 , inference loss is 9.04, weight deacy loss is 6.38, training accuracy is 0.312500, time 207.586 samples/sec
+epoch 15, total_step 242460, total loss is 12.38 , inference loss is 6.00, weight deacy loss is 6.38, training accuracy is 0.531250, time 228.659 samples/sec
+epoch 15, total_step 242480, total loss is 20.69 , inference loss is 14.31, weight deacy loss is 6.38, training accuracy is 0.250000, time 221.767 samples/sec
+epoch 15, total_step 242500, total loss is 14.42 , inference loss is 8.05, weight deacy loss is 6.38, training accuracy is 0.343750, time 210.552 samples/sec
+epoch 15, total_step 242520, total loss is 9.49 , inference loss is 3.12, weight deacy loss is 6.38, training accuracy is 0.562500, time 220.199 samples/sec
+epoch 15, total_step 242540, total loss is 14.73 , inference loss is 8.36, weight deacy loss is 6.38, training accuracy is 0.406250, time 214.927 samples/sec
+epoch 15, total_step 242560, total loss is 11.16 , inference loss is 4.79, weight deacy loss is 6.38, training accuracy is 0.531250, time 215.621 samples/sec
+epoch 15, total_step 242580, total loss is 17.22 , inference loss is 10.85, weight deacy loss is 6.37, training accuracy is 0.343750, time 210.654 samples/sec
+epoch 15, total_step 242600, total loss is 9.67 , inference loss is 3.30, weight deacy loss is 6.37, training accuracy is 0.562500, time 213.998 samples/sec
+epoch 15, total_step 242620, total loss is 12.15 , inference loss is 5.78, weight deacy loss is 6.37, training accuracy is 0.468750, time 207.079 samples/sec
+epoch 15, total_step 242640, total loss is 16.79 , inference loss is 10.42, weight deacy loss is 6.37, training accuracy is 0.281250, time 208.623 samples/sec
+epoch 15, total_step 242660, total loss is 16.86 , inference loss is 10.48, weight deacy loss is 6.37, training accuracy is 0.250000, time 211.086 samples/sec
+epoch 15, total_step 242680, total loss is 14.61 , inference loss is 8.23, weight deacy loss is 6.37, training accuracy is 0.375000, time 216.160 samples/sec
+epoch 15, total_step 242700, total loss is 10.07 , inference loss is 3.69, weight deacy loss is 6.37, training accuracy is 0.437500, time 13.329 samples/sec
+epoch 15, total_step 242720, total loss is 11.52 , inference loss is 5.15, weight deacy loss is 6.37, training accuracy is 0.531250, time 237.162 samples/sec
+epoch 15, total_step 242740, total loss is 11.11 , inference loss is 4.74, weight deacy loss is 6.37, training accuracy is 0.562500, time 212.806 samples/sec
+epoch 15, total_step 242760, total loss is 13.06 , inference loss is 6.69, weight deacy loss is 6.37, training accuracy is 0.406250, time 225.904 samples/sec
+epoch 15, total_step 242780, total loss is 12.90 , inference loss is 6.53, weight deacy loss is 6.37, training accuracy is 0.500000, time 234.680 samples/sec
+epoch 15, total_step 242800, total loss is 12.91 , inference loss is 6.54, weight deacy loss is 6.37, training accuracy is 0.437500, time 226.307 samples/sec
+epoch 15, total_step 242820, total loss is 13.22 , inference loss is 6.85, weight deacy loss is 6.37, training accuracy is 0.281250, time 221.247 samples/sec
+epoch 15, total_step 242840, total loss is 13.43 , inference loss is 7.06, weight deacy loss is 6.37, training accuracy is 0.406250, time 218.020 samples/sec
+epoch 15, total_step 242860, total loss is 12.59 , inference loss is 6.21, weight deacy loss is 6.37, training accuracy is 0.406250, time 216.174 samples/sec
+epoch 15, total_step 242880, total loss is 12.18 , inference loss is 5.81, weight deacy loss is 6.37, training accuracy is 0.468750, time 216.563 samples/sec
+epoch 15, total_step 242900, total loss is 14.12 , inference loss is 7.75, weight deacy loss is 6.37, training accuracy is 0.343750, time 221.524 samples/sec
+epoch 15, total_step 242920, total loss is 15.79 , inference loss is 9.42, weight deacy loss is 6.37, training accuracy is 0.250000, time 216.560 samples/sec
+epoch 15, total_step 242940, total loss is 17.23 , inference loss is 10.85, weight deacy loss is 6.37, training accuracy is 0.281250, time 211.673 samples/sec
+epoch 15, total_step 242960, total loss is 13.17 , inference loss is 6.80, weight deacy loss is 6.37, training accuracy is 0.312500, time 217.764 samples/sec
+epoch 15, total_step 242980, total loss is 13.34 , inference loss is 6.96, weight deacy loss is 6.37, training accuracy is 0.500000, time 218.105 samples/sec
+epoch 15, total_step 243000, total loss is 13.01 , inference loss is 6.64, weight deacy loss is 6.37, training accuracy is 0.531250, time 12.418 samples/sec
+epoch 15, total_step 243020, total loss is 13.66 , inference loss is 7.29, weight deacy loss is 6.37, training accuracy is 0.468750, time 220.127 samples/sec
+epoch 15, total_step 243040, total loss is 15.18 , inference loss is 8.81, weight deacy loss is 6.37, training accuracy is 0.437500, time 213.103 samples/sec
+epoch 15, total_step 243060, total loss is 11.49 , inference loss is 5.12, weight deacy loss is 6.37, training accuracy is 0.625000, time 210.076 samples/sec
+epoch 15, total_step 243080, total loss is 15.29 , inference loss is 8.91, weight deacy loss is 6.37, training accuracy is 0.406250, time 220.117 samples/sec
+epoch 15, total_step 243100, total loss is 15.53 , inference loss is 9.15, weight deacy loss is 6.37, training accuracy is 0.406250, time 218.276 samples/sec
+epoch 15, total_step 243120, total loss is 11.42 , inference loss is 5.05, weight deacy loss is 6.37, training accuracy is 0.437500, time 214.621 samples/sec
+epoch 15, total_step 243140, total loss is 12.98 , inference loss is 6.60, weight deacy loss is 6.37, training accuracy is 0.531250, time 217.870 samples/sec
+epoch 15, total_step 243160, total loss is 13.80 , inference loss is 7.43, weight deacy loss is 6.37, training accuracy is 0.437500, time 218.231 samples/sec
+epoch 15, total_step 243180, total loss is 9.16 , inference loss is 2.79, weight deacy loss is 6.37, training accuracy is 0.500000, time 213.636 samples/sec
+epoch 15, total_step 243200, total loss is 13.78 , inference loss is 7.41, weight deacy loss is 6.37, training accuracy is 0.343750, time 224.255 samples/sec
+epoch 15, total_step 243220, total loss is 13.45 , inference loss is 7.08, weight deacy loss is 6.37, training accuracy is 0.406250, time 211.381 samples/sec
+epoch 15, total_step 243240, total loss is 16.58 , inference loss is 10.20, weight deacy loss is 6.37, training accuracy is 0.375000, time 224.189 samples/sec
+epoch 15, total_step 243260, total loss is 15.10 , inference loss is 8.73, weight deacy loss is 6.37, training accuracy is 0.531250, time 193.184 samples/sec
+epoch 15, total_step 243280, total loss is 10.04 , inference loss is 3.66, weight deacy loss is 6.37, training accuracy is 0.562500, time 199.966 samples/sec
+epoch 15, total_step 243300, total loss is 13.04 , inference loss is 6.66, weight deacy loss is 6.37, training accuracy is 0.375000, time 13.899 samples/sec
+epoch 15, total_step 243320, total loss is 13.93 , inference loss is 7.56, weight deacy loss is 6.37, training accuracy is 0.531250, time 220.765 samples/sec
+epoch 15, total_step 243340, total loss is 16.18 , inference loss is 9.80, weight deacy loss is 6.37, training accuracy is 0.437500, time 227.838 samples/sec
+epoch 15, total_step 243360, total loss is 12.83 , inference loss is 6.45, weight deacy loss is 6.37, training accuracy is 0.375000, time 207.985 samples/sec
+epoch 15, total_step 243380, total loss is 13.11 , inference loss is 6.73, weight deacy loss is 6.37, training accuracy is 0.500000, time 213.783 samples/sec
+epoch 15, total_step 243400, total loss is 19.72 , inference loss is 13.35, weight deacy loss is 6.37, training accuracy is 0.218750, time 216.852 samples/sec
+epoch 15, total_step 243420, total loss is 15.72 , inference loss is 9.34, weight deacy loss is 6.37, training accuracy is 0.406250, time 211.087 samples/sec
+epoch 15, total_step 243440, total loss is 14.11 , inference loss is 7.74, weight deacy loss is 6.37, training accuracy is 0.468750, time 225.230 samples/sec
+epoch 15, total_step 243460, total loss is 14.80 , inference loss is 8.43, weight deacy loss is 6.37, training accuracy is 0.375000, time 224.192 samples/sec
+epoch 15, total_step 243480, total loss is 17.08 , inference loss is 10.71, weight deacy loss is 6.37, training accuracy is 0.312500, time 215.562 samples/sec
+epoch 15, total_step 243500, total loss is 17.21 , inference loss is 10.84, weight deacy loss is 6.37, training accuracy is 0.343750, time 226.334 samples/sec
+epoch 15, total_step 243520, total loss is 14.78 , inference loss is 8.41, weight deacy loss is 6.37, training accuracy is 0.437500, time 222.612 samples/sec
+epoch 15, total_step 243540, total loss is 10.74 , inference loss is 4.37, weight deacy loss is 6.37, training accuracy is 0.406250, time 298.406 samples/sec
+epoch 15, total_step 243560, total loss is 14.06 , inference loss is 7.69, weight deacy loss is 6.37, training accuracy is 0.593750, time 212.312 samples/sec
+epoch 15, total_step 243580, total loss is 14.22 , inference loss is 7.85, weight deacy loss is 6.37, training accuracy is 0.468750, time 215.828 samples/sec
+epoch 15, total_step 243600, total loss is 11.16 , inference loss is 4.79, weight deacy loss is 6.37, training accuracy is 0.468750, time 12.914 samples/sec
+epoch 15, total_step 243620, total loss is 11.37 , inference loss is 5.00, weight deacy loss is 6.37, training accuracy is 0.500000, time 236.434 samples/sec
+epoch 15, total_step 243640, total loss is 12.25 , inference loss is 5.88, weight deacy loss is 6.37, training accuracy is 0.437500, time 209.584 samples/sec
+epoch 15, total_step 243660, total loss is 13.75 , inference loss is 7.38, weight deacy loss is 6.37, training accuracy is 0.562500, time 222.500 samples/sec
+epoch 15, total_step 243680, total loss is 14.30 , inference loss is 7.93, weight deacy loss is 6.37, training accuracy is 0.593750, time 233.843 samples/sec
+epoch 15, total_step 243700, total loss is 13.57 , inference loss is 7.19, weight deacy loss is 6.37, training accuracy is 0.406250, time 216.888 samples/sec
+epoch 15, total_step 243720, total loss is 11.84 , inference loss is 5.46, weight deacy loss is 6.37, training accuracy is 0.593750, time 212.561 samples/sec
+epoch 15, total_step 243740, total loss is 14.94 , inference loss is 8.57, weight deacy loss is 6.37, training accuracy is 0.312500, time 219.425 samples/sec
+epoch 15, total_step 243760, total loss is 14.94 , inference loss is 8.57, weight deacy loss is 6.37, training accuracy is 0.500000, time 215.910 samples/sec
+epoch 15, total_step 243780, total loss is 12.67 , inference loss is 6.30, weight deacy loss is 6.37, training accuracy is 0.468750, time 221.436 samples/sec
+epoch 15, total_step 243800, total loss is 13.03 , inference loss is 6.66, weight deacy loss is 6.37, training accuracy is 0.312500, time 296.953 samples/sec
+epoch 15, total_step 243820, total loss is 13.72 , inference loss is 7.35, weight deacy loss is 6.37, training accuracy is 0.406250, time 215.690 samples/sec
+epoch 15, total_step 243840, total loss is 14.52 , inference loss is 8.15, weight deacy loss is 6.37, training accuracy is 0.312500, time 218.438 samples/sec
+epoch 15, total_step 243860, total loss is 13.35 , inference loss is 6.98, weight deacy loss is 6.37, training accuracy is 0.531250, time 221.299 samples/sec
+epoch 15, total_step 243880, total loss is 13.70 , inference loss is 7.33, weight deacy loss is 6.37, training accuracy is 0.437500, time 226.808 samples/sec
+epoch 15, total_step 243900, total loss is 11.19 , inference loss is 4.81, weight deacy loss is 6.37, training accuracy is 0.531250, time 8.096 samples/sec
+epoch 15, total_step 243920, total loss is 13.51 , inference loss is 7.14, weight deacy loss is 6.37, training accuracy is 0.281250, time 212.397 samples/sec
+epoch 15, total_step 243940, total loss is 15.15 , inference loss is 8.78, weight deacy loss is 6.37, training accuracy is 0.343750, time 212.109 samples/sec
+epoch 15, total_step 243960, total loss is 12.99 , inference loss is 6.62, weight deacy loss is 6.37, training accuracy is 0.406250, time 210.472 samples/sec
+epoch 15, total_step 243980, total loss is 13.78 , inference loss is 7.41, weight deacy loss is 6.37, training accuracy is 0.375000, time 213.146 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.440799000000002
+best_threshold_index 109 0.9857407407407407
+best_threshold_index 111 0.985
+best_threshold_index 111 0.9846296296296296
+best_threshold_index 109 0.987037037037037
+best_threshold_index 111 0.9859259259259259
+best_threshold_index 111 0.9842592592592593
+best_threshold_index 111 0.9851851851851852
+best_threshold_index 111 0.9842592592592593
+best_threshold_index 111 0.9835185185185186
+best_threshold_index 111 0.9846296296296296
+[lfw][244000]XNorm: 20.884777
+[lfw][244000]Accuracy-Flip: 0.98450+-0.00910
+test accuracy is: 0.9845
+epoch 15, total_step 244000, total loss is 17.65 , inference loss is 11.28, weight deacy loss is 6.37, training accuracy is 0.343750, time 8.998 samples/sec
+epoch 15, total_step 244020, total loss is 16.53 , inference loss is 10.16, weight deacy loss is 6.37, training accuracy is 0.281250, time 207.327 samples/sec
+epoch 15, total_step 244040, total loss is 19.29 , inference loss is 12.92, weight deacy loss is 6.37, training accuracy is 0.281250, time 215.592 samples/sec
+epoch 15, total_step 244060, total loss is 12.97 , inference loss is 6.59, weight deacy loss is 6.37, training accuracy is 0.406250, time 203.922 samples/sec
+epoch 15, total_step 244080, total loss is 19.34 , inference loss is 12.97, weight deacy loss is 6.37, training accuracy is 0.218750, time 214.079 samples/sec
+epoch 15, total_step 244100, total loss is 15.02 , inference loss is 8.65, weight deacy loss is 6.37, training accuracy is 0.468750, time 214.017 samples/sec
+epoch 15, total_step 244120, total loss is 15.78 , inference loss is 9.41, weight deacy loss is 6.37, training accuracy is 0.281250, time 212.674 samples/sec
+epoch 15, total_step 244140, total loss is 17.45 , inference loss is 11.08, weight deacy loss is 6.37, training accuracy is 0.312500, time 202.365 samples/sec
+epoch 15, total_step 244160, total loss is 11.24 , inference loss is 4.87, weight deacy loss is 6.37, training accuracy is 0.468750, time 216.027 samples/sec
+epoch 15, total_step 244180, total loss is 13.90 , inference loss is 7.53, weight deacy loss is 6.37, training accuracy is 0.500000, time 218.618 samples/sec
+epoch 15, total_step 244200, total loss is 14.56 , inference loss is 8.19, weight deacy loss is 6.37, training accuracy is 0.437500, time 8.946 samples/sec
+epoch 15, total_step 244220, total loss is 16.10 , inference loss is 9.73, weight deacy loss is 6.37, training accuracy is 0.406250, time 239.831 samples/sec
+epoch 15, total_step 244240, total loss is 13.71 , inference loss is 7.34, weight deacy loss is 6.37, training accuracy is 0.250000, time 212.250 samples/sec
+epoch 15, total_step 244260, total loss is 14.58 , inference loss is 8.21, weight deacy loss is 6.37, training accuracy is 0.406250, time 215.825 samples/sec2022-12-01 16:37:21.565021: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7192 of 10000
+2022-12-01 16:37:25.400989: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 15, total_step 244280, total loss is 12.38 , inference loss is 6.01, weight deacy loss is 6.37, training accuracy is 0.500000, time 212.311 samples/sec
+epoch 15, total_step 244300, total loss is 12.31 , inference loss is 5.94, weight deacy loss is 6.37, training accuracy is 0.562500, time 225.271 samples/sec
+epoch 15, total_step 244320, total loss is 15.28 , inference loss is 8.91, weight deacy loss is 6.37, training accuracy is 0.375000, time 217.990 samples/sec
+epoch 15, total_step 244340, total loss is 15.52 , inference loss is 9.15, weight deacy loss is 6.37, training accuracy is 0.468750, time 215.923 samples/sec
+epoch 15, total_step 244360, total loss is 19.84 , inference loss is 13.47, weight deacy loss is 6.37, training accuracy is 0.187500, time 206.856 samples/sec
+epoch 15, total_step 244380, total loss is 12.17 , inference loss is 5.80, weight deacy loss is 6.37, training accuracy is 0.500000, time 212.992 samples/sec
+epoch 15, total_step 244400, total loss is 13.30 , inference loss is 6.93, weight deacy loss is 6.37, training accuracy is 0.343750, time 211.266 samples/sec
+epoch 15, total_step 244420, total loss is 14.14 , inference loss is 7.77, weight deacy loss is 6.37, training accuracy is 0.343750, time 196.302 samples/sec
+epoch 15, total_step 244440, total loss is 10.66 , inference loss is 4.29, weight deacy loss is 6.37, training accuracy is 0.593750, time 211.900 samples/sec
+epoch 15, total_step 244460, total loss is 15.86 , inference loss is 9.49, weight deacy loss is 6.37, training accuracy is 0.343750, time 198.405 samples/sec
+epoch 15, total_step 244480, total loss is 16.11 , inference loss is 9.74, weight deacy loss is 6.37, training accuracy is 0.312500, time 223.427 samples/sec
+epoch 15, total_step 244500, total loss is 13.33 , inference loss is 6.96, weight deacy loss is 6.37, training accuracy is 0.437500, time 10.855 samples/sec
+epoch 15, total_step 244520, total loss is 12.45 , inference loss is 6.08, weight deacy loss is 6.37, training accuracy is 0.437500, time 214.085 samples/sec
+epoch 15, total_step 244540, total loss is 18.68 , inference loss is 12.31, weight deacy loss is 6.37, training accuracy is 0.406250, time 208.196 samples/sec
+epoch 15, total_step 244560, total loss is 11.30 , inference loss is 4.93, weight deacy loss is 6.37, training accuracy is 0.468750, time 212.419 samples/sec
+epoch 15, total_step 244580, total loss is 13.52 , inference loss is 7.15, weight deacy loss is 6.37, training accuracy is 0.406250, time 202.288 samples/sec
+epoch 15, total_step 244600, total loss is 15.97 , inference loss is 9.60, weight deacy loss is 6.37, training accuracy is 0.500000, time 199.561 samples/sec
+epoch 15, total_step 244620, total loss is 14.97 , inference loss is 8.60, weight deacy loss is 6.37, training accuracy is 0.437500, time 219.019 samples/sec
+epoch 15, total_step 244640, total loss is 16.87 , inference loss is 10.50, weight deacy loss is 6.37, training accuracy is 0.281250, time 216.155 samples/sec
+epoch 15, total_step 244660, total loss is 14.59 , inference loss is 8.22, weight deacy loss is 6.37, training accuracy is 0.375000, time 215.947 samples/sec
+epoch 15, total_step 244680, total loss is 10.00 , inference loss is 3.63, weight deacy loss is 6.37, training accuracy is 0.468750, time 220.962 samples/sec
+epoch 15, total_step 244700, total loss is 10.58 , inference loss is 4.21, weight deacy loss is 6.37, training accuracy is 0.562500, time 225.391 samples/sec
+epoch 15, total_step 244720, total loss is 11.76 , inference loss is 5.39, weight deacy loss is 6.37, training accuracy is 0.343750, time 237.316 samples/sec
+epoch 15, total_step 244740, total loss is 17.62 , inference loss is 11.25, weight deacy loss is 6.37, training accuracy is 0.312500, time 299.655 samples/sec
+epoch 15, total_step 244760, total loss is 16.06 , inference loss is 9.69, weight deacy loss is 6.37, training accuracy is 0.343750, time 226.205 samples/sec
+epoch 15, total_step 244780, total loss is 14.34 , inference loss is 7.97, weight deacy loss is 6.37, training accuracy is 0.406250, time 210.743 samples/sec
+epoch 15, total_step 244800, total loss is 14.45 , inference loss is 8.08, weight deacy loss is 6.37, training accuracy is 0.312500, time 13.250 samples/sec
+epoch 15, total_step 244820, total loss is 13.68 , inference loss is 7.32, weight deacy loss is 6.37, training accuracy is 0.468750, time 214.665 samples/sec
+epoch 15, total_step 244840, total loss is 13.68 , inference loss is 7.31, weight deacy loss is 6.37, training accuracy is 0.343750, time 210.200 samples/sec
+epoch 15, total_step 244860, total loss is 15.41 , inference loss is 9.04, weight deacy loss is 6.37, training accuracy is 0.437500, time 219.349 samples/sec
+epoch 15, total_step 244880, total loss is 10.97 , inference loss is 4.60, weight deacy loss is 6.37, training accuracy is 0.562500, time 213.221 samples/sec
+epoch 15, total_step 244900, total loss is 15.77 , inference loss is 9.40, weight deacy loss is 6.37, training accuracy is 0.218750, time 235.134 samples/sec
+epoch 15, total_step 244920, total loss is 16.21 , inference loss is 9.84, weight deacy loss is 6.37, training accuracy is 0.312500, time 215.408 samples/sec
+epoch 15, total_step 244940, total loss is 12.30 , inference loss is 5.93, weight deacy loss is 6.37, training accuracy is 0.375000, time 221.896 samples/sec
+epoch 15, total_step 244960, total loss is 17.24 , inference loss is 10.87, weight deacy loss is 6.37, training accuracy is 0.281250, time 218.775 samples/sec
+epoch 15, total_step 244980, total loss is 13.97 , inference loss is 7.60, weight deacy loss is 6.37, training accuracy is 0.375000, time 211.503 samples/sec
+epoch 15, total_step 245000, total loss is 15.68 , inference loss is 9.31, weight deacy loss is 6.37, training accuracy is 0.437500, time 211.626 samples/sec
+epoch 15, total_step 245020, total loss is 16.75 , inference loss is 10.38, weight deacy loss is 6.37, training accuracy is 0.281250, time 296.817 samples/sec
+epoch 15, total_step 245040, total loss is 11.64 , inference loss is 5.27, weight deacy loss is 6.37, training accuracy is 0.406250, time 193.597 samples/sec
+epoch 15, total_step 245060, total loss is 15.82 , inference loss is 9.45, weight deacy loss is 6.37, training accuracy is 0.250000, time 225.909 samples/sec
+epoch 15, total_step 245080, total loss is 10.27 , inference loss is 3.90, weight deacy loss is 6.37, training accuracy is 0.437500, time 193.147 samples/sec
+epoch 15, total_step 245100, total loss is 12.22 , inference loss is 5.85, weight deacy loss is 6.37, training accuracy is 0.437500, time 12.977 samples/sec
+epoch 15, total_step 245120, total loss is 12.94 , inference loss is 6.57, weight deacy loss is 6.37, training accuracy is 0.406250, time 201.107 samples/sec
+epoch 15, total_step 245140, total loss is 11.38 , inference loss is 5.01, weight deacy loss is 6.37, training accuracy is 0.562500, time 175.544 samples/sec
+epoch 15, total_step 245160, total loss is 13.19 , inference loss is 6.82, weight deacy loss is 6.37, training accuracy is 0.437500, time 200.213 samples/sec
+epoch 15, total_step 245180, total loss is 14.88 , inference loss is 8.51, weight deacy loss is 6.37, training accuracy is 0.375000, time 198.433 samples/sec
+epoch 15, total_step 245200, total loss is 15.06 , inference loss is 8.69, weight deacy loss is 6.37, training accuracy is 0.250000, time 202.932 samples/sec
+epoch 15, total_step 245220, total loss is 12.83 , inference loss is 6.46, weight deacy loss is 6.37, training accuracy is 0.312500, time 179.370 samples/sec
+epoch 15, total_step 245240, total loss is 15.50 , inference loss is 9.13, weight deacy loss is 6.37, training accuracy is 0.406250, time 202.106 samples/sec
+epoch 15, total_step 245260, total loss is 15.20 , inference loss is 8.83, weight deacy loss is 6.37, training accuracy is 0.531250, time 218.858 samples/sec
+epoch 15, total_step 245280, total loss is 11.82 , inference loss is 5.45, weight deacy loss is 6.37, training accuracy is 0.625000, time 192.609 samples/sec
+End of epoch 15
+epoch 16, total_step 245300, total loss is 19.98 , inference loss is 13.62, weight deacy loss is 6.37, training accuracy is 0.093750, time 218.770 samples/sec
+epoch 16, total_step 245320, total loss is 12.23 , inference loss is 5.86, weight deacy loss is 6.37, training accuracy is 0.593750, time 214.747 samples/sec
+epoch 16, total_step 245340, total loss is 13.35 , inference loss is 6.99, weight deacy loss is 6.37, training accuracy is 0.656250, time 214.087 samples/sec
+epoch 16, total_step 245360, total loss is 9.28 , inference loss is 2.91, weight deacy loss is 6.37, training accuracy is 0.812500, time 225.427 samples/sec
+epoch 16, total_step 245380, total loss is 9.67 , inference loss is 3.30, weight deacy loss is 6.37, training accuracy is 0.875000, time 214.978 samples/sec
+epoch 16, total_step 245400, total loss is 11.56 , inference loss is 5.19, weight deacy loss is 6.37, training accuracy is 0.718750, time 13.568 samples/sec
+epoch 16, total_step 245420, total loss is 10.01 , inference loss is 3.64, weight deacy loss is 6.37, training accuracy is 0.812500, time 214.327 samples/sec
+epoch 16, total_step 245440, total loss is 10.20 , inference loss is 3.84, weight deacy loss is 6.37, training accuracy is 0.843750, time 225.385 samples/sec
+epoch 16, total_step 245460, total loss is 8.37 , inference loss is 2.00, weight deacy loss is 6.37, training accuracy is 0.812500, time 218.193 samples/sec
+epoch 16, total_step 245480, total loss is 9.53 , inference loss is 3.17, weight deacy loss is 6.37, training accuracy is 0.812500, time 205.157 samples/sec
+epoch 16, total_step 245500, total loss is 11.79 , inference loss is 5.42, weight deacy loss is 6.37, training accuracy is 0.687500, time 216.138 samples/sec
+epoch 16, total_step 245520, total loss is 9.48 , inference loss is 3.12, weight deacy loss is 6.37, training accuracy is 0.718750, time 298.128 samples/sec
+epoch 16, total_step 245540, total loss is 9.81 , inference loss is 3.45, weight deacy loss is 6.37, training accuracy is 0.687500, time 298.037 samples/sec
+epoch 16, total_step 245560, total loss is 8.90 , inference loss is 2.53, weight deacy loss is 6.37, training accuracy is 0.750000, time 215.577 samples/sec
+epoch 16, total_step 245580, total loss is 13.92 , inference loss is 7.56, weight deacy loss is 6.37, training accuracy is 0.531250, time 213.096 samples/sec
+epoch 16, total_step 245600, total loss is 11.38 , inference loss is 5.01, weight deacy loss is 6.37, training accuracy is 0.593750, time 217.479 samples/sec
+epoch 16, total_step 245620, total loss is 8.96 , inference loss is 2.59, weight deacy loss is 6.37, training accuracy is 0.781250, time 214.949 samples/sec
+epoch 16, total_step 245640, total loss is 11.70 , inference loss is 5.33, weight deacy loss is 6.37, training accuracy is 0.781250, time 216.774 samples/sec
+epoch 16, total_step 245660, total loss is 7.52 , inference loss is 1.15, weight deacy loss is 6.37, training accuracy is 0.875000, time 213.471 samples/sec
+epoch 16, total_step 245680, total loss is 10.32 , inference loss is 3.95, weight deacy loss is 6.37, training accuracy is 0.718750, time 210.970 samples/sec
+epoch 16, total_step 245700, total loss is 9.75 , inference loss is 3.38, weight deacy loss is 6.37, training accuracy is 0.625000, time 13.597 samples/sec
+epoch 16, total_step 245720, total loss is 11.37 , inference loss is 5.00, weight deacy loss is 6.37, training accuracy is 0.656250, time 215.752 samples/sec
+epoch 16, total_step 245740, total loss is 9.99 , inference loss is 3.62, weight deacy loss is 6.37, training accuracy is 0.875000, time 217.511 samples/sec
+epoch 16, total_step 245760, total loss is 11.96 , inference loss is 5.59, weight deacy loss is 6.37, training accuracy is 0.593750, time 229.309 samples/sec
+epoch 16, total_step 245780, total loss is 11.25 , inference loss is 4.88, weight deacy loss is 6.37, training accuracy is 0.656250, time 216.179 samples/sec
+epoch 16, total_step 245800, total loss is 13.90 , inference loss is 7.53, weight deacy loss is 6.37, training accuracy is 0.656250, time 224.548 samples/sec
+epoch 16, total_step 245820, total loss is 12.15 , inference loss is 5.78, weight deacy loss is 6.37, training accuracy is 0.625000, time 213.905 samples/sec
+epoch 16, total_step 245840, total loss is 8.60 , inference loss is 2.23, weight deacy loss is 6.37, training accuracy is 0.781250, time 226.903 samples/sec
+epoch 16, total_step 245860, total loss is 11.30 , inference loss is 4.94, weight deacy loss is 6.37, training accuracy is 0.656250, time 223.536 samples/sec
+epoch 16, total_step 245880, total loss is 8.57 , inference loss is 2.20, weight deacy loss is 6.37, training accuracy is 0.781250, time 232.095 samples/sec
+epoch 16, total_step 245900, total loss is 14.59 , inference loss is 8.22, weight deacy loss is 6.37, training accuracy is 0.500000, time 218.629 samples/sec
+epoch 16, total_step 245920, total loss is 11.11 , inference loss is 4.75, weight deacy loss is 6.37, training accuracy is 0.625000, time 213.784 samples/sec
+epoch 16, total_step 245940, total loss is 9.10 , inference loss is 2.73, weight deacy loss is 6.37, training accuracy is 0.625000, time 219.661 samples/sec
+epoch 16, total_step 245960, total loss is 11.55 , inference loss is 5.18, weight deacy loss is 6.37, training accuracy is 0.687500, time 225.252 samples/sec
+epoch 16, total_step 245980, total loss is 10.15 , inference loss is 3.78, weight deacy loss is 6.37, training accuracy is 0.593750, time 298.558 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.718526999999963
+best_threshold_index 115 0.9829629629629629
+best_threshold_index 112 0.9827777777777778
+best_threshold_index 115 0.982037037037037
+best_threshold_index 109 0.9837037037037037
+best_threshold_index 112 0.9818518518518519
+best_threshold_index 115 0.9818518518518519
+best_threshold_index 112 0.9827777777777778
+best_threshold_index 115 0.9837037037037037
+best_threshold_index 112 0.9814814814814815
+best_threshold_index 112 0.982037037037037
+[lfw][246000]XNorm: 18.760091
+[lfw][246000]Accuracy-Flip: 0.98017+-0.00762
+test accuracy is: 0.9801666666666666
+epoch 16, total_step 246000, total loss is 12.24 , inference loss is 5.87, weight deacy loss is 6.37, training accuracy is 0.718750, time 16.318 samples/sec
+epoch 16, total_step 246020, total loss is 12.87 , inference loss is 6.50, weight deacy loss is 6.37, training accuracy is 0.562500, time 220.655 samples/sec
+epoch 16, total_step 246040, total loss is 10.34 , inference loss is 3.97, weight deacy loss is 6.37, training accuracy is 0.687500, time 224.515 samples/sec
+epoch 16, total_step 246060, total loss is 10.69 , inference loss is 4.32, weight deacy loss is 6.37, training accuracy is 0.718750, time 204.710 samples/sec
+epoch 16, total_step 246080, total loss is 13.11 , inference loss is 6.74, weight deacy loss is 6.37, training accuracy is 0.625000, time 215.688 samples/sec
+epoch 16, total_step 246100, total loss is 12.90 , inference loss is 6.53, weight deacy loss is 6.37, training accuracy is 0.687500, time 222.557 samples/sec
+epoch 16, total_step 246120, total loss is 12.00 , inference loss is 5.63, weight deacy loss is 6.37, training accuracy is 0.687500, time 230.181 samples/sec
+epoch 16, total_step 246140, total loss is 12.15 , inference loss is 5.77, weight deacy loss is 6.37, training accuracy is 0.656250, time 215.096 samples/sec
+epoch 16, total_step 246160, total loss is 8.04 , inference loss is 1.67, weight deacy loss is 6.37, training accuracy is 0.750000, time 215.151 samples/sec
+epoch 16, total_step 246180, total loss is 13.07 , inference loss is 6.70, weight deacy loss is 6.37, training accuracy is 0.562500, time 218.708 samples/sec
+epoch 16, total_step 246200, total loss is 10.30 , inference loss is 3.93, weight deacy loss is 6.37, training accuracy is 0.718750, time 210.077 samples/sec
+epoch 16, total_step 246220, total loss is 11.64 , inference loss is 5.27, weight deacy loss is 6.37, training accuracy is 0.562500, time 221.541 samples/sec
+epoch 16, total_step 246240, total loss is 11.98 , inference loss is 5.61, weight deacy loss is 6.37, training accuracy is 0.656250, time 216.804 samples/sec
+epoch 16, total_step 246260, total loss is 9.60 , inference loss is 3.23, weight deacy loss is 6.37, training accuracy is 0.781250, time 226.696 samples/sec
+epoch 16, total_step 246280, total loss is 8.50 , inference loss is 2.12, weight deacy loss is 6.37, training accuracy is 0.781250, time 216.120 samples/sec
+epoch 16, total_step 246300, total loss is 11.03 , inference loss is 4.65, weight deacy loss is 6.38, training accuracy is 0.750000, time 12.265 samples/sec
+epoch 16, total_step 246320, total loss is 8.56 , inference loss is 2.19, weight deacy loss is 6.38, training accuracy is 0.718750, time 300.372 samples/sec
+epoch 16, total_step 246340, total loss is 10.58 , inference loss is 4.21, weight deacy loss is 6.38, training accuracy is 0.781250, time 231.951 samples/sec
+epoch 16, total_step 246360, total loss is 9.30 , inference loss is 2.92, weight deacy loss is 6.38, training accuracy is 0.656250, time 218.554 samples/sec
+epoch 16, total_step 246380, total loss is 10.44 , inference loss is 4.07, weight deacy loss is 6.38, training accuracy is 0.593750, time 215.741 samples/sec
+epoch 16, total_step 246400, total loss is 9.34 , inference loss is 2.96, weight deacy loss is 6.38, training accuracy is 0.656250, time 223.242 samples/sec
+epoch 16, total_step 246420, total loss is 10.26 , inference loss is 3.89, weight deacy loss is 6.38, training accuracy is 0.656250, time 223.940 samples/sec
+epoch 16, total_step 246440, total loss is 11.02 , inference loss is 4.64, weight deacy loss is 6.38, training accuracy is 0.625000, time 231.626 samples/sec
+epoch 16, total_step 246460, total loss is 11.83 , inference loss is 5.45, weight deacy loss is 6.38, training accuracy is 0.562500, time 225.293 samples/sec
+epoch 16, total_step 246480, total loss is 16.89 , inference loss is 10.52, weight deacy loss is 6.38, training accuracy is 0.562500, time 220.133 samples/sec
+epoch 16, total_step 246500, total loss is 12.23 , inference loss is 5.85, weight deacy loss is 6.38, training accuracy is 0.593750, time 228.118 samples/sec
+epoch 16, total_step 246520, total loss is 13.66 , inference loss is 7.28, weight deacy loss is 6.38, training accuracy is 0.500000, time 219.962 samples/sec
+epoch 16, total_step 246540, total loss is 12.63 , inference loss is 6.25, weight deacy loss is 6.38, training accuracy is 0.500000, time 224.288 samples/sec
+epoch 16, total_step 246560, total loss is 11.70 , inference loss is 5.32, weight deacy loss is 6.38, training accuracy is 0.593750, time 221.734 samples/sec
+epoch 16, total_step 246580, total loss is 12.26 , inference loss is 5.88, weight deacy loss is 6.38, training accuracy is 0.656250, time 215.684 samples/sec
+epoch 16, total_step 246600, total loss is 14.78 , inference loss is 8.40, weight deacy loss is 6.38, training accuracy is 0.562500, time 12.781 samples/sec
+epoch 16, total_step 246620, total loss is 13.41 , inference loss is 7.03, weight deacy loss is 6.38, training accuracy is 0.562500, time 222.448 samples/sec
+epoch 16, total_step 246640, total loss is 11.43 , inference loss is 5.05, weight deacy loss is 6.38, training accuracy is 0.750000, time 215.809 samples/sec
+epoch 16, total_step 246660, total loss is 9.87 , inference loss is 3.49, weight deacy loss is 6.38, training accuracy is 0.562500, time 219.925 samples/sec
+epoch 16, total_step 246680, total loss is 7.95 , inference loss is 1.56, weight deacy loss is 6.38, training accuracy is 0.718750, time 211.669 samples/sec
+epoch 16, total_step 246700, total loss is 10.69 , inference loss is 4.31, weight deacy loss is 6.38, training accuracy is 0.656250, time 218.200 samples/sec
+epoch 16, total_step 246720, total loss is 11.72 , inference loss is 5.34, weight deacy loss is 6.38, training accuracy is 0.625000, time 211.577 samples/sec
+epoch 16, total_step 246740, total loss is 15.78 , inference loss is 9.40, weight deacy loss is 6.38, training accuracy is 0.531250, time 221.878 samples/sec
+epoch 16, total_step 246760, total loss is 9.83 , inference loss is 3.45, weight deacy loss is 6.38, training accuracy is 0.656250, time 221.554 samples/sec
+epoch 16, total_step 246780, total loss is 16.68 , inference loss is 10.29, weight deacy loss is 6.38, training accuracy is 0.437500, time 228.435 samples/sec
+epoch 16, total_step 246800, total loss is 10.07 , inference loss is 3.68, weight deacy loss is 6.38, training accuracy is 0.718750, time 226.930 samples/sec
+epoch 16, total_step 246820, total loss is 12.51 , inference loss is 6.12, weight deacy loss is 6.38, training accuracy is 0.500000, time 216.415 samples/sec
+epoch 16, total_step 246840, total loss is 17.44 , inference loss is 11.05, weight deacy loss is 6.38, training accuracy is 0.468750, time 207.500 samples/sec
+epoch 16, total_step 246860, total loss is 12.89 , inference loss is 6.50, weight deacy loss is 6.38, training accuracy is 0.593750, time 228.178 samples/sec
+epoch 16, total_step 246880, total loss is 12.76 , inference loss is 6.38, weight deacy loss is 6.38, training accuracy is 0.562500, time 217.865 samples/sec
+epoch 16, total_step 246900, total loss is 14.38 , inference loss is 7.99, weight deacy loss is 6.38, training accuracy is 0.656250, time 12.531 samples/sec
+epoch 16, total_step 246920, total loss is 11.53 , inference loss is 5.15, weight deacy loss is 6.39, training accuracy is 0.562500, time 210.791 samples/sec
+epoch 16, total_step 246940, total loss is 15.10 , inference loss is 8.72, weight deacy loss is 6.39, training accuracy is 0.562500, time 214.664 samples/sec
+epoch 16, total_step 246960, total loss is 11.49 , inference loss is 5.10, weight deacy loss is 6.39, training accuracy is 0.687500, time 218.196 samples/sec
+epoch 16, total_step 246980, total loss is 16.47 , inference loss is 10.08, weight deacy loss is 6.39, training accuracy is 0.437500, time 218.500 samples/sec
+epoch 16, total_step 247000, total loss is 10.82 , inference loss is 4.43, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.917 samples/sec
+epoch 16, total_step 247020, total loss is 12.91 , inference loss is 6.53, weight deacy loss is 6.39, training accuracy is 0.593750, time 217.033 samples/sec
+epoch 16, total_step 247040, total loss is 11.40 , inference loss is 5.01, weight deacy loss is 6.39, training accuracy is 0.593750, time 223.293 samples/sec
+epoch 16, total_step 247060, total loss is 11.84 , inference loss is 5.45, weight deacy loss is 6.39, training accuracy is 0.687500, time 223.233 samples/sec
+epoch 16, total_step 247080, total loss is 14.24 , inference loss is 7.86, weight deacy loss is 6.39, training accuracy is 0.531250, time 218.674 samples/sec
+epoch 16, total_step 247100, total loss is 15.93 , inference loss is 9.55, weight deacy loss is 6.39, training accuracy is 0.531250, time 221.718 samples/sec
+epoch 16, total_step 247120, total loss is 12.04 , inference loss is 5.65, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.884 samples/sec
+epoch 16, total_step 247140, total loss is 9.43 , inference loss is 3.04, weight deacy loss is 6.39, training accuracy is 0.625000, time 222.682 samples/sec
+epoch 16, total_step 247160, total loss is 14.51 , inference loss is 8.12, weight deacy loss is 6.39, training accuracy is 0.593750, time 298.188 samples/sec
+epoch 16, total_step 247180, total loss is 7.98 , inference loss is 1.60, weight deacy loss is 6.39, training accuracy is 0.843750, time 216.162 samples/sec
+epoch 16, total_step 247200, total loss is 9.40 , inference loss is 3.01, weight deacy loss is 6.39, training accuracy is 0.656250, time 13.009 samples/sec
+epoch 16, total_step 247220, total loss is 13.30 , inference loss is 6.91, weight deacy loss is 6.39, training accuracy is 0.593750, time 218.483 samples/sec
+epoch 16, total_step 247240, total loss is 12.68 , inference loss is 6.29, weight deacy loss is 6.39, training accuracy is 0.500000, time 219.016 samples/sec
+epoch 16, total_step 247260, total loss is 11.23 , inference loss is 4.84, weight deacy loss is 6.39, training accuracy is 0.625000, time 213.569 samples/sec
+epoch 16, total_step 247280, total loss is 10.08 , inference loss is 3.69, weight deacy loss is 6.39, training accuracy is 0.843750, time 214.100 samples/sec
+epoch 16, total_step 247300, total loss is 10.14 , inference loss is 3.76, weight deacy loss is 6.39, training accuracy is 0.718750, time 219.466 samples/sec
+epoch 16, total_step 247320, total loss is 8.41 , inference loss is 2.02, weight deacy loss is 6.39, training accuracy is 0.718750, time 216.097 samples/sec
+epoch 16, total_step 247340, total loss is 11.46 , inference loss is 5.07, weight deacy loss is 6.39, training accuracy is 0.625000, time 214.675 samples/sec
+epoch 16, total_step 247360, total loss is 10.03 , inference loss is 3.65, weight deacy loss is 6.39, training accuracy is 0.687500, time 217.587 samples/sec
+epoch 16, total_step 247380, total loss is 10.71 , inference loss is 4.32, weight deacy loss is 6.39, training accuracy is 0.593750, time 223.823 samples/sec
+epoch 16, total_step 247400, total loss is 9.46 , inference loss is 3.07, weight deacy loss is 6.39, training accuracy is 0.625000, time 216.596 samples/sec
+epoch 16, total_step 247420, total loss is 8.80 , inference loss is 2.41, weight deacy loss is 6.39, training accuracy is 0.718750, time 213.993 samples/sec
+epoch 16, total_step 247440, total loss is 8.40 , inference loss is 2.02, weight deacy loss is 6.39, training accuracy is 0.750000, time 217.094 samples/sec
+epoch 16, total_step 247460, total loss is 9.90 , inference loss is 3.51, weight deacy loss is 6.39, training accuracy is 0.687500, time 299.317 samples/sec
+epoch 16, total_step 247480, total loss is 8.46 , inference loss is 2.08, weight deacy loss is 6.39, training accuracy is 0.750000, time 298.680 samples/sec
+epoch 16, total_step 247500, total loss is 8.89 , inference loss is 2.50, weight deacy loss is 6.39, training accuracy is 0.718750, time 13.816 samples/sec
+epoch 16, total_step 247520, total loss is 14.28 , inference loss is 7.90, weight deacy loss is 6.39, training accuracy is 0.625000, time 214.535 samples/sec
+epoch 16, total_step 247540, total loss is 10.21 , inference loss is 3.82, weight deacy loss is 6.39, training accuracy is 0.656250, time 295.319 samples/sec
+epoch 16, total_step 247560, total loss is 10.99 , inference loss is 4.60, weight deacy loss is 6.39, training accuracy is 0.625000, time 216.298 samples/sec
+epoch 16, total_step 247580, total loss is 8.94 , inference loss is 2.55, weight deacy loss is 6.39, training accuracy is 0.687500, time 213.438 samples/sec
+epoch 16, total_step 247600, total loss is 14.92 , inference loss is 8.53, weight deacy loss is 6.39, training accuracy is 0.468750, time 227.382 samples/sec
+epoch 16, total_step 247620, total loss is 9.77 , inference loss is 3.38, weight deacy loss is 6.39, training accuracy is 0.625000, time 223.060 samples/sec
+epoch 16, total_step 247640, total loss is 11.29 , inference loss is 4.90, weight deacy loss is 6.39, training accuracy is 0.687500, time 227.528 samples/sec
+epoch 16, total_step 247660, total loss is 13.07 , inference loss is 6.68, weight deacy loss is 6.39, training accuracy is 0.625000, time 226.011 samples/sec
+epoch 16, total_step 247680, total loss is 7.99 , inference loss is 1.60, weight deacy loss is 6.39, training accuracy is 0.750000, time 220.067 samples/sec
+epoch 16, total_step 247700, total loss is 9.66 , inference loss is 3.28, weight deacy loss is 6.39, training accuracy is 0.750000, time 214.002 samples/sec
+epoch 16, total_step 247720, total loss is 12.08 , inference loss is 5.69, weight deacy loss is 6.39, training accuracy is 0.656250, time 211.126 samples/sec
+epoch 16, total_step 247740, total loss is 10.99 , inference loss is 4.60, weight deacy loss is 6.39, training accuracy is 0.812500, time 213.424 samples/sec
+epoch 16, total_step 247760, total loss is 11.06 , inference loss is 4.67, weight deacy loss is 6.39, training accuracy is 0.781250, time 214.043 samples/sec
+epoch 16, total_step 247780, total loss is 8.72 , inference loss is 2.34, weight deacy loss is 6.39, training accuracy is 0.687500, time 231.028 samples/sec
+epoch 16, total_step 247800, total loss is 9.86 , inference loss is 3.47, weight deacy loss is 6.39, training accuracy is 0.656250, time 13.362 samples/sec
+epoch 16, total_step 247820, total loss is 12.01 , inference loss is 5.62, weight deacy loss is 6.39, training accuracy is 0.687500, time 201.957 samples/sec
+epoch 16, total_step 247840, total loss is 10.62 , inference loss is 4.23, weight deacy loss is 6.39, training accuracy is 0.718750, time 221.639 samples/sec
+epoch 16, total_step 247860, total loss is 8.68 , inference loss is 2.29, weight deacy loss is 6.39, training accuracy is 0.781250, time 224.690 samples/sec
+epoch 16, total_step 247880, total loss is 9.46 , inference loss is 3.07, weight deacy loss is 6.39, training accuracy is 0.625000, time 216.679 samples/sec
+epoch 16, total_step 247900, total loss is 8.46 , inference loss is 2.07, weight deacy loss is 6.39, training accuracy is 0.750000, time 223.511 samples/sec
+epoch 16, total_step 247920, total loss is 9.44 , inference loss is 3.06, weight deacy loss is 6.39, training accuracy is 0.687500, time 230.785 samples/sec
+epoch 16, total_step 247940, total loss is 9.54 , inference loss is 3.15, weight deacy loss is 6.39, training accuracy is 0.562500, time 221.830 samples/sec
+epoch 16, total_step 247960, total loss is 14.35 , inference loss is 7.96, weight deacy loss is 6.39, training accuracy is 0.531250, time 218.232 samples/sec
+epoch 16, total_step 247980, total loss is 12.86 , inference loss is 6.48, weight deacy loss is 6.39, training accuracy is 0.656250, time 217.710 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.705411999999995
+best_threshold_index 121 0.9868518518518519
+best_threshold_index 121 0.9855555555555555
+best_threshold_index 123 0.9855555555555555
+best_threshold_index 121 0.9861111111111112
+best_threshold_index 123 0.9861111111111112
+best_threshold_index 123 0.9851851851851852
+best_threshold_index 121 0.9851851851851852
+best_threshold_index 121 0.985
+best_threshold_index 121 0.9848148148148148
+best_threshold_index 121 0.9851851851851852
+[lfw][248000]XNorm: 18.430484
+[lfw][248000]Accuracy-Flip: 0.98500+-0.00537
+test accuracy is: 0.985
+epoch 16, total_step 248000, total loss is 12.96 , inference loss is 6.58, weight deacy loss is 6.39, training accuracy is 0.750000, time 16.593 samples/sec
+epoch 16, total_step 248020, total loss is 10.54 , inference loss is 4.15, weight deacy loss is 6.39, training accuracy is 0.656250, time 226.317 samples/sec
+epoch 16, total_step 248040, total loss is 10.59 , inference loss is 4.21, weight deacy loss is 6.39, training accuracy is 0.750000, time 221.208 samples/sec
+epoch 16, total_step 248060, total loss is 12.71 , inference loss is 6.33, weight deacy loss is 6.39, training accuracy is 0.687500, time 240.700 samples/sec
+epoch 16, total_step 248080, total loss is 9.03 , inference loss is 2.65, weight deacy loss is 6.39, training accuracy is 0.718750, time 209.612 samples/sec
+epoch 16, total_step 248100, total loss is 11.99 , inference loss is 5.60, weight deacy loss is 6.39, training accuracy is 0.625000, time 13.714 samples/sec
+epoch 16, total_step 248120, total loss is 8.18 , inference loss is 1.79, weight deacy loss is 6.39, training accuracy is 0.781250, time 218.792 samples/sec
+epoch 16, total_step 248140, total loss is 8.78 , inference loss is 2.40, weight deacy loss is 6.39, training accuracy is 0.718750, time 226.689 samples/sec
+epoch 16, total_step 248160, total loss is 11.70 , inference loss is 5.31, weight deacy loss is 6.39, training accuracy is 0.593750, time 225.538 samples/sec
+epoch 16, total_step 248180, total loss is 9.99 , inference loss is 3.60, weight deacy loss is 6.39, training accuracy is 0.656250, time 297.418 samples/sec
+epoch 16, total_step 248200, total loss is 11.43 , inference loss is 5.05, weight deacy loss is 6.39, training accuracy is 0.531250, time 221.714 samples/sec
+epoch 16, total_step 248220, total loss is 8.99 , inference loss is 2.60, weight deacy loss is 6.39, training accuracy is 0.718750, time 216.755 samples/sec
+epoch 16, total_step 248240, total loss is 10.01 , inference loss is 3.63, weight deacy loss is 6.39, training accuracy is 0.750000, time 208.904 samples/sec
+epoch 16, total_step 248260, total loss is 8.61 , inference loss is 2.22, weight deacy loss is 6.39, training accuracy is 0.750000, time 215.157 samples/sec
+epoch 16, total_step 248280, total loss is 9.73 , inference loss is 3.34, weight deacy loss is 6.39, training accuracy is 0.718750, time 212.052 samples/sec
+epoch 16, total_step 248300, total loss is 9.20 , inference loss is 2.81, weight deacy loss is 6.39, training accuracy is 0.750000, time 220.922 samples/sec
+epoch 16, total_step 248320, total loss is 11.29 , inference loss is 4.91, weight deacy loss is 6.39, training accuracy is 0.625000, time 204.636 samples/sec
+epoch 16, total_step 248340, total loss is 10.22 , inference loss is 3.84, weight deacy loss is 6.39, training accuracy is 0.593750, time 210.343 samples/sec
+epoch 16, total_step 248360, total loss is 14.20 , inference loss is 7.81, weight deacy loss is 6.39, training accuracy is 0.437500, time 223.182 samples/sec
+epoch 16, total_step 248380, total loss is 12.17 , inference loss is 5.79, weight deacy loss is 6.39, training accuracy is 0.718750, time 228.232 samples/sec
+epoch 16, total_step 248400, total loss is 9.61 , inference loss is 3.23, weight deacy loss is 6.39, training accuracy is 0.750000, time 13.459 samples/sec
+epoch 16, total_step 248420, total loss is 10.66 , inference loss is 4.27, weight deacy loss is 6.39, training accuracy is 0.718750, time 217.642 samples/sec
+epoch 16, total_step 248440, total loss is 10.64 , inference loss is 4.25, weight deacy loss is 6.39, training accuracy is 0.625000, time 297.645 samples/sec
+epoch 16, total_step 248460, total loss is 8.85 , inference loss is 2.47, weight deacy loss is 6.39, training accuracy is 0.812500, time 227.502 samples/sec
+epoch 16, total_step 248480, total loss is 17.47 , inference loss is 11.09, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.322 samples/sec
+epoch 16, total_step 248500, total loss is 13.57 , inference loss is 7.18, weight deacy loss is 6.39, training accuracy is 0.562500, time 221.603 samples/sec
+epoch 16, total_step 248520, total loss is 12.40 , inference loss is 6.02, weight deacy loss is 6.39, training accuracy is 0.687500, time 225.078 samples/sec
+epoch 16, total_step 248540, total loss is 11.66 , inference loss is 5.28, weight deacy loss is 6.39, training accuracy is 0.625000, time 210.246 samples/sec
+epoch 16, total_step 248560, total loss is 11.19 , inference loss is 4.80, weight deacy loss is 6.39, training accuracy is 0.656250, time 209.326 samples/sec
+epoch 16, total_step 248580, total loss is 13.79 , inference loss is 7.41, weight deacy loss is 6.39, training accuracy is 0.437500, time 223.306 samples/sec
+epoch 16, total_step 248600, total loss is 13.04 , inference loss is 6.65, weight deacy loss is 6.39, training accuracy is 0.625000, time 215.748 samples/sec
+epoch 16, total_step 248620, total loss is 12.62 , inference loss is 6.24, weight deacy loss is 6.39, training accuracy is 0.500000, time 219.647 samples/sec
+epoch 16, total_step 248640, total loss is 14.59 , inference loss is 8.21, weight deacy loss is 6.39, training accuracy is 0.468750, time 218.538 samples/sec
+epoch 16, total_step 248660, total loss is 16.13 , inference loss is 9.75, weight deacy loss is 6.39, training accuracy is 0.343750, time 211.231 samples/sec
+epoch 16, total_step 248680, total loss is 12.13 , inference loss is 5.74, weight deacy loss is 6.39, training accuracy is 0.562500, time 235.384 samples/sec
+epoch 16, total_step 248700, total loss is 14.38 , inference loss is 7.99, weight deacy loss is 6.39, training accuracy is 0.593750, time 12.262 samples/sec
+epoch 16, total_step 248720, total loss is 10.39 , inference loss is 4.00, weight deacy loss is 6.39, training accuracy is 0.656250, time 213.147 samples/sec
+epoch 16, total_step 248740, total loss is 12.11 , inference loss is 5.72, weight deacy loss is 6.39, training accuracy is 0.562500, time 230.216 samples/sec
+epoch 16, total_step 248760, total loss is 10.38 , inference loss is 3.99, weight deacy loss is 6.39, training accuracy is 0.593750, time 218.389 samples/sec
+epoch 16, total_step 248780, total loss is 15.05 , inference loss is 8.66, weight deacy loss is 6.39, training accuracy is 0.468750, time 212.381 samples/sec
+epoch 16, total_step 248800, total loss is 14.46 , inference loss is 8.08, weight deacy loss is 6.39, training accuracy is 0.500000, time 211.396 samples/sec
+epoch 16, total_step 248820, total loss is 13.44 , inference loss is 7.06, weight deacy loss is 6.39, training accuracy is 0.562500, time 213.639 samples/sec
+epoch 16, total_step 248840, total loss is 15.29 , inference loss is 8.91, weight deacy loss is 6.39, training accuracy is 0.625000, time 233.963 samples/sec
+epoch 16, total_step 248860, total loss is 10.29 , inference loss is 3.91, weight deacy loss is 6.39, training accuracy is 0.562500, time 221.595 samples/sec
+epoch 16, total_step 248880, total loss is 12.46 , inference loss is 6.07, weight deacy loss is 6.39, training accuracy is 0.656250, time 242.328 samples/sec
+epoch 16, total_step 248900, total loss is 13.86 , inference loss is 7.47, weight deacy loss is 6.39, training accuracy is 0.375000, time 209.240 samples/sec
+epoch 16, total_step 248920, total loss is 10.91 , inference loss is 4.52, weight deacy loss is 6.39, training accuracy is 0.593750, time 227.334 samples/sec
+epoch 16, total_step 248940, total loss is 9.53 , inference loss is 3.14, weight deacy loss is 6.39, training accuracy is 0.750000, time 221.105 samples/sec
+epoch 16, total_step 248960, total loss is 17.48 , inference loss is 11.09, weight deacy loss is 6.39, training accuracy is 0.375000, time 219.749 samples/sec
+epoch 16, total_step 248980, total loss is 11.58 , inference loss is 5.19, weight deacy loss is 6.39, training accuracy is 0.625000, time 214.573 samples/sec
+epoch 16, total_step 249000, total loss is 12.77 , inference loss is 6.38, weight deacy loss is 6.39, training accuracy is 0.593750, time 12.599 samples/sec
+epoch 16, total_step 249020, total loss is 10.70 , inference loss is 4.31, weight deacy loss is 6.39, training accuracy is 0.687500, time 211.895 samples/sec
+epoch 16, total_step 249040, total loss is 11.24 , inference loss is 4.86, weight deacy loss is 6.39, training accuracy is 0.593750, time 219.440 samples/sec
+epoch 16, total_step 249060, total loss is 14.76 , inference loss is 8.37, weight deacy loss is 6.39, training accuracy is 0.625000, time 212.396 samples/sec
+epoch 16, total_step 249080, total loss is 12.29 , inference loss is 5.90, weight deacy loss is 6.39, training accuracy is 0.593750, time 218.902 samples/sec
+epoch 16, total_step 249100, total loss is 14.05 , inference loss is 7.66, weight deacy loss is 6.39, training accuracy is 0.468750, time 214.811 samples/sec
+epoch 16, total_step 249120, total loss is 18.50 , inference loss is 12.11, weight deacy loss is 6.39, training accuracy is 0.437500, time 217.662 samples/sec
+epoch 16, total_step 249140, total loss is 11.64 , inference loss is 5.25, weight deacy loss is 6.39, training accuracy is 0.593750, time 227.220 samples/sec
+epoch 16, total_step 249160, total loss is 13.22 , inference loss is 6.83, weight deacy loss is 6.39, training accuracy is 0.468750, time 217.134 samples/sec
+epoch 16, total_step 249180, total loss is 9.25 , inference loss is 2.86, weight deacy loss is 6.39, training accuracy is 0.843750, time 213.962 samples/sec
+epoch 16, total_step 249200, total loss is 13.50 , inference loss is 7.11, weight deacy loss is 6.39, training accuracy is 0.468750, time 212.078 samples/sec
+epoch 16, total_step 249220, total loss is 15.60 , inference loss is 9.22, weight deacy loss is 6.39, training accuracy is 0.625000, time 225.890 samples/sec
+epoch 16, total_step 249240, total loss is 11.84 , inference loss is 5.46, weight deacy loss is 6.39, training accuracy is 0.593750, time 299.467 samples/sec
+epoch 16, total_step 249260, total loss is 17.23 , inference loss is 10.84, weight deacy loss is 6.39, training accuracy is 0.375000, time 299.281 samples/sec
+epoch 16, total_step 249280, total loss is 14.67 , inference loss is 8.28, weight deacy loss is 6.39, training accuracy is 0.468750, time 210.364 samples/sec
+epoch 16, total_step 249300, total loss is 11.92 , inference loss is 5.53, weight deacy loss is 6.39, training accuracy is 0.593750, time 12.441 samples/sec
+epoch 16, total_step 249320, total loss is 12.21 , inference loss is 5.82, weight deacy loss is 6.39, training accuracy is 0.562500, time 218.435 samples/sec
+epoch 16, total_step 249340, total loss is 12.76 , inference loss is 6.37, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.793 samples/sec
+epoch 16, total_step 249360, total loss is 12.30 , inference loss is 5.91, weight deacy loss is 6.39, training accuracy is 0.562500, time 210.855 samples/sec
+epoch 16, total_step 249380, total loss is 12.54 , inference loss is 6.15, weight deacy loss is 6.39, training accuracy is 0.531250, time 218.630 samples/sec
+epoch 16, total_step 249400, total loss is 12.73 , inference loss is 6.34, weight deacy loss is 6.39, training accuracy is 0.562500, time 224.186 samples/sec
+epoch 16, total_step 249420, total loss is 14.31 , inference loss is 7.92, weight deacy loss is 6.39, training accuracy is 0.500000, time 299.755 samples/sec
+epoch 16, total_step 249440, total loss is 12.27 , inference loss is 5.88, weight deacy loss is 6.39, training accuracy is 0.593750, time 215.817 samples/sec
+epoch 16, total_step 249460, total loss is 10.41 , inference loss is 4.02, weight deacy loss is 6.39, training accuracy is 0.718750, time 213.114 samples/sec
+epoch 16, total_step 249480, total loss is 15.63 , inference loss is 9.24, weight deacy loss is 6.39, training accuracy is 0.406250, time 212.730 samples/sec
+epoch 16, total_step 249500, total loss is 18.85 , inference loss is 12.47, weight deacy loss is 6.39, training accuracy is 0.343750, time 215.713 samples/sec
+epoch 16, total_step 249520, total loss is 13.02 , inference loss is 6.63, weight deacy loss is 6.39, training accuracy is 0.531250, time 233.621 samples/sec
+epoch 16, total_step 249540, total loss is 11.27 , inference loss is 4.88, weight deacy loss is 6.39, training accuracy is 0.468750, time 223.020 samples/sec
+epoch 16, total_step 249560, total loss is 12.03 , inference loss is 5.65, weight deacy loss is 6.39, training accuracy is 0.625000, time 217.347 samples/sec
+epoch 16, total_step 249580, total loss is 11.05 , inference loss is 4.67, weight deacy loss is 6.39, training accuracy is 0.437500, time 221.246 samples/sec
+epoch 16, total_step 249600, total loss is 13.48 , inference loss is 7.09, weight deacy loss is 6.39, training accuracy is 0.437500, time 12.710 samples/sec
+epoch 16, total_step 249620, total loss is 12.01 , inference loss is 5.62, weight deacy loss is 6.39, training accuracy is 0.468750, time 213.735 samples/sec
+epoch 16, total_step 249640, total loss is 13.84 , inference loss is 7.45, weight deacy loss is 6.39, training accuracy is 0.625000, time 212.539 samples/sec
+epoch 16, total_step 249660, total loss is 11.85 , inference loss is 5.46, weight deacy loss is 6.39, training accuracy is 0.625000, time 228.964 samples/sec
+epoch 16, total_step 249680, total loss is 12.65 , inference loss is 6.26, weight deacy loss is 6.39, training accuracy is 0.593750, time 296.858 samples/sec
+epoch 16, total_step 249700, total loss is 12.60 , inference loss is 6.21, weight deacy loss is 6.39, training accuracy is 0.656250, time 212.679 samples/sec
+epoch 16, total_step 249720, total loss is 16.58 , inference loss is 10.20, weight deacy loss is 6.39, training accuracy is 0.312500, time 225.667 samples/sec
+epoch 16, total_step 249740, total loss is 13.42 , inference loss is 7.03, weight deacy loss is 6.39, training accuracy is 0.593750, time 218.190 samples/sec
+epoch 16, total_step 249760, total loss is 15.55 , inference loss is 9.17, weight deacy loss is 6.39, training accuracy is 0.531250, time 210.870 samples/sec
+epoch 16, total_step 249780, total loss is 12.18 , inference loss is 5.79, weight deacy loss is 6.39, training accuracy is 0.562500, time 216.374 samples/sec
+epoch 16, total_step 249800, total loss is 11.58 , inference loss is 5.19, weight deacy loss is 6.39, training accuracy is 0.593750, time 213.026 samples/sec
+epoch 16, total_step 249820, total loss is 10.46 , inference loss is 4.07, weight deacy loss is 6.39, training accuracy is 0.562500, time 227.150 samples/sec
+epoch 16, total_step 249840, total loss is 13.59 , inference loss is 7.20, weight deacy loss is 6.39, training accuracy is 0.375000, time 220.952 samples/sec
+epoch 16, total_step 249860, total loss is 12.92 , inference loss is 6.53, weight deacy loss is 6.39, training accuracy is 0.531250, time 215.127 samples/sec
+epoch 16, total_step 249880, total loss is 10.56 , inference loss is 4.17, weight deacy loss is 6.39, training accuracy is 0.625000, time 214.382 samples/sec
+epoch 16, total_step 249900, total loss is 13.29 , inference loss is 6.90, weight deacy loss is 6.39, training accuracy is 0.500000, time 12.833 samples/sec
+epoch 16, total_step 249920, total loss is 11.15 , inference loss is 4.76, weight deacy loss is 6.39, training accuracy is 0.593750, time 221.613 samples/sec
+epoch 16, total_step 249940, total loss is 14.64 , inference loss is 8.25, weight deacy loss is 6.39, training accuracy is 0.406250, time 295.632 samples/sec
+epoch 16, total_step 249960, total loss is 12.34 , inference loss is 5.95, weight deacy loss is 6.39, training accuracy is 0.531250, time 213.469 samples/sec
+epoch 16, total_step 249980, total loss is 15.68 , inference loss is 9.29, weight deacy loss is 6.39, training accuracy is 0.375000, time 211.780 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.864779
+best_threshold_index 122 0.9877777777777778
+best_threshold_index 123 0.9874074074074074
+best_threshold_index 123 0.9885185185185185
+best_threshold_index 123 0.9875925925925926
+best_threshold_index 123 0.9872222222222222
+best_threshold_index 122 0.9868518518518519
+best_threshold_index 123 0.9877777777777778
+best_threshold_index 123 0.9868518518518519
+best_threshold_index 126 0.9868518518518519
+best_threshold_index 123 0.9866666666666667
+[lfw][250000]XNorm: 22.257239
+[lfw][250000]Accuracy-Flip: 0.98667+-0.00483
+test accuracy is: 0.9866666666666667
+epoch 16, total_step 250000, total loss is 9.73 , inference loss is 3.34, weight deacy loss is 6.39, training accuracy is 0.718750, time 16.748 samples/sec
+epoch 16, total_step 250020, total loss is 11.21 , inference loss is 4.82, weight deacy loss is 6.39, training accuracy is 0.500000, time 212.758 samples/sec
+epoch 16, total_step 250040, total loss is 13.61 , inference loss is 7.22, weight deacy loss is 6.39, training accuracy is 0.468750, time 242.029 samples/sec
+epoch 16, total_step 250060, total loss is 11.16 , inference loss is 4.77, weight deacy loss is 6.39, training accuracy is 0.562500, time 221.125 samples/sec
+epoch 16, total_step 250080, total loss is 11.05 , inference loss is 4.66, weight deacy loss is 6.39, training accuracy is 0.593750, time 224.384 samples/sec
+epoch 16, total_step 250100, total loss is 10.06 , inference loss is 3.67, weight deacy loss is 6.39, training accuracy is 0.687500, time 211.361 samples/sec
+epoch 16, total_step 250120, total loss is 10.04 , inference loss is 3.65, weight deacy loss is 6.39, training accuracy is 0.593750, time 222.548 samples/sec
+epoch 16, total_step 250140, total loss is 15.85 , inference loss is 9.46, weight deacy loss is 6.39, training accuracy is 0.593750, time 208.875 samples/sec
+epoch 16, total_step 250160, total loss is 12.44 , inference loss is 6.05, weight deacy loss is 6.39, training accuracy is 0.562500, time 225.102 samples/sec
+epoch 16, total_step 250180, total loss is 10.57 , inference loss is 4.18, weight deacy loss is 6.39, training accuracy is 0.625000, time 218.698 samples/sec
+epoch 16, total_step 250200, total loss is 12.63 , inference loss is 6.24, weight deacy loss is 6.39, training accuracy is 0.500000, time 12.775 samples/sec
+epoch 16, total_step 250220, total loss is 13.90 , inference loss is 7.51, weight deacy loss is 6.39, training accuracy is 0.500000, time 228.365 samples/sec
+epoch 16, total_step 250240, total loss is 10.20 , inference loss is 3.81, weight deacy loss is 6.39, training accuracy is 0.531250, time 299.092 samples/sec
+epoch 16, total_step 250260, total loss is 11.59 , inference loss is 5.20, weight deacy loss is 6.39, training accuracy is 0.468750, time 211.455 samples/sec
+epoch 16, total_step 250280, total loss is 9.88 , inference loss is 3.49, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.224 samples/sec
+epoch 16, total_step 250300, total loss is 10.47 , inference loss is 4.09, weight deacy loss is 6.39, training accuracy is 0.593750, time 221.920 samples/sec
+epoch 16, total_step 250320, total loss is 10.53 , inference loss is 4.14, weight deacy loss is 6.39, training accuracy is 0.656250, time 212.024 samples/sec
+epoch 16, total_step 250340, total loss is 10.46 , inference loss is 4.07, weight deacy loss is 6.39, training accuracy is 0.656250, time 229.030 samples/sec
+epoch 16, total_step 250360, total loss is 12.94 , inference loss is 6.55, weight deacy loss is 6.39, training accuracy is 0.625000, time 225.153 samples/sec
+epoch 16, total_step 250380, total loss is 10.07 , inference loss is 3.68, weight deacy loss is 6.39, training accuracy is 0.625000, time 209.033 samples/sec
+epoch 16, total_step 250400, total loss is 15.55 , inference loss is 9.16, weight deacy loss is 6.39, training accuracy is 0.468750, time 216.818 samples/sec
+epoch 16, total_step 250420, total loss is 10.04 , inference loss is 3.65, weight deacy loss is 6.39, training accuracy is 0.750000, time 232.442 samples/sec
+epoch 16, total_step 250440, total loss is 12.56 , inference loss is 6.17, weight deacy loss is 6.39, training accuracy is 0.656250, time 213.951 samples/sec
+epoch 16, total_step 250460, total loss is 11.50 , inference loss is 5.11, weight deacy loss is 6.39, training accuracy is 0.625000, time 216.469 samples/sec
+epoch 16, total_step 250480, total loss is 21.23 , inference loss is 14.84, weight deacy loss is 6.39, training accuracy is 0.375000, time 216.642 samples/sec
+epoch 16, total_step 250500, total loss is 14.37 , inference loss is 7.98, weight deacy loss is 6.39, training accuracy is 0.531250, time 6.951 samples/sec
+epoch 16, total_step 250520, total loss is 14.04 , inference loss is 7.65, weight deacy loss is 6.39, training accuracy is 0.656250, time 203.377 samples/sec
+epoch 16, total_step 250540, total loss is 12.30 , inference loss is 5.91, weight deacy loss is 6.39, training accuracy is 0.468750, time 206.139 samples/sec
+epoch 16, total_step 250560, total loss is 10.38 , inference loss is 3.99, weight deacy loss is 6.39, training accuracy is 0.625000, time 204.204 samples/sec
+epoch 16, total_step 250580, total loss is 10.44 , inference loss is 4.05, weight deacy loss is 6.39, training accuracy is 0.531250, time 198.411 samples/sec
+epoch 16, total_step 250600, total loss is 14.71 , inference loss is 8.32, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.034 samples/sec
+epoch 16, total_step 250620, total loss is 11.72 , inference loss is 5.33, weight deacy loss is 6.39, training accuracy is 0.562500, time 206.494 samples/sec
+epoch 16, total_step 250640, total loss is 11.83 , inference loss is 5.44, weight deacy loss is 6.39, training accuracy is 0.562500, time 222.518 samples/sec
+epoch 16, total_step 250660, total loss is 13.79 , inference loss is 7.40, weight deacy loss is 6.39, training accuracy is 0.593750, time 197.179 samples/sec
+epoch 16, total_step 250680, total loss is 11.70 , inference loss is 5.31, weight deacy loss is 6.39, training accuracy is 0.531250, time 192.541 samples/sec
+epoch 16, total_step 250700, total loss is 13.86 , inference loss is 7.47, weight deacy loss is 6.39, training accuracy is 0.500000, time 214.359 samples/sec
+epoch 16, total_step 250720, total loss is 14.32 , inference loss is 7.93, weight deacy loss is 6.39, training accuracy is 0.593750, time 200.029 samples/sec
+epoch 16, total_step 250740, total loss is 11.86 , inference loss is 5.47, weight deacy loss is 6.39, training accuracy is 0.500000, time 208.479 samples/sec
+epoch 16, total_step 250760, total loss is 11.36 , inference loss is 4.97, weight deacy loss is 6.39, training accuracy is 0.593750, time 207.683 samples/sec
+epoch 16, total_step 250780, total loss is 13.41 , inference loss is 7.02, weight deacy loss is 6.39, training accuracy is 0.593750, time 299.928 samples/sec
+epoch 16, total_step 250800, total loss is 11.43 , inference loss is 5.04, weight deacy loss is 6.39, training accuracy is 0.625000, time 8.689 samples/sec
+epoch 16, total_step 250820, total loss is 13.11 , inference loss is 6.73, weight deacy loss is 6.39, training accuracy is 0.531250, time 211.273 samples/sec
+epoch 16, total_step 250840, total loss is 14.99 , inference loss is 8.60, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.550 samples/sec
+epoch 16, total_step 250860, total loss is 13.38 , inference loss is 6.99, weight deacy loss is 6.39, training accuracy is 0.531250, time 247.571 samples/sec
+epoch 16, total_step 250880, total loss is 13.94 , inference loss is 7.55, weight deacy loss is 6.39, training accuracy is 0.562500, time 297.281 samples/sec
+epoch 16, total_step 250900, total loss is 11.91 , inference loss is 5.52, weight deacy loss is 6.39, training accuracy is 0.562500, time 299.792 samples/sec
+epoch 16, total_step 250920, total loss is 10.43 , inference loss is 4.05, weight deacy loss is 6.39, training accuracy is 0.593750, time 299.268 samples/sec
+epoch 16, total_step 250940, total loss is 10.83 , inference loss is 4.44, weight deacy loss is 6.39, training accuracy is 0.531250, time 199.960 samples/sec
+epoch 16, total_step 250960, total loss is 13.01 , inference loss is 6.62, weight deacy loss is 6.39, training accuracy is 0.562500, time 243.706 samples/sec
+epoch 16, total_step 250980, total loss is 9.65 , inference loss is 3.27, weight deacy loss is 6.39, training accuracy is 0.562500, time 212.465 samples/sec
+epoch 16, total_step 251000, total loss is 13.68 , inference loss is 7.29, weight deacy loss is 6.39, training accuracy is 0.531250, time 207.414 samples/sec
+epoch 16, total_step 251020, total loss is 9.82 , inference loss is 3.43, weight deacy loss is 6.39, training accuracy is 0.718750, time 222.548 samples/sec
+epoch 16, total_step 251040, total loss is 11.91 , inference loss is 5.52, weight deacy loss is 6.39, training accuracy is 0.500000, time 208.453 samples/sec
+epoch 16, total_step 251060, total loss is 12.90 , inference loss is 6.51, weight deacy loss is 6.39, training accuracy is 0.625000, time 204.888 samples/sec
+epoch 16, total_step 251080, total loss is 14.13 , inference loss is 7.74, weight deacy loss is 6.39, training accuracy is 0.500000, time 202.481 samples/sec
+epoch 16, total_step 251100, total loss is 13.52 , inference loss is 7.13, weight deacy loss is 6.39, training accuracy is 0.500000, time 10.576 samples/sec
+epoch 16, total_step 251120, total loss is 9.20 , inference loss is 2.81, weight deacy loss is 6.39, training accuracy is 0.687500, time 213.799 samples/sec
+epoch 16, total_step 251140, total loss is 18.03 , inference loss is 11.64, weight deacy loss is 6.39, training accuracy is 0.375000, time 297.178 samples/sec
+epoch 16, total_step 251160, total loss is 12.16 , inference loss is 5.77, weight deacy loss is 6.39, training accuracy is 0.562500, time 219.212 samples/sec
+epoch 16, total_step 251180, total loss is 8.64 , inference loss is 2.25, weight deacy loss is 6.39, training accuracy is 0.718750, time 204.636 samples/sec
+epoch 16, total_step 251200, total loss is 12.36 , inference loss is 5.97, weight deacy loss is 6.39, training accuracy is 0.562500, time 200.002 samples/sec
+epoch 16, total_step 251220, total loss is 16.57 , inference loss is 10.19, weight deacy loss is 6.39, training accuracy is 0.593750, time 195.040 samples/sec
+epoch 16, total_step 251240, total loss is 17.24 , inference loss is 10.85, weight deacy loss is 6.39, training accuracy is 0.406250, time 216.982 samples/sec
+epoch 16, total_step 251260, total loss is 11.77 , inference loss is 5.38, weight deacy loss is 6.39, training accuracy is 0.406250, time 205.316 samples/sec
+epoch 16, total_step 251280, total loss is 12.93 , inference loss is 6.54, weight deacy loss is 6.39, training accuracy is 0.468750, time 209.781 samples/sec
+epoch 16, total_step 251300, total loss is 10.16 , inference loss is 3.78, weight deacy loss is 6.39, training accuracy is 0.593750, time 215.191 samples/sec
+epoch 16, total_step 251320, total loss is 13.22 , inference loss is 6.83, weight deacy loss is 6.39, training accuracy is 0.531250, time 202.892 samples/sec
+epoch 16, total_step 251340, total loss is 14.62 , inference loss is 8.23, weight deacy loss is 6.39, training accuracy is 0.406250, time 219.890 samples/sec
+epoch 16, total_step 251360, total loss is 11.00 , inference loss is 4.61, weight deacy loss is 6.39, training accuracy is 0.500000, time 208.263 samples/sec
+epoch 16, total_step 251380, total loss is 10.62 , inference loss is 4.23, weight deacy loss is 6.39, training accuracy is 0.562500, time 207.963 samples/sec
+epoch 16, total_step 251400, total loss is 10.61 , inference loss is 4.22, weight deacy loss is 6.39, training accuracy is 0.500000, time 9.963 samples/sec
+epoch 16, total_step 251420, total loss is 12.61 , inference loss is 6.22, weight deacy loss is 6.39, training accuracy is 0.531250, time 243.283 samples/sec
+epoch 16, total_step 251440, total loss is 14.46 , inference loss is 8.07, weight deacy loss is 6.39, training accuracy is 0.562500, time 206.923 samples/sec
+epoch 16, total_step 251460, total loss is 14.76 , inference loss is 8.37, weight deacy loss is 6.39, training accuracy is 0.406250, time 227.940 samples/sec
+epoch 16, total_step 251480, total loss is 15.70 , inference loss is 9.31, weight deacy loss is 6.39, training accuracy is 0.468750, time 219.098 samples/sec
+epoch 16, total_step 251500, total loss is 10.35 , inference loss is 3.96, weight deacy loss is 6.39, training accuracy is 0.656250, time 214.757 samples/sec
+epoch 16, total_step 251520, total loss is 11.11 , inference loss is 4.72, weight deacy loss is 6.39, training accuracy is 0.593750, time 205.056 samples/sec
+epoch 16, total_step 251540, total loss is 11.66 , inference loss is 5.27, weight deacy loss is 6.39, training accuracy is 0.562500, time 196.512 samples/sec
+epoch 16, total_step 251560, total loss is 11.11 , inference loss is 4.72, weight deacy loss is 6.39, training accuracy is 0.656250, time 197.463 samples/sec
+epoch 16, total_step 251580, total loss is 10.92 , inference loss is 4.54, weight deacy loss is 6.39, training accuracy is 0.437500, time 201.555 samples/sec
+epoch 16, total_step 251600, total loss is 13.16 , inference loss is 6.77, weight deacy loss is 6.39, training accuracy is 0.656250, time 205.468 samples/sec
+epoch 16, total_step 251620, total loss is 11.17 , inference loss is 4.78, weight deacy loss is 6.39, training accuracy is 0.593750, time 206.768 samples/sec
+epoch 16, total_step 251640, total loss is 10.50 , inference loss is 4.11, weight deacy loss is 6.39, training accuracy is 0.687500, time 213.833 samples/sec
+epoch 16, total_step 251660, total loss is 11.01 , inference loss is 4.63, weight deacy loss is 6.39, training accuracy is 0.718750, time 217.000 samples/sec
+epoch 16, total_step 251680, total loss is 12.43 , inference loss is 6.04, weight deacy loss is 6.39, training accuracy is 0.468750, time 211.171 samples/sec
+epoch 16, total_step 251700, total loss is 10.60 , inference loss is 4.21, weight deacy loss is 6.39, training accuracy is 0.593750, time 9.816 samples/sec
+epoch 16, total_step 251720, total loss is 10.88 , inference loss is 4.49, weight deacy loss is 6.39, training accuracy is 0.500000, time 206.769 samples/sec
+epoch 16, total_step 251740, total loss is 10.75 , inference loss is 4.36, weight deacy loss is 6.39, training accuracy is 0.625000, time 208.731 samples/sec
+epoch 16, total_step 251760, total loss is 17.93 , inference loss is 11.54, weight deacy loss is 6.39, training accuracy is 0.406250, time 210.854 samples/sec
+epoch 16, total_step 251780, total loss is 12.91 , inference loss is 6.52, weight deacy loss is 6.39, training accuracy is 0.406250, time 207.366 samples/sec
+epoch 16, total_step 251800, total loss is 14.19 , inference loss is 7.80, weight deacy loss is 6.39, training accuracy is 0.468750, time 220.711 samples/sec
+epoch 16, total_step 251820, total loss is 11.03 , inference loss is 4.65, weight deacy loss is 6.39, training accuracy is 0.500000, time 209.498 samples/sec
+epoch 16, total_step 251840, total loss is 9.69 , inference loss is 3.30, weight deacy loss is 6.39, training accuracy is 0.625000, time 213.392 samples/sec
+epoch 16, total_step 251860, total loss is 12.99 , inference loss is 6.60, weight deacy loss is 6.39, training accuracy is 0.531250, time 216.020 samples/sec
+epoch 16, total_step 251880, total loss is 10.80 , inference loss is 4.41, weight deacy loss is 6.39, training accuracy is 0.531250, time 216.120 samples/sec
+epoch 16, total_step 251900, total loss is 14.16 , inference loss is 7.77, weight deacy loss is 6.39, training accuracy is 0.593750, time 217.667 samples/sec
+epoch 16, total_step 251920, total loss is 9.74 , inference loss is 3.36, weight deacy loss is 6.39, training accuracy is 0.500000, time 209.751 samples/sec
+epoch 16, total_step 251940, total loss is 15.39 , inference loss is 9.00, weight deacy loss is 6.39, training accuracy is 0.500000, time 224.979 samples/sec
+epoch 16, total_step 251960, total loss is 11.30 , inference loss is 4.92, weight deacy loss is 6.39, training accuracy is 0.468750, time 294.656 samples/sec
+epoch 16, total_step 251980, total loss is 13.90 , inference loss is 7.52, weight deacy loss is 6.39, training accuracy is 0.468750, time 222.101 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.156855000000014
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9875925925925926
+best_threshold_index 114 0.9868518518518519
+best_threshold_index 115 0.9885185185185185
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9866666666666667
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 115 0.9868518518518519
+best_threshold_index 115 0.9861111111111112
+best_threshold_index 115 0.9864814814814815
+[lfw][252000]XNorm: 22.372356
+[lfw][252000]Accuracy-Flip: 0.98683+-0.00598
+test accuracy is: 0.9868333333333335
+epoch 16, total_step 252000, total loss is 13.20 , inference loss is 6.81, weight deacy loss is 6.39, training accuracy is 0.468750, time 17.212 samples/sec
+epoch 16, total_step 252020, total loss is 12.21 , inference loss is 5.83, weight deacy loss is 6.39, training accuracy is 0.531250, time 298.578 samples/sec
+epoch 16, total_step 252040, total loss is 8.35 , inference loss is 1.96, weight deacy loss is 6.39, training accuracy is 0.687500, time 199.832 samples/sec
+epoch 16, total_step 252060, total loss is 10.92 , inference loss is 4.53, weight deacy loss is 6.39, training accuracy is 0.625000, time 210.483 samples/sec
+epoch 16, total_step 252080, total loss is 10.25 , inference loss is 3.86, weight deacy loss is 6.39, training accuracy is 0.687500, time 210.068 samples/sec
+epoch 16, total_step 252100, total loss is 10.76 , inference loss is 4.38, weight deacy loss is 6.39, training accuracy is 0.562500, time 216.227 samples/sec
+epoch 16, total_step 252120, total loss is 13.77 , inference loss is 7.38, weight deacy loss is 6.39, training accuracy is 0.468750, time 223.651 samples/sec
+epoch 16, total_step 252140, total loss is 11.83 , inference loss is 5.45, weight deacy loss is 6.39, training accuracy is 0.437500, time 219.920 samples/sec
+epoch 16, total_step 252160, total loss is 11.23 , inference loss is 4.84, weight deacy loss is 6.39, training accuracy is 0.656250, time 213.475 samples/sec
+epoch 16, total_step 252180, total loss is 13.82 , inference loss is 7.43, weight deacy loss is 6.39, training accuracy is 0.406250, time 223.884 samples/sec
+epoch 16, total_step 252200, total loss is 12.10 , inference loss is 5.72, weight deacy loss is 6.39, training accuracy is 0.500000, time 213.468 samples/sec
+epoch 16, total_step 252220, total loss is 15.69 , inference loss is 9.30, weight deacy loss is 6.39, training accuracy is 0.656250, time 220.743 samples/sec
+epoch 16, total_step 252240, total loss is 15.70 , inference loss is 9.31, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.707 samples/sec
+epoch 16, total_step 252260, total loss is 8.65 , inference loss is 2.26, weight deacy loss is 6.39, training accuracy is 0.625000, time 229.220 samples/sec
+epoch 16, total_step 252280, total loss is 16.87 , inference loss is 10.49, weight deacy loss is 6.39, training accuracy is 0.562500, time 217.788 samples/sec
+epoch 16, total_step 252300, total loss is 12.83 , inference loss is 6.44, weight deacy loss is 6.39, training accuracy is 0.437500, time 15.317 samples/sec
+epoch 16, total_step 252320, total loss is 13.28 , inference loss is 6.89, weight deacy loss is 6.39, training accuracy is 0.562500, time 216.127 samples/sec
+epoch 16, total_step 252340, total loss is 14.23 , inference loss is 7.85, weight deacy loss is 6.39, training accuracy is 0.406250, time 217.933 samples/sec
+epoch 16, total_step 252360, total loss is 10.41 , inference loss is 4.02, weight deacy loss is 6.39, training accuracy is 0.656250, time 210.684 samples/sec
+epoch 16, total_step 252380, total loss is 10.46 , inference loss is 4.07, weight deacy loss is 6.39, training accuracy is 0.562500, time 215.691 samples/sec
+epoch 16, total_step 252400, total loss is 12.49 , inference loss is 6.10, weight deacy loss is 6.39, training accuracy is 0.656250, time 218.555 samples/sec
+epoch 16, total_step 252420, total loss is 12.56 , inference loss is 6.17, weight deacy loss is 6.39, training accuracy is 0.531250, time 225.081 samples/sec
+epoch 16, total_step 252440, total loss is 9.36 , inference loss is 2.97, weight deacy loss is 6.39, training accuracy is 0.562500, time 213.990 samples/sec
+epoch 16, total_step 252460, total loss is 12.47 , inference loss is 6.08, weight deacy loss is 6.39, training accuracy is 0.437500, time 218.691 samples/sec
+epoch 16, total_step 252480, total loss is 12.09 , inference loss is 5.70, weight deacy loss is 6.39, training accuracy is 0.687500, time 231.584 samples/sec
+epoch 16, total_step 252500, total loss is 12.70 , inference loss is 6.31, weight deacy loss is 6.39, training accuracy is 0.656250, time 218.978 samples/sec
+epoch 16, total_step 252520, total loss is 13.79 , inference loss is 7.40, weight deacy loss is 6.39, training accuracy is 0.562500, time 297.638 samples/sec
+epoch 16, total_step 252540, total loss is 15.17 , inference loss is 8.78, weight deacy loss is 6.39, training accuracy is 0.406250, time 216.749 samples/sec
+epoch 16, total_step 252560, total loss is 11.51 , inference loss is 5.12, weight deacy loss is 6.39, training accuracy is 0.468750, time 212.466 samples/sec
+epoch 16, total_step 252580, total loss is 11.05 , inference loss is 4.66, weight deacy loss is 6.39, training accuracy is 0.500000, time 214.559 samples/sec
+epoch 16, total_step 252600, total loss is 10.61 , inference loss is 4.22, weight deacy loss is 6.39, training accuracy is 0.531250, time 13.788 samples/sec
+epoch 16, total_step 252620, total loss is 13.70 , inference loss is 7.31, weight deacy loss is 6.39, training accuracy is 0.375000, time 294.557 samples/sec
+epoch 16, total_step 252640, total loss is 13.75 , inference loss is 7.36, weight deacy loss is 6.39, training accuracy is 0.562500, time 298.083 samples/sec
+epoch 16, total_step 252660, total loss is 11.71 , inference loss is 5.32, weight deacy loss is 6.39, training accuracy is 0.593750, time 213.291 samples/sec
+epoch 16, total_step 252680, total loss is 12.65 , inference loss is 6.26, weight deacy loss is 6.39, training accuracy is 0.437500, time 221.881 samples/sec
+epoch 16, total_step 252700, total loss is 12.82 , inference loss is 6.43, weight deacy loss is 6.39, training accuracy is 0.562500, time 217.416 samples/sec
+epoch 16, total_step 252720, total loss is 12.97 , inference loss is 6.58, weight deacy loss is 6.39, training accuracy is 0.562500, time 217.717 samples/sec
+epoch 16, total_step 252740, total loss is 13.47 , inference loss is 7.09, weight deacy loss is 6.39, training accuracy is 0.437500, time 217.687 samples/sec
+epoch 16, total_step 252760, total loss is 10.62 , inference loss is 4.23, weight deacy loss is 6.39, training accuracy is 0.656250, time 294.149 samples/sec
+epoch 16, total_step 252780, total loss is 11.75 , inference loss is 5.36, weight deacy loss is 6.39, training accuracy is 0.625000, time 211.774 samples/sec
+epoch 16, total_step 252800, total loss is 14.43 , inference loss is 8.04, weight deacy loss is 6.39, training accuracy is 0.500000, time 212.049 samples/sec
+epoch 16, total_step 252820, total loss is 12.47 , inference loss is 6.08, weight deacy loss is 6.39, training accuracy is 0.500000, time 213.093 samples/sec
+epoch 16, total_step 252840, total loss is 12.06 , inference loss is 5.67, weight deacy loss is 6.39, training accuracy is 0.687500, time 209.692 samples/sec
+epoch 16, total_step 252860, total loss is 10.28 , inference loss is 3.89, weight deacy loss is 6.39, training accuracy is 0.625000, time 214.166 samples/sec
+epoch 16, total_step 252880, total loss is 13.92 , inference loss is 7.54, weight deacy loss is 6.39, training accuracy is 0.625000, time 212.964 samples/sec
+epoch 16, total_step 252900, total loss is 10.31 , inference loss is 3.92, weight deacy loss is 6.39, training accuracy is 0.593750, time 13.866 samples/sec
+epoch 16, total_step 252920, total loss is 11.37 , inference loss is 4.98, weight deacy loss is 6.39, training accuracy is 0.562500, time 216.805 samples/sec
+epoch 16, total_step 252940, total loss is 11.24 , inference loss is 4.85, weight deacy loss is 6.39, training accuracy is 0.625000, time 226.739 samples/sec
+epoch 16, total_step 252960, total loss is 12.59 , inference loss is 6.20, weight deacy loss is 6.39, training accuracy is 0.531250, time 214.618 samples/sec
+epoch 16, total_step 252980, total loss is 12.24 , inference loss is 5.85, weight deacy loss is 6.39, training accuracy is 0.593750, time 215.330 samples/sec
+epoch 16, total_step 253000, total loss is 13.14 , inference loss is 6.75, weight deacy loss is 6.39, training accuracy is 0.593750, time 214.095 samples/sec
+epoch 16, total_step 253020, total loss is 12.49 , inference loss is 6.10, weight deacy loss is 6.39, training accuracy is 0.500000, time 181.623 samples/sec
+epoch 16, total_step 253040, total loss is 14.87 , inference loss is 8.48, weight deacy loss is 6.39, training accuracy is 0.437500, time 185.286 samples/sec
+epoch 16, total_step 253060, total loss is 14.32 , inference loss is 7.93, weight deacy loss is 6.39, training accuracy is 0.500000, time 212.387 samples/sec
+epoch 16, total_step 253080, total loss is 13.57 , inference loss is 7.18, weight deacy loss is 6.39, training accuracy is 0.593750, time 213.522 samples/sec
+epoch 16, total_step 253100, total loss is 9.39 , inference loss is 3.00, weight deacy loss is 6.39, training accuracy is 0.718750, time 211.090 samples/sec
+epoch 16, total_step 253120, total loss is 9.73 , inference loss is 3.35, weight deacy loss is 6.39, training accuracy is 0.687500, time 222.873 samples/sec
+epoch 16, total_step 253140, total loss is 11.54 , inference loss is 5.15, weight deacy loss is 6.39, training accuracy is 0.562500, time 213.403 samples/sec
+epoch 16, total_step 253160, total loss is 11.10 , inference loss is 4.71, weight deacy loss is 6.39, training accuracy is 0.593750, time 217.318 samples/sec
+epoch 16, total_step 253180, total loss is 10.27 , inference loss is 3.88, weight deacy loss is 6.39, training accuracy is 0.656250, time 212.301 samples/sec
+epoch 16, total_step 253200, total loss is 12.75 , inference loss is 6.36, weight deacy loss is 6.39, training accuracy is 0.593750, time 12.435 samples/sec
+epoch 16, total_step 253220, total loss is 12.14 , inference loss is 5.75, weight deacy loss is 6.39, training accuracy is 0.531250, time 216.226 samples/sec
+epoch 16, total_step 253240, total loss is 12.88 , inference loss is 6.49, weight deacy loss is 6.39, training accuracy is 0.531250, time 230.377 samples/sec
+epoch 16, total_step 253260, total loss is 13.96 , inference loss is 7.57, weight deacy loss is 6.39, training accuracy is 0.500000, time 236.336 samples/sec
+epoch 16, total_step 253280, total loss is 13.38 , inference loss is 6.99, weight deacy loss is 6.39, training accuracy is 0.500000, time 210.204 samples/sec
+epoch 16, total_step 253300, total loss is 10.96 , inference loss is 4.57, weight deacy loss is 6.39, training accuracy is 0.687500, time 198.141 samples/sec
+epoch 16, total_step 253320, total loss is 13.10 , inference loss is 6.71, weight deacy loss is 6.39, training accuracy is 0.500000, time 220.300 samples/sec
+epoch 16, total_step 253340, total loss is 10.24 , inference loss is 3.85, weight deacy loss is 6.39, training accuracy is 0.781250, time 213.275 samples/sec
+epoch 16, total_step 253360, total loss is 9.26 , inference loss is 2.87, weight deacy loss is 6.39, training accuracy is 0.750000, time 213.594 samples/sec
+epoch 16, total_step 253380, total loss is 11.23 , inference loss is 4.84, weight deacy loss is 6.39, training accuracy is 0.562500, time 236.307 samples/sec
+epoch 16, total_step 253400, total loss is 12.74 , inference loss is 6.36, weight deacy loss is 6.39, training accuracy is 0.625000, time 214.794 samples/sec
+epoch 16, total_step 253420, total loss is 12.57 , inference loss is 6.19, weight deacy loss is 6.39, training accuracy is 0.562500, time 212.779 samples/sec
+epoch 16, total_step 253440, total loss is 15.75 , inference loss is 9.36, weight deacy loss is 6.39, training accuracy is 0.468750, time 228.567 samples/sec
+epoch 16, total_step 253460, total loss is 10.93 , inference loss is 4.54, weight deacy loss is 6.39, training accuracy is 0.468750, time 213.696 samples/sec
+epoch 16, total_step 253480, total loss is 15.23 , inference loss is 8.85, weight deacy loss is 6.39, training accuracy is 0.437500, time 215.980 samples/sec
+epoch 16, total_step 253500, total loss is 10.35 , inference loss is 3.97, weight deacy loss is 6.39, training accuracy is 0.437500, time 12.585 samples/sec
+epoch 16, total_step 253520, total loss is 11.74 , inference loss is 5.36, weight deacy loss is 6.39, training accuracy is 0.531250, time 217.511 samples/sec
+epoch 16, total_step 253540, total loss is 15.44 , inference loss is 9.06, weight deacy loss is 6.39, training accuracy is 0.500000, time 213.768 samples/sec
+epoch 16, total_step 253560, total loss is 13.37 , inference loss is 6.99, weight deacy loss is 6.39, training accuracy is 0.562500, time 297.648 samples/sec
+epoch 16, total_step 253580, total loss is 12.69 , inference loss is 6.30, weight deacy loss is 6.39, training accuracy is 0.468750, time 210.214 samples/sec
+epoch 16, total_step 253600, total loss is 11.14 , inference loss is 4.75, weight deacy loss is 6.39, training accuracy is 0.562500, time 214.604 samples/sec
+epoch 16, total_step 253620, total loss is 14.94 , inference loss is 8.55, weight deacy loss is 6.39, training accuracy is 0.562500, time 235.601 samples/sec
+epoch 16, total_step 253640, total loss is 9.45 , inference loss is 3.06, weight deacy loss is 6.39, training accuracy is 0.687500, time 218.413 samples/sec
+epoch 16, total_step 253660, total loss is 13.79 , inference loss is 7.41, weight deacy loss is 6.39, training accuracy is 0.406250, time 216.904 samples/sec
+epoch 16, total_step 253680, total loss is 18.04 , inference loss is 11.65, weight deacy loss is 6.39, training accuracy is 0.406250, time 220.557 samples/sec
+epoch 16, total_step 253700, total loss is 11.34 , inference loss is 4.96, weight deacy loss is 6.39, training accuracy is 0.437500, time 229.677 samples/sec
+epoch 16, total_step 253720, total loss is 13.64 , inference loss is 7.25, weight deacy loss is 6.39, training accuracy is 0.593750, time 224.492 samples/sec
+epoch 16, total_step 253740, total loss is 8.78 , inference loss is 2.39, weight deacy loss is 6.39, training accuracy is 0.750000, time 219.039 samples/sec
+epoch 16, total_step 253760, total loss is 11.89 , inference loss is 5.50, weight deacy loss is 6.39, training accuracy is 0.500000, time 205.127 samples/sec
+epoch 16, total_step 253780, total loss is 13.83 , inference loss is 7.44, weight deacy loss is 6.39, training accuracy is 0.500000, time 224.563 samples/sec
+epoch 16, total_step 253800, total loss is 13.57 , inference loss is 7.18, weight deacy loss is 6.39, training accuracy is 0.593750, time 12.720 samples/sec
+epoch 16, total_step 253820, total loss is 9.77 , inference loss is 3.38, weight deacy loss is 6.39, training accuracy is 0.656250, time 299.581 samples/sec
+epoch 16, total_step 253840, total loss is 12.69 , inference loss is 6.31, weight deacy loss is 6.39, training accuracy is 0.656250, time 224.834 samples/sec
+epoch 16, total_step 253860, total loss is 12.57 , inference loss is 6.18, weight deacy loss is 6.39, training accuracy is 0.500000, time 227.735 samples/sec
+epoch 16, total_step 253880, total loss is 15.79 , inference loss is 9.41, weight deacy loss is 6.39, training accuracy is 0.500000, time 212.125 samples/sec
+epoch 16, total_step 253900, total loss is 12.97 , inference loss is 6.58, weight deacy loss is 6.39, training accuracy is 0.593750, time 223.238 samples/sec
+epoch 16, total_step 253920, total loss is 12.55 , inference loss is 6.17, weight deacy loss is 6.39, training accuracy is 0.593750, time 219.842 samples/sec
+epoch 16, total_step 253940, total loss is 10.32 , inference loss is 3.93, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.968 samples/sec
+epoch 16, total_step 253960, total loss is 9.29 , inference loss is 2.90, weight deacy loss is 6.39, training accuracy is 0.656250, time 213.658 samples/sec
+epoch 16, total_step 253980, total loss is 17.18 , inference loss is 10.79, weight deacy loss is 6.39, training accuracy is 0.406250, time 226.469 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.486558000000013
+best_threshold_index 119 0.9887037037037038
+best_threshold_index 119 0.9885185185185185
+best_threshold_index 119 0.9896296296296296
+best_threshold_index 119 0.9901851851851852
+best_threshold_index 119 0.9894444444444445
+best_threshold_index 119 0.9888888888888889
+best_threshold_index 119 0.9896296296296296
+best_threshold_index 119 0.9883333333333333
+best_threshold_index 119 0.9881481481481481
+best_threshold_index 119 0.9885185185185185
+[lfw][254000]XNorm: 21.660630
+[lfw][254000]Accuracy-Flip: 0.98900+-0.00578
+test accuracy is: 0.9890000000000001
+epoch 16, total_step 254000, total loss is 14.74 , inference loss is 8.35, weight deacy loss is 6.39, training accuracy is 0.468750, time 16.756 samples/sec
+epoch 16, total_step 254020, total loss is 12.14 , inference loss is 5.75, weight deacy loss is 6.39, training accuracy is 0.593750, time 215.881 samples/sec
+epoch 16, total_step 254040, total loss is 15.26 , inference loss is 8.87, weight deacy loss is 6.39, training accuracy is 0.437500, time 222.048 samples/sec
+epoch 16, total_step 254060, total loss is 10.77 , inference loss is 4.38, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.420 samples/sec
+epoch 16, total_step 254080, total loss is 13.13 , inference loss is 6.74, weight deacy loss is 6.39, training accuracy is 0.437500, time 222.963 samples/sec
+epoch 16, total_step 254100, total loss is 15.61 , inference loss is 9.22, weight deacy loss is 6.39, training accuracy is 0.531250, time 12.780 samples/sec
+epoch 16, total_step 254120, total loss is 13.49 , inference loss is 7.10, weight deacy loss is 6.39, training accuracy is 0.562500, time 225.473 samples/sec
+epoch 16, total_step 254140, total loss is 10.51 , inference loss is 4.12, weight deacy loss is 6.39, training accuracy is 0.531250, time 231.865 samples/sec
+epoch 16, total_step 254160, total loss is 16.00 , inference loss is 9.61, weight deacy loss is 6.39, training accuracy is 0.593750, time 217.079 samples/sec
+epoch 16, total_step 254180, total loss is 9.64 , inference loss is 3.25, weight deacy loss is 6.39, training accuracy is 0.593750, time 299.633 samples/sec
+epoch 16, total_step 254200, total loss is 13.51 , inference loss is 7.12, weight deacy loss is 6.39, training accuracy is 0.500000, time 216.537 samples/sec
+epoch 16, total_step 254220, total loss is 11.70 , inference loss is 5.31, weight deacy loss is 6.39, training accuracy is 0.500000, time 218.840 samples/sec
+epoch 16, total_step 254240, total loss is 12.80 , inference loss is 6.42, weight deacy loss is 6.39, training accuracy is 0.562500, time 212.103 samples/sec
+epoch 16, total_step 254260, total loss is 11.53 , inference loss is 5.14, weight deacy loss is 6.39, training accuracy is 0.468750, time 222.806 samples/sec
+epoch 16, total_step 254280, total loss is 10.87 , inference loss is 4.48, weight deacy loss is 6.39, training accuracy is 0.562500, time 220.536 samples/sec
+epoch 16, total_step 254300, total loss is 13.91 , inference loss is 7.52, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.704 samples/sec
+epoch 16, total_step 254320, total loss is 11.95 , inference loss is 5.56, weight deacy loss is 6.39, training accuracy is 0.562500, time 216.515 samples/sec
+epoch 16, total_step 254340, total loss is 18.31 , inference loss is 11.92, weight deacy loss is 6.39, training accuracy is 0.437500, time 212.862 samples/sec
+epoch 16, total_step 254360, total loss is 13.95 , inference loss is 7.56, weight deacy loss is 6.39, training accuracy is 0.375000, time 221.269 samples/sec
+epoch 16, total_step 254380, total loss is 13.69 , inference loss is 7.30, weight deacy loss is 6.39, training accuracy is 0.406250, time 215.588 samples/sec
+epoch 16, total_step 254400, total loss is 10.87 , inference loss is 4.48, weight deacy loss is 6.39, training accuracy is 0.593750, time 13.909 samples/sec
+epoch 16, total_step 254420, total loss is 10.72 , inference loss is 4.33, weight deacy loss is 6.39, training accuracy is 0.687500, time 294.840 samples/sec
+epoch 16, total_step 254440, total loss is 12.13 , inference loss is 5.74, weight deacy loss is 6.39, training accuracy is 0.531250, time 211.736 samples/sec
+epoch 16, total_step 254460, total loss is 14.31 , inference loss is 7.92, weight deacy loss is 6.39, training accuracy is 0.468750, time 219.472 samples/sec
+epoch 16, total_step 254480, total loss is 15.57 , inference loss is 9.18, weight deacy loss is 6.39, training accuracy is 0.468750, time 215.769 samples/sec
+epoch 16, total_step 254500, total loss is 13.10 , inference loss is 6.71, weight deacy loss is 6.39, training accuracy is 0.531250, time 230.081 samples/sec
+epoch 16, total_step 254520, total loss is 9.91 , inference loss is 3.52, weight deacy loss is 6.39, training accuracy is 0.750000, time 208.245 samples/sec
+epoch 16, total_step 254540, total loss is 12.21 , inference loss is 5.82, weight deacy loss is 6.39, training accuracy is 0.531250, time 218.710 samples/sec
+epoch 16, total_step 254560, total loss is 12.66 , inference loss is 6.27, weight deacy loss is 6.39, training accuracy is 0.500000, time 216.312 samples/sec
+epoch 16, total_step 254580, total loss is 11.65 , inference loss is 5.26, weight deacy loss is 6.39, training accuracy is 0.562500, time 226.899 samples/sec
+epoch 16, total_step 254600, total loss is 12.63 , inference loss is 6.24, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.856 samples/sec
+epoch 16, total_step 254620, total loss is 11.92 , inference loss is 5.53, weight deacy loss is 6.39, training accuracy is 0.562500, time 234.040 samples/sec
+epoch 16, total_step 254640, total loss is 13.84 , inference loss is 7.45, weight deacy loss is 6.39, training accuracy is 0.500000, time 298.315 samples/sec
+epoch 16, total_step 254660, total loss is 11.74 , inference loss is 5.35, weight deacy loss is 6.39, training accuracy is 0.500000, time 219.046 samples/sec
+epoch 16, total_step 254680, total loss is 14.45 , inference loss is 8.07, weight deacy loss is 6.39, training accuracy is 0.562500, time 213.120 samples/sec
+epoch 16, total_step 254700, total loss is 14.14 , inference loss is 7.75, weight deacy loss is 6.39, training accuracy is 0.437500, time 13.393 samples/sec
+epoch 16, total_step 254720, total loss is 16.31 , inference loss is 9.92, weight deacy loss is 6.39, training accuracy is 0.406250, time 219.838 samples/sec
+epoch 16, total_step 254740, total loss is 14.55 , inference loss is 8.16, weight deacy loss is 6.39, training accuracy is 0.468750, time 206.459 samples/sec
+epoch 16, total_step 254760, total loss is 18.36 , inference loss is 11.97, weight deacy loss is 6.39, training accuracy is 0.406250, time 223.579 samples/sec
+epoch 16, total_step 254780, total loss is 13.72 , inference loss is 7.33, weight deacy loss is 6.39, training accuracy is 0.531250, time 223.270 samples/sec
+epoch 16, total_step 254800, total loss is 11.96 , inference loss is 5.57, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.174 samples/sec
+epoch 16, total_step 254820, total loss is 15.96 , inference loss is 9.57, weight deacy loss is 6.39, training accuracy is 0.406250, time 218.293 samples/sec
+epoch 16, total_step 254840, total loss is 13.87 , inference loss is 7.47, weight deacy loss is 6.39, training accuracy is 0.343750, time 217.214 samples/sec
+epoch 16, total_step 254860, total loss is 11.45 , inference loss is 5.06, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.227 samples/sec
+epoch 16, total_step 254880, total loss is 15.09 , inference loss is 8.70, weight deacy loss is 6.39, training accuracy is 0.375000, time 211.262 samples/sec
+epoch 16, total_step 254900, total loss is 13.16 , inference loss is 6.77, weight deacy loss is 6.39, training accuracy is 0.437500, time 297.716 samples/sec
+epoch 16, total_step 254920, total loss is 10.33 , inference loss is 3.94, weight deacy loss is 6.39, training accuracy is 0.593750, time 210.637 samples/sec
+epoch 16, total_step 254940, total loss is 10.81 , inference loss is 4.42, weight deacy loss is 6.39, training accuracy is 0.781250, time 220.413 samples/sec
+epoch 16, total_step 254960, total loss is 12.09 , inference loss is 5.70, weight deacy loss is 6.39, training accuracy is 0.687500, time 219.140 samples/sec
+epoch 16, total_step 254980, total loss is 13.43 , inference loss is 7.04, weight deacy loss is 6.39, training accuracy is 0.500000, time 211.201 samples/sec
+epoch 16, total_step 255000, total loss is 7.90 , inference loss is 1.51, weight deacy loss is 6.39, training accuracy is 0.812500, time 12.531 samples/sec
+epoch 16, total_step 255020, total loss is 13.98 , inference loss is 7.59, weight deacy loss is 6.39, training accuracy is 0.531250, time 207.767 samples/sec
+epoch 16, total_step 255040, total loss is 12.58 , inference loss is 6.19, weight deacy loss is 6.39, training accuracy is 0.531250, time 217.451 samples/sec
+epoch 16, total_step 255060, total loss is 10.74 , inference loss is 4.35, weight deacy loss is 6.39, training accuracy is 0.468750, time 218.597 samples/sec
+epoch 16, total_step 255080, total loss is 18.13 , inference loss is 11.74, weight deacy loss is 6.39, training accuracy is 0.343750, time 221.111 samples/sec
+epoch 16, total_step 255100, total loss is 16.16 , inference loss is 9.77, weight deacy loss is 6.39, training accuracy is 0.500000, time 236.953 samples/sec
+epoch 16, total_step 255120, total loss is 12.40 , inference loss is 6.01, weight deacy loss is 6.39, training accuracy is 0.562500, time 207.193 samples/sec
+epoch 16, total_step 255140, total loss is 12.14 , inference loss is 5.75, weight deacy loss is 6.39, training accuracy is 0.656250, time 220.681 samples/sec
+epoch 16, total_step 255160, total loss is 9.26 , inference loss is 2.87, weight deacy loss is 6.39, training accuracy is 0.593750, time 298.570 samples/sec
+epoch 16, total_step 255180, total loss is 10.72 , inference loss is 4.33, weight deacy loss is 6.39, training accuracy is 0.468750, time 217.941 samples/sec
+epoch 16, total_step 255200, total loss is 12.03 , inference loss is 5.63, weight deacy loss is 6.39, training accuracy is 0.468750, time 215.060 samples/sec
+epoch 16, total_step 255220, total loss is 13.00 , inference loss is 6.61, weight deacy loss is 6.39, training accuracy is 0.593750, time 205.813 samples/sec
+epoch 16, total_step 255240, total loss is 11.57 , inference loss is 5.18, weight deacy loss is 6.39, training accuracy is 0.593750, time 208.540 samples/sec
+epoch 16, total_step 255260, total loss is 11.97 , inference loss is 5.58, weight deacy loss is 6.39, training accuracy is 0.531250, time 213.959 samples/sec
+epoch 16, total_step 255280, total loss is 14.20 , inference loss is 7.81, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.211 samples/sec
+epoch 16, total_step 255300, total loss is 13.39 , inference loss is 7.00, weight deacy loss is 6.39, training accuracy is 0.500000, time 12.798 samples/sec
+epoch 16, total_step 255320, total loss is 14.75 , inference loss is 8.36, weight deacy loss is 6.39, training accuracy is 0.468750, time 214.688 samples/sec
+epoch 16, total_step 255340, total loss is 12.84 , inference loss is 6.45, weight deacy loss is 6.39, training accuracy is 0.625000, time 211.741 samples/sec
+epoch 16, total_step 255360, total loss is 13.35 , inference loss is 6.96, weight deacy loss is 6.39, training accuracy is 0.500000, time 210.559 samples/sec
+epoch 16, total_step 255380, total loss is 16.62 , inference loss is 10.23, weight deacy loss is 6.39, training accuracy is 0.406250, time 217.219 samples/sec
+epoch 16, total_step 255400, total loss is 8.77 , inference loss is 2.38, weight deacy loss is 6.39, training accuracy is 0.687500, time 216.194 samples/sec
+epoch 16, total_step 255420, total loss is 14.40 , inference loss is 8.01, weight deacy loss is 6.39, training accuracy is 0.625000, time 297.000 samples/sec
+epoch 16, total_step 255440, total loss is 14.26 , inference loss is 7.87, weight deacy loss is 6.39, training accuracy is 0.500000, time 221.375 samples/sec
+epoch 16, total_step 255460, total loss is 13.70 , inference loss is 7.31, weight deacy loss is 6.39, training accuracy is 0.531250, time 213.969 samples/sec
+epoch 16, total_step 255480, total loss is 10.85 , inference loss is 4.46, weight deacy loss is 6.39, training accuracy is 0.468750, time 225.281 samples/sec
+epoch 16, total_step 255500, total loss is 10.63 , inference loss is 4.24, weight deacy loss is 6.39, training accuracy is 0.500000, time 215.222 samples/sec
+epoch 16, total_step 255520, total loss is 13.86 , inference loss is 7.47, weight deacy loss is 6.39, training accuracy is 0.468750, time 216.144 samples/sec
+epoch 16, total_step 255540, total loss is 10.73 , inference loss is 4.34, weight deacy loss is 6.39, training accuracy is 0.562500, time 227.359 samples/sec
+epoch 16, total_step 255560, total loss is 14.92 , inference loss is 8.53, weight deacy loss is 6.39, training accuracy is 0.531250, time 219.462 samples/sec
+epoch 16, total_step 255580, total loss is 11.31 , inference loss is 4.92, weight deacy loss is 6.39, training accuracy is 0.656250, time 215.680 samples/sec
+epoch 16, total_step 255600, total loss is 12.03 , inference loss is 5.64, weight deacy loss is 6.39, training accuracy is 0.656250, time 13.010 samples/sec
+epoch 16, total_step 255620, total loss is 11.13 , inference loss is 4.74, weight deacy loss is 6.39, training accuracy is 0.500000, time 215.534 samples/sec
+epoch 16, total_step 255640, total loss is 11.38 , inference loss is 4.99, weight deacy loss is 6.39, training accuracy is 0.562500, time 229.568 samples/sec
+epoch 16, total_step 255660, total loss is 11.51 , inference loss is 5.12, weight deacy loss is 6.39, training accuracy is 0.562500, time 214.345 samples/sec
+epoch 16, total_step 255680, total loss is 13.27 , inference loss is 6.88, weight deacy loss is 6.39, training accuracy is 0.531250, time 299.704 samples/sec
+epoch 16, total_step 255700, total loss is 11.63 , inference loss is 5.24, weight deacy loss is 6.39, training accuracy is 0.500000, time 228.098 samples/sec
+epoch 16, total_step 255720, total loss is 15.80 , inference loss is 9.41, weight deacy loss is 6.39, training accuracy is 0.468750, time 218.042 samples/sec
+epoch 16, total_step 255740, total loss is 12.20 , inference loss is 5.81, weight deacy loss is 6.39, training accuracy is 0.625000, time 216.972 samples/sec
+epoch 16, total_step 255760, total loss is 12.11 , inference loss is 5.73, weight deacy loss is 6.39, training accuracy is 0.562500, time 213.076 samples/sec
+epoch 16, total_step 255780, total loss is 13.84 , inference loss is 7.45, weight deacy loss is 6.39, training accuracy is 0.593750, time 221.442 samples/sec
+epoch 16, total_step 255800, total loss is 15.33 , inference loss is 8.94, weight deacy loss is 6.39, training accuracy is 0.500000, time 225.512 samples/sec
+epoch 16, total_step 255820, total loss is 10.81 , inference loss is 4.42, weight deacy loss is 6.39, training accuracy is 0.656250, time 227.162 samples/sec
+epoch 16, total_step 255840, total loss is 12.65 , inference loss is 6.26, weight deacy loss is 6.39, training accuracy is 0.531250, time 213.010 samples/sec
+epoch 16, total_step 255860, total loss is 18.16 , inference loss is 11.78, weight deacy loss is 6.39, training accuracy is 0.468750, time 207.327 samples/sec
+epoch 16, total_step 255880, total loss is 13.33 , inference loss is 6.94, weight deacy loss is 6.39, training accuracy is 0.468750, time 232.873 samples/sec
+epoch 16, total_step 255900, total loss is 13.13 , inference loss is 6.74, weight deacy loss is 6.39, training accuracy is 0.531250, time 12.603 samples/sec
+epoch 16, total_step 255920, total loss is 17.68 , inference loss is 11.29, weight deacy loss is 6.39, training accuracy is 0.468750, time 212.091 samples/sec
+epoch 16, total_step 255940, total loss is 13.60 , inference loss is 7.21, weight deacy loss is 6.39, training accuracy is 0.593750, time 182.359 samples/sec
+epoch 16, total_step 255960, total loss is 12.01 , inference loss is 5.62, weight deacy loss is 6.39, training accuracy is 0.562500, time 215.665 samples/sec
+epoch 16, total_step 255980, total loss is 11.80 , inference loss is 5.41, weight deacy loss is 6.39, training accuracy is 0.562500, time 208.647 samples/sec
+testing verification..
+(12000, 512)
+infer time 26.01503499999998
+best_threshold_index 115 0.9844444444444445
+best_threshold_index 115 0.9835185185185186
+best_threshold_index 115 0.9831481481481481
+best_threshold_index 116 0.9851851851851852
+best_threshold_index 115 0.9840740740740741
+best_threshold_index 116 0.9829629629629629
+best_threshold_index 115 0.9833333333333333
+best_threshold_index 115 0.9831481481481481
+best_threshold_index 115 0.9824074074074074
+best_threshold_index 115 0.9831481481481481
+[lfw][256000]XNorm: 21.499211
+[lfw][256000]Accuracy-Flip: 0.98317+-0.00693
+test accuracy is: 0.9831666666666667
+epoch 16, total_step 256000, total loss is 16.69 , inference loss is 10.30, weight deacy loss is 6.39, training accuracy is 0.437500, time 16.079 samples/sec
+epoch 16, total_step 256020, total loss is 11.87 , inference loss is 5.48, weight deacy loss is 6.39, training accuracy is 0.468750, time 207.021 samples/sec
+epoch 16, total_step 256040, total loss is 16.41 , inference loss is 10.02, weight deacy loss is 6.39, training accuracy is 0.406250, time 299.960 samples/sec
+epoch 16, total_step 256060, total loss is 11.31 , inference loss is 4.92, weight deacy loss is 6.39, training accuracy is 0.562500, time 219.459 samples/sec
+epoch 16, total_step 256080, total loss is 17.56 , inference loss is 11.17, weight deacy loss is 6.39, training accuracy is 0.531250, time 218.991 samples/sec
+epoch 16, total_step 256100, total loss is 13.47 , inference loss is 7.08, weight deacy loss is 6.39, training accuracy is 0.593750, time 215.923 samples/sec
+epoch 16, total_step 256120, total loss is 14.91 , inference loss is 8.52, weight deacy loss is 6.39, training accuracy is 0.500000, time 227.510 samples/sec
+epoch 16, total_step 256140, total loss is 13.01 , inference loss is 6.62, weight deacy loss is 6.39, training accuracy is 0.531250, time 229.210 samples/sec
+epoch 16, total_step 256160, total loss is 11.87 , inference loss is 5.48, weight deacy loss is 6.39, training accuracy is 0.593750, time 219.230 samples/sec
+epoch 16, total_step 256180, total loss is 13.84 , inference loss is 7.45, weight deacy loss is 6.39, training accuracy is 0.406250, time 219.259 samples/sec
+epoch 16, total_step 256200, total loss is 15.60 , inference loss is 9.21, weight deacy loss is 6.39, training accuracy is 0.468750, time 13.540 samples/sec
+epoch 16, total_step 256220, total loss is 11.88 , inference loss is 5.49, weight deacy loss is 6.39, training accuracy is 0.625000, time 224.199 samples/sec
+epoch 16, total_step 256240, total loss is 14.95 , inference loss is 8.56, weight deacy loss is 6.39, training accuracy is 0.437500, time 218.248 samples/sec
+epoch 16, total_step 256260, total loss is 12.52 , inference loss is 6.13, weight deacy loss is 6.39, training accuracy is 0.593750, time 219.537 samples/sec
+epoch 16, total_step 256280, total loss is 12.79 , inference loss is 6.40, weight deacy loss is 6.39, training accuracy is 0.468750, time 213.872 samples/sec
+epoch 16, total_step 256300, total loss is 12.47 , inference loss is 6.08, weight deacy loss is 6.39, training accuracy is 0.593750, time 214.085 samples/sec
+epoch 16, total_step 256320, total loss is 11.38 , inference loss is 4.99, weight deacy loss is 6.39, training accuracy is 0.562500, time 213.695 samples/sec
+epoch 16, total_step 256340, total loss is 15.51 , inference loss is 9.12, weight deacy loss is 6.39, training accuracy is 0.406250, time 212.540 samples/sec
+epoch 16, total_step 256360, total loss is 11.93 , inference loss is 5.54, weight deacy loss is 6.39, training accuracy is 0.500000, time 213.562 samples/sec
+epoch 16, total_step 256380, total loss is 15.24 , inference loss is 8.85, weight deacy loss is 6.39, training accuracy is 0.343750, time 210.099 samples/sec
+epoch 16, total_step 256400, total loss is 11.80 , inference loss is 5.41, weight deacy loss is 6.39, training accuracy is 0.437500, time 217.584 samples/sec
+epoch 16, total_step 256420, total loss is 11.81 , inference loss is 5.42, weight deacy loss is 6.39, training accuracy is 0.468750, time 210.996 samples/sec
+epoch 16, total_step 256440, total loss is 14.87 , inference loss is 8.48, weight deacy loss is 6.39, training accuracy is 0.343750, time 217.616 samples/sec
+epoch 16, total_step 256460, total loss is 12.49 , inference loss is 6.10, weight deacy loss is 6.39, training accuracy is 0.625000, time 211.606 samples/sec
+epoch 16, total_step 256480, total loss is 12.51 , inference loss is 6.12, weight deacy loss is 6.39, training accuracy is 0.562500, time 215.240 samples/sec
+epoch 16, total_step 256500, total loss is 11.61 , inference loss is 5.22, weight deacy loss is 6.39, training accuracy is 0.500000, time 12.810 samples/sec
+epoch 16, total_step 256520, total loss is 17.28 , inference loss is 10.89, weight deacy loss is 6.39, training accuracy is 0.468750, time 213.027 samples/sec
+epoch 16, total_step 256540, total loss is 13.71 , inference loss is 7.32, weight deacy loss is 6.39, training accuracy is 0.531250, time 212.273 samples/sec
+epoch 16, total_step 256560, total loss is 13.94 , inference loss is 7.55, weight deacy loss is 6.39, training accuracy is 0.562500, time 213.196 samples/sec
+epoch 16, total_step 256580, total loss is 10.99 , inference loss is 4.61, weight deacy loss is 6.39, training accuracy is 0.562500, time 224.447 samples/sec
+epoch 16, total_step 256600, total loss is 16.76 , inference loss is 10.37, weight deacy loss is 6.39, training accuracy is 0.406250, time 215.995 samples/sec
+epoch 16, total_step 256620, total loss is 18.37 , inference loss is 11.98, weight deacy loss is 6.39, training accuracy is 0.406250, time 223.852 samples/sec
+epoch 16, total_step 256640, total loss is 11.42 , inference loss is 5.03, weight deacy loss is 6.39, training accuracy is 0.500000, time 227.649 samples/sec
+epoch 16, total_step 256660, total loss is 11.19 , inference loss is 4.80, weight deacy loss is 6.39, training accuracy is 0.593750, time 218.882 samples/sec
+epoch 16, total_step 256680, total loss is 15.67 , inference loss is 9.28, weight deacy loss is 6.39, training accuracy is 0.562500, time 207.362 samples/sec
+epoch 16, total_step 256700, total loss is 12.03 , inference loss is 5.64, weight deacy loss is 6.39, training accuracy is 0.437500, time 221.349 samples/sec
+epoch 16, total_step 256720, total loss is 10.99 , inference loss is 4.60, weight deacy loss is 6.39, training accuracy is 0.656250, time 226.279 samples/sec
+epoch 16, total_step 256740, total loss is 14.79 , inference loss is 8.40, weight deacy loss is 6.39, training accuracy is 0.437500, time 210.899 samples/sec
+epoch 16, total_step 256760, total loss is 11.54 , inference loss is 5.15, weight deacy loss is 6.39, training accuracy is 0.562500, time 221.473 samples/sec
+epoch 16, total_step 256780, total loss is 16.01 , inference loss is 9.62, weight deacy loss is 6.39, training accuracy is 0.437500, time 298.414 samples/sec
+epoch 16, total_step 256800, total loss is 13.15 , inference loss is 6.76, weight deacy loss is 6.39, training accuracy is 0.500000, time 12.853 samples/sec
+epoch 16, total_step 256820, total loss is 11.71 , inference loss is 5.32, weight deacy loss is 6.39, training accuracy is 0.500000, time 215.845 samples/sec
+epoch 16, total_step 256840, total loss is 12.14 , inference loss is 5.75, weight deacy loss is 6.39, training accuracy is 0.500000, time 216.415 samples/sec
+epoch 16, total_step 256860, total loss is 14.28 , inference loss is 7.89, weight deacy loss is 6.39, training accuracy is 0.500000, time 224.815 samples/sec
+epoch 16, total_step 256880, total loss is 17.78 , inference loss is 11.39, weight deacy loss is 6.39, training accuracy is 0.375000, time 221.186 samples/sec
+epoch 16, total_step 256900, total loss is 14.00 , inference loss is 7.61, weight deacy loss is 6.39, training accuracy is 0.468750, time 218.698 samples/sec
+epoch 16, total_step 256920, total loss is 14.26 , inference loss is 7.87, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.372 samples/sec
+epoch 16, total_step 256940, total loss is 17.39 , inference loss is 11.00, weight deacy loss is 6.39, training accuracy is 0.312500, time 209.400 samples/sec
+epoch 16, total_step 256960, total loss is 12.30 , inference loss is 5.91, weight deacy loss is 6.39, training accuracy is 0.593750, time 215.140 samples/sec
+epoch 16, total_step 256980, total loss is 9.76 , inference loss is 3.37, weight deacy loss is 6.39, training accuracy is 0.718750, time 211.994 samples/sec
+epoch 16, total_step 257000, total loss is 14.44 , inference loss is 8.05, weight deacy loss is 6.39, training accuracy is 0.562500, time 222.110 samples/sec
+epoch 16, total_step 257020, total loss is 12.23 , inference loss is 5.84, weight deacy loss is 6.39, training accuracy is 0.468750, time 204.194 samples/sec
+epoch 16, total_step 257040, total loss is 12.93 , inference loss is 6.54, weight deacy loss is 6.39, training accuracy is 0.500000, time 299.537 samples/sec
+epoch 16, total_step 257060, total loss is 13.60 , inference loss is 7.21, weight deacy loss is 6.39, training accuracy is 0.500000, time 203.033 samples/sec
+epoch 16, total_step 257080, total loss is 14.67 , inference loss is 8.28, weight deacy loss is 6.39, training accuracy is 0.500000, time 210.404 samples/sec
+epoch 16, total_step 257100, total loss is 14.63 , inference loss is 8.24, weight deacy loss is 6.39, training accuracy is 0.500000, time 12.705 samples/sec
+epoch 16, total_step 257120, total loss is 10.26 , inference loss is 3.87, weight deacy loss is 6.39, training accuracy is 0.562500, time 196.457 samples/sec
+epoch 16, total_step 257140, total loss is 15.98 , inference loss is 9.59, weight deacy loss is 6.39, training accuracy is 0.375000, time 220.374 samples/sec
+epoch 16, total_step 257160, total loss is 11.63 , inference loss is 5.24, weight deacy loss is 6.39, training accuracy is 0.406250, time 197.999 samples/sec
+epoch 16, total_step 257180, total loss is 10.45 , inference loss is 4.06, weight deacy loss is 6.39, training accuracy is 0.625000, time 202.448 samples/sec
+epoch 16, total_step 257200, total loss is 14.25 , inference loss is 7.86, weight deacy loss is 6.39, training accuracy is 0.406250, time 210.583 samples/sec
+epoch 16, total_step 257220, total loss is 11.85 , inference loss is 5.46, weight deacy loss is 6.39, training accuracy is 0.500000, time 216.132 samples/sec
+epoch 16, total_step 257240, total loss is 12.99 , inference loss is 6.60, weight deacy loss is 6.39, training accuracy is 0.468750, time 216.144 samples/sec
+epoch 16, total_step 257260, total loss is 12.94 , inference loss is 6.55, weight deacy loss is 6.39, training accuracy is 0.437500, time 205.160 samples/sec
+epoch 16, total_step 257280, total loss is 12.98 , inference loss is 6.59, weight deacy loss is 6.39, training accuracy is 0.437500, time 212.433 samples/sec
+epoch 16, total_step 257300, total loss is 10.87 , inference loss is 4.48, weight deacy loss is 6.39, training accuracy is 0.625000, time 297.551 samples/sec
+epoch 16, total_step 257320, total loss is 11.69 , inference loss is 5.30, weight deacy loss is 6.39, training accuracy is 0.687500, time 225.618 samples/sec
+epoch 16, total_step 257340, total loss is 12.24 , inference loss is 5.86, weight deacy loss is 6.39, training accuracy is 0.531250, time 211.819 samples/sec
+epoch 16, total_step 257360, total loss is 10.42 , inference loss is 4.03, weight deacy loss is 6.39, training accuracy is 0.562500, time 174.671 samples/sec
+epoch 16, total_step 257380, total loss is 12.70 , inference loss is 6.31, weight deacy loss is 6.39, training accuracy is 0.437500, time 221.985 samples/sec
+epoch 16, total_step 257400, total loss is 11.97 , inference loss is 5.58, weight deacy loss is 6.39, training accuracy is 0.531250, time 11.606 samples/sec
+epoch 16, total_step 257420, total loss is 12.84 , inference loss is 6.45, weight deacy loss is 6.39, training accuracy is 0.468750, time 210.810 samples/sec
+epoch 16, total_step 257440, total loss is 12.15 , inference loss is 5.76, weight deacy loss is 6.39, training accuracy is 0.500000, time 208.665 samples/sec
+epoch 16, total_step 257460, total loss is 12.97 , inference loss is 6.58, weight deacy loss is 6.39, training accuracy is 0.500000, time 217.073 samples/sec
+epoch 16, total_step 257480, total loss is 14.09 , inference loss is 7.70, weight deacy loss is 6.39, training accuracy is 0.625000, time 225.510 samples/sec
+epoch 16, total_step 257500, total loss is 9.99 , inference loss is 3.61, weight deacy loss is 6.39, training accuracy is 0.625000, time 208.490 samples/sec
+epoch 16, total_step 257520, total loss is 14.64 , inference loss is 8.26, weight deacy loss is 6.39, training accuracy is 0.500000, time 186.974 samples/sec
+epoch 16, total_step 257540, total loss is 12.92 , inference loss is 6.54, weight deacy loss is 6.39, training accuracy is 0.500000, time 222.184 samples/sec
+epoch 16, total_step 257560, total loss is 9.76 , inference loss is 3.38, weight deacy loss is 6.39, training accuracy is 0.750000, time 211.492 samples/sec
+epoch 16, total_step 257580, total loss is 11.48 , inference loss is 5.09, weight deacy loss is 6.39, training accuracy is 0.468750, time 196.614 samples/sec
+epoch 16, total_step 257600, total loss is 18.23 , inference loss is 11.84, weight deacy loss is 6.39, training accuracy is 0.343750, time 211.990 samples/sec
+epoch 16, total_step 257620, total loss is 13.53 , inference loss is 7.14, weight deacy loss is 6.39, training accuracy is 0.375000, time 204.193 samples/sec
+epoch 16, total_step 257640, total loss is 15.75 , inference loss is 9.36, weight deacy loss is 6.39, training accuracy is 0.437500, time 299.633 samples/sec
+epoch 16, total_step 257660, total loss is 15.82 , inference loss is 9.43, weight deacy loss is 6.39, training accuracy is 0.343750, time 235.356 samples/sec
+epoch 16, total_step 257680, total loss is 17.89 , inference loss is 11.50, weight deacy loss is 6.39, training accuracy is 0.281250, time 204.312 samples/sec
+epoch 16, total_step 257700, total loss is 13.48 , inference loss is 7.09, weight deacy loss is 6.39, training accuracy is 0.500000, time 13.275 samples/sec
+epoch 16, total_step 257720, total loss is 13.69 , inference loss is 7.31, weight deacy loss is 6.39, training accuracy is 0.406250, time 209.034 samples/sec
+epoch 16, total_step 257740, total loss is 13.51 , inference loss is 7.13, weight deacy loss is 6.39, training accuracy is 0.375000, time 211.743 samples/sec
+epoch 16, total_step 257760, total loss is 13.52 , inference loss is 7.14, weight deacy loss is 6.39, training accuracy is 0.562500, time 197.277 samples/sec
+epoch 16, total_step 257780, total loss is 14.19 , inference loss is 7.80, weight deacy loss is 6.39, training accuracy is 0.593750, time 202.233 samples/sec
+epoch 16, total_step 257800, total loss is 11.16 , inference loss is 4.77, weight deacy loss is 6.39, training accuracy is 0.343750, time 217.788 samples/sec
+epoch 16, total_step 257820, total loss is 13.57 , inference loss is 7.18, weight deacy loss is 6.39, training accuracy is 0.625000, time 216.213 samples/sec
+epoch 16, total_step 257840, total loss is 15.35 , inference loss is 8.97, weight deacy loss is 6.39, training accuracy is 0.468750, time 215.997 samples/sec
+epoch 16, total_step 257860, total loss is 12.93 , inference loss is 6.54, weight deacy loss is 6.39, training accuracy is 0.437500, time 225.389 samples/sec
+epoch 16, total_step 257880, total loss is 12.64 , inference loss is 6.25, weight deacy loss is 6.39, training accuracy is 0.406250, time 213.529 samples/sec
+epoch 16, total_step 257900, total loss is 12.82 , inference loss is 6.43, weight deacy loss is 6.39, training accuracy is 0.406250, time 210.973 samples/sec
+epoch 16, total_step 257920, total loss is 11.75 , inference loss is 5.36, weight deacy loss is 6.39, training accuracy is 0.562500, time 216.983 samples/sec
+epoch 16, total_step 257940, total loss is 14.06 , inference loss is 7.67, weight deacy loss is 6.39, training accuracy is 0.343750, time 220.328 samples/sec
+epoch 16, total_step 257960, total loss is 15.53 , inference loss is 9.15, weight deacy loss is 6.39, training accuracy is 0.343750, time 300.560 samples/sec
+epoch 16, total_step 257980, total loss is 12.10 , inference loss is 5.71, weight deacy loss is 6.39, training accuracy is 0.531250, time 300.936 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.371736999999996
+best_threshold_index 110 0.9875925925925926
+best_threshold_index 110 0.9861111111111112
+best_threshold_index 110 0.9864814814814815
+best_threshold_index 110 0.9875925925925926
+best_threshold_index 110 0.987037037037037
+best_threshold_index 110 0.9857407407407407
+best_threshold_index 110 0.9868518518518519
+best_threshold_index 110 0.9862962962962963
+best_threshold_index 110 0.9855555555555555
+best_threshold_index 110 0.9857407407407407
+[lfw][258000]XNorm: 21.391791
+[lfw][258000]Accuracy-Flip: 0.98650+-0.00639
+test accuracy is: 0.9865
+epoch 16, total_step 258000, total loss is 14.74 , inference loss is 8.36, weight deacy loss is 6.39, training accuracy is 0.406250, time 17.809 samples/sec
+epoch 16, total_step 258020, total loss is 15.60 , inference loss is 9.21, weight deacy loss is 6.39, training accuracy is 0.406250, time 296.375 samples/sec
+epoch 16, total_step 258040, total loss is 14.60 , inference loss is 8.21, weight deacy loss is 6.39, training accuracy is 0.375000, time 300.538 samples/sec
+epoch 16, total_step 258060, total loss is 16.85 , inference loss is 10.46, weight deacy loss is 6.39, training accuracy is 0.437500, time 206.648 samples/sec
+epoch 16, total_step 258080, total loss is 16.88 , inference loss is 10.49, weight deacy loss is 6.39, training accuracy is 0.406250, time 219.089 samples/sec
+epoch 16, total_step 258100, total loss is 12.74 , inference loss is 6.35, weight deacy loss is 6.39, training accuracy is 0.437500, time 216.255 samples/sec
+epoch 16, total_step 258120, total loss is 12.70 , inference loss is 6.32, weight deacy loss is 6.39, training accuracy is 0.500000, time 215.193 samples/sec
+epoch 16, total_step 258140, total loss is 14.43 , inference loss is 8.05, weight deacy loss is 6.39, training accuracy is 0.437500, time 203.237 samples/sec
+epoch 16, total_step 258160, total loss is 8.88 , inference loss is 2.49, weight deacy loss is 6.39, training accuracy is 0.468750, time 210.977 samples/sec
+epoch 16, total_step 258180, total loss is 13.95 , inference loss is 7.56, weight deacy loss is 6.39, training accuracy is 0.562500, time 222.314 samples/sec
+epoch 16, total_step 258200, total loss is 15.85 , inference loss is 9.46, weight deacy loss is 6.39, training accuracy is 0.343750, time 210.509 samples/sec
+epoch 16, total_step 258220, total loss is 15.46 , inference loss is 9.08, weight deacy loss is 6.39, training accuracy is 0.343750, time 206.130 samples/sec
+epoch 16, total_step 258240, total loss is 16.71 , inference loss is 10.33, weight deacy loss is 6.39, training accuracy is 0.312500, time 195.083 samples/sec
+epoch 16, total_step 258260, total loss is 17.18 , inference loss is 10.80, weight deacy loss is 6.39, training accuracy is 0.312500, time 204.037 samples/sec
+epoch 16, total_step 258280, total loss is 10.70 , inference loss is 4.32, weight deacy loss is 6.39, training accuracy is 0.562500, time 210.351 samples/sec
+epoch 16, total_step 258300, total loss is 14.97 , inference loss is 8.58, weight deacy loss is 6.39, training accuracy is 0.312500, time 12.610 samples/sec
+epoch 16, total_step 258320, total loss is 15.01 , inference loss is 8.62, weight deacy loss is 6.39, training accuracy is 0.500000, time 175.053 samples/sec
+epoch 16, total_step 258340, total loss is 11.18 , inference loss is 4.79, weight deacy loss is 6.39, training accuracy is 0.531250, time 214.772 samples/sec
+epoch 16, total_step 258360, total loss is 12.38 , inference loss is 6.00, weight deacy loss is 6.39, training accuracy is 0.406250, time 208.639 samples/sec
+epoch 16, total_step 258380, total loss is 14.13 , inference loss is 7.74, weight deacy loss is 6.39, training accuracy is 0.312500, time 194.728 samples/sec
+epoch 16, total_step 258400, total loss is 14.56 , inference loss is 8.17, weight deacy loss is 6.39, training accuracy is 0.437500, time 202.146 samples/sec
+epoch 16, total_step 258420, total loss is 13.08 , inference loss is 6.69, weight deacy loss is 6.39, training accuracy is 0.500000, time 198.886 samples/sec
+epoch 16, total_step 258440, total loss is 13.00 , inference loss is 6.61, weight deacy loss is 6.39, training accuracy is 0.281250, time 221.869 samples/sec
+epoch 16, total_step 258460, total loss is 13.77 , inference loss is 7.38, weight deacy loss is 6.39, training accuracy is 0.375000, time 211.809 samples/sec
+epoch 16, total_step 258480, total loss is 11.82 , inference loss is 5.43, weight deacy loss is 6.39, training accuracy is 0.343750, time 204.527 samples/sec
+epoch 16, total_step 258500, total loss is 10.75 , inference loss is 4.36, weight deacy loss is 6.39, training accuracy is 0.468750, time 200.133 samples/sec
+epoch 16, total_step 258520, total loss is 15.25 , inference loss is 8.86, weight deacy loss is 6.39, training accuracy is 0.375000, time 209.620 samples/sec
+epoch 16, total_step 258540, total loss is 18.27 , inference loss is 11.88, weight deacy loss is 6.39, training accuracy is 0.375000, time 214.196 samples/sec
+epoch 16, total_step 258560, total loss is 12.81 , inference loss is 6.42, weight deacy loss is 6.39, training accuracy is 0.593750, time 225.233 samples/sec
+epoch 16, total_step 258580, total loss is 14.91 , inference loss is 8.52, weight deacy loss is 6.39, training accuracy is 0.375000, time 192.427 samples/sec
+epoch 16, total_step 258600, total loss is 13.35 , inference loss is 6.97, weight deacy loss is 6.39, training accuracy is 0.500000, time 7.891 samples/sec
+epoch 16, total_step 258620, total loss is 15.64 , inference loss is 9.25, weight deacy loss is 6.39, training accuracy is 0.406250, time 196.998 samples/sec
+epoch 16, total_step 258640, total loss is 12.81 , inference loss is 6.43, weight deacy loss is 6.39, training accuracy is 0.531250, time 207.939 samples/sec
+epoch 16, total_step 258660, total loss is 15.61 , inference loss is 9.22, weight deacy loss is 6.39, training accuracy is 0.500000, time 205.483 samples/sec
+epoch 16, total_step 258680, total loss is 10.34 , inference loss is 3.95, weight deacy loss is 6.39, training accuracy is 0.531250, time 214.777 samples/sec
+epoch 16, total_step 258700, total loss is 12.69 , inference loss is 6.31, weight deacy loss is 6.39, training accuracy is 0.500000, time 204.503 samples/sec
+epoch 16, total_step 258720, total loss is 14.86 , inference loss is 8.47, weight deacy loss is 6.39, training accuracy is 0.406250, time 231.065 samples/sec
+epoch 16, total_step 258740, total loss is 12.18 , inference loss is 5.79, weight deacy loss is 6.39, training accuracy is 0.468750, time 209.398 samples/sec
+epoch 16, total_step 258760, total loss is 12.59 , inference loss is 6.21, weight deacy loss is 6.39, training accuracy is 0.468750, time 204.277 samples/sec
+epoch 16, total_step 258780, total loss is 13.39 , inference loss is 7.01, weight deacy loss is 6.39, training accuracy is 0.468750, time 192.378 samples/sec
+epoch 16, total_step 258800, total loss is 13.66 , inference loss is 7.27, weight deacy loss is 6.39, training accuracy is 0.406250, time 298.590 samples/sec
+epoch 16, total_step 258820, total loss is 13.10 , inference loss is 6.71, weight deacy loss is 6.39, training accuracy is 0.437500, time 299.365 samples/sec
+epoch 16, total_step 258840, total loss is 17.26 , inference loss is 10.87, weight deacy loss is 6.39, training accuracy is 0.343750, time 197.279 samples/sec
+epoch 16, total_step 258860, total loss is 11.54 , inference loss is 5.16, weight deacy loss is 6.39, training accuracy is 0.562500, time 205.456 samples/sec
+epoch 16, total_step 258880, total loss is 14.30 , inference loss is 7.92, weight deacy loss is 6.39, training accuracy is 0.468750, time 207.656 samples/sec
+epoch 16, total_step 258900, total loss is 10.28 , inference loss is 3.90, weight deacy loss is 6.39, training accuracy is 0.625000, time 9.720 samples/sec
+epoch 16, total_step 258920, total loss is 11.47 , inference loss is 5.09, weight deacy loss is 6.39, training accuracy is 0.468750, time 222.733 samples/sec
+epoch 16, total_step 258940, total loss is 12.30 , inference loss is 5.92, weight deacy loss is 6.39, training accuracy is 0.562500, time 174.803 samples/sec
+epoch 16, total_step 258960, total loss is 15.27 , inference loss is 8.88, weight deacy loss is 6.39, training accuracy is 0.343750, time 208.841 samples/sec
+epoch 16, total_step 258980, total loss is 13.68 , inference loss is 7.29, weight deacy loss is 6.39, training accuracy is 0.500000, time 180.649 samples/sec
+epoch 16, total_step 259000, total loss is 11.35 , inference loss is 4.96, weight deacy loss is 6.39, training accuracy is 0.531250, time 215.180 samples/sec
+epoch 16, total_step 259020, total loss is 11.33 , inference loss is 4.95, weight deacy loss is 6.39, training accuracy is 0.468750, time 214.855 samples/sec
+epoch 16, total_step 259040, total loss is 15.57 , inference loss is 9.18, weight deacy loss is 6.39, training accuracy is 0.437500, time 297.531 samples/sec
+epoch 16, total_step 259060, total loss is 15.48 , inference loss is 9.09, weight deacy loss is 6.39, training accuracy is 0.437500, time 217.985 samples/sec
+epoch 16, total_step 259080, total loss is 10.00 , inference loss is 3.62, weight deacy loss is 6.39, training accuracy is 0.625000, time 227.832 samples/sec
+epoch 16, total_step 259100, total loss is 13.55 , inference loss is 7.17, weight deacy loss is 6.39, training accuracy is 0.343750, time 213.305 samples/sec
+epoch 16, total_step 259120, total loss is 12.24 , inference loss is 5.85, weight deacy loss is 6.39, training accuracy is 0.531250, time 202.814 samples/sec
+epoch 16, total_step 259140, total loss is 11.98 , inference loss is 5.60, weight deacy loss is 6.39, training accuracy is 0.500000, time 203.895 samples/sec
+epoch 16, total_step 259160, total loss is 11.24 , inference loss is 4.86, weight deacy loss is 6.39, training accuracy is 0.406250, time 228.576 samples/sec
+epoch 16, total_step 259180, total loss is 15.25 , inference loss is 8.86, weight deacy loss is 6.39, training accuracy is 0.500000, time 298.533 samples/sec
+epoch 16, total_step 259200, total loss is 10.90 , inference loss is 4.51, weight deacy loss is 6.38, training accuracy is 0.531250, time 9.083 samples/sec
+epoch 16, total_step 259220, total loss is 17.09 , inference loss is 10.71, weight deacy loss is 6.38, training accuracy is 0.343750, time 299.675 samples/sec
+epoch 16, total_step 259240, total loss is 14.08 , inference loss is 7.70, weight deacy loss is 6.38, training accuracy is 0.437500, time 300.830 samples/sec
+epoch 16, total_step 259260, total loss is 14.92 , inference loss is 8.54, weight deacy loss is 6.38, training accuracy is 0.312500, time 302.652 samples/sec
+epoch 16, total_step 259280, total loss is 14.17 , inference loss is 7.79, weight deacy loss is 6.38, training accuracy is 0.375000, time 299.956 samples/sec
+epoch 16, total_step 259300, total loss is 17.29 , inference loss is 10.91, weight deacy loss is 6.38, training accuracy is 0.375000, time 219.217 samples/sec
+epoch 16, total_step 259320, total loss is 17.66 , inference loss is 11.28, weight deacy loss is 6.38, training accuracy is 0.312500, time 229.595 samples/sec
+epoch 16, total_step 259340, total loss is 15.57 , inference loss is 9.18, weight deacy loss is 6.38, training accuracy is 0.375000, time 232.561 samples/sec
+epoch 16, total_step 259360, total loss is 12.53 , inference loss is 6.14, weight deacy loss is 6.38, training accuracy is 0.468750, time 226.677 samples/sec
+epoch 16, total_step 259380, total loss is 16.18 , inference loss is 9.80, weight deacy loss is 6.38, training accuracy is 0.500000, time 221.120 samples/sec
+epoch 16, total_step 259400, total loss is 11.67 , inference loss is 5.28, weight deacy loss is 6.38, training accuracy is 0.375000, time 214.579 samples/sec
+epoch 16, total_step 259420, total loss is 16.38 , inference loss is 9.99, weight deacy loss is 6.38, training accuracy is 0.406250, time 223.479 samples/sec
+epoch 16, total_step 259440, total loss is 13.95 , inference loss is 7.57, weight deacy loss is 6.38, training accuracy is 0.375000, time 224.557 samples/sec
+epoch 16, total_step 259460, total loss is 16.45 , inference loss is 10.06, weight deacy loss is 6.38, training accuracy is 0.406250, time 299.229 samples/sec
+epoch 16, total_step 259480, total loss is 12.91 , inference loss is 6.52, weight deacy loss is 6.38, training accuracy is 0.500000, time 301.523 samples/sec
+epoch 16, total_step 259500, total loss is 12.83 , inference loss is 6.44, weight deacy loss is 6.38, training accuracy is 0.531250, time 14.395 samples/sec
+epoch 16, total_step 259520, total loss is 15.89 , inference loss is 9.51, weight deacy loss is 6.38, training accuracy is 0.375000, time 296.744 samples/sec
+epoch 16, total_step 259540, total loss is 18.59 , inference loss is 12.21, weight deacy loss is 6.38, training accuracy is 0.250000, time 213.565 samples/sec
+epoch 16, total_step 259560, total loss is 13.68 , inference loss is 7.30, weight deacy loss is 6.38, training accuracy is 0.500000, time 209.256 samples/sec
+epoch 16, total_step 259580, total loss is 14.88 , inference loss is 8.50, weight deacy loss is 6.38, training accuracy is 0.343750, time 196.988 samples/sec
+epoch 16, total_step 259600, total loss is 15.20 , inference loss is 8.81, weight deacy loss is 6.38, training accuracy is 0.343750, time 187.711 samples/sec
+epoch 16, total_step 259620, total loss is 15.47 , inference loss is 9.09, weight deacy loss is 6.38, training accuracy is 0.375000, time 213.058 samples/sec
+epoch 16, total_step 259640, total loss is 14.82 , inference loss is 8.43, weight deacy loss is 6.38, training accuracy is 0.343750, time 209.955 samples/sec
+epoch 16, total_step 259660, total loss is 14.51 , inference loss is 8.12, weight deacy loss is 6.38, training accuracy is 0.312500, time 200.794 samples/sec
+epoch 16, total_step 259680, total loss is 12.70 , inference loss is 6.32, weight deacy loss is 6.38, training accuracy is 0.437500, time 213.619 samples/sec
+epoch 16, total_step 259700, total loss is 12.09 , inference loss is 5.70, weight deacy loss is 6.38, training accuracy is 0.312500, time 215.414 samples/sec
+epoch 16, total_step 259720, total loss is 14.63 , inference loss is 8.24, weight deacy loss is 6.38, training accuracy is 0.312500, time 298.904 samples/sec
+epoch 16, total_step 259740, total loss is 12.75 , inference loss is 6.37, weight deacy loss is 6.38, training accuracy is 0.531250, time 195.198 samples/sec
+epoch 16, total_step 259760, total loss is 11.93 , inference loss is 5.55, weight deacy loss is 6.38, training accuracy is 0.593750, time 213.383 samples/sec
+epoch 16, total_step 259780, total loss is 15.93 , inference loss is 9.54, weight deacy loss is 6.38, training accuracy is 0.562500, time 202.882 samples/sec
+epoch 16, total_step 259800, total loss is 14.25 , inference loss is 7.87, weight deacy loss is 6.38, training accuracy is 0.468750, time 12.368 samples/sec
+epoch 16, total_step 259820, total loss is 13.14 , inference loss is 6.75, weight deacy loss is 6.38, training accuracy is 0.312500, time 218.310 samples/sec
+epoch 16, total_step 259840, total loss is 13.31 , inference loss is 6.92, weight deacy loss is 6.38, training accuracy is 0.562500, time 231.130 samples/sec
+epoch 16, total_step 259860, total loss is 15.89 , inference loss is 9.50, weight deacy loss is 6.38, training accuracy is 0.250000, time 211.557 samples/sec
+epoch 16, total_step 259880, total loss is 14.77 , inference loss is 8.39, weight deacy loss is 6.38, training accuracy is 0.437500, time 200.470 samples/sec
+epoch 16, total_step 259900, total loss is 10.90 , inference loss is 4.52, weight deacy loss is 6.38, training accuracy is 0.500000, time 216.796 samples/sec
+epoch 16, total_step 259920, total loss is 14.74 , inference loss is 8.35, weight deacy loss is 6.38, training accuracy is 0.375000, time 202.366 samples/sec
+epoch 16, total_step 259940, total loss is 10.79 , inference loss is 4.40, weight deacy loss is 6.38, training accuracy is 0.437500, time 186.583 samples/sec
+epoch 16, total_step 259960, total loss is 12.74 , inference loss is 6.35, weight deacy loss is 6.38, training accuracy is 0.281250, time 214.369 samples/sec
+epoch 16, total_step 259980, total loss is 14.93 , inference loss is 8.54, weight deacy loss is 6.38, training accuracy is 0.468750, time 301.469 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.478475000000024
+best_threshold_index 109 0.9844444444444445
+best_threshold_index 111 0.9835185185185186
+best_threshold_index 109 0.9837037037037037
+best_threshold_index 109 0.9857407407407407
+best_threshold_index 112 0.9846296296296296
+best_threshold_index 111 0.9831481481481481
+best_threshold_index 109 0.9842592592592593
+best_threshold_index 109 0.9831481481481481
+best_threshold_index 109 0.9822222222222222
+best_threshold_index 109 0.9827777777777778
+[lfw][260000]XNorm: 22.419194
+[lfw][260000]Accuracy-Flip: 0.98283+-0.00879
+test accuracy is: 0.9828333333333333
+epoch 16, total_step 260000, total loss is 12.40 , inference loss is 6.01, weight deacy loss is 6.38, training accuracy is 0.375000, time 17.499 samples/sec
+epoch 16, total_step 260020, total loss is 13.91 , inference loss is 7.53, weight deacy loss is 6.38, training accuracy is 0.437500, time 202.251 samples/sec
+epoch 16, total_step 260040, total loss is 14.01 , inference loss is 7.63, weight deacy loss is 6.38, training accuracy is 0.437500, time 224.494 samples/sec
+epoch 16, total_step 260060, total loss is 14.38 , inference loss is 7.99, weight deacy loss is 6.38, training accuracy is 0.281250, time 196.438 samples/sec
+epoch 16, total_step 260080, total loss is 16.66 , inference loss is 10.27, weight deacy loss is 6.38, training accuracy is 0.312500, time 225.019 samples/sec
+epoch 16, total_step 260100, total loss is 14.68 , inference loss is 8.30, weight deacy loss is 6.38, training accuracy is 0.500000, time 11.085 samples/sec
+epoch 16, total_step 260120, total loss is 12.54 , inference loss is 6.16, weight deacy loss is 6.38, training accuracy is 0.562500, time 214.825 samples/sec
+epoch 16, total_step 260140, total loss is 17.03 , inference loss is 10.65, weight deacy loss is 6.38, training accuracy is 0.281250, time 211.543 samples/sec
+epoch 16, total_step 260160, total loss is 11.24 , inference loss is 4.86, weight deacy loss is 6.38, training accuracy is 0.437500, time 211.089 samples/sec
+epoch 16, total_step 260180, total loss is 10.93 , inference loss is 4.54, weight deacy loss is 6.38, training accuracy is 0.437500, time 224.356 samples/sec
+epoch 16, total_step 260200, total loss is 13.09 , inference loss is 6.71, weight deacy loss is 6.38, training accuracy is 0.468750, time 225.688 samples/sec
+epoch 16, total_step 260220, total loss is 13.02 , inference loss is 6.64, weight deacy loss is 6.38, training accuracy is 0.437500, time 223.076 samples/sec
+epoch 16, total_step 260240, total loss is 12.10 , inference loss is 5.71, weight deacy loss is 6.38, training accuracy is 0.531250, time 205.503 samples/sec
+epoch 16, total_step 260260, total loss is 12.80 , inference loss is 6.42, weight deacy loss is 6.38, training accuracy is 0.437500, time 202.075 samples/sec
+epoch 16, total_step 260280, total loss is 12.15 , inference loss is 5.77, weight deacy loss is 6.38, training accuracy is 0.437500, time 223.921 samples/sec
+epoch 16, total_step 260300, total loss is 14.81 , inference loss is 8.42, weight deacy loss is 6.38, training accuracy is 0.406250, time 224.555 samples/sec
+epoch 16, total_step 260320, total loss is 14.89 , inference loss is 8.50, weight deacy loss is 6.38, training accuracy is 0.468750, time 208.152 samples/sec
+epoch 16, total_step 260340, total loss is 13.68 , inference loss is 7.30, weight deacy loss is 6.38, training accuracy is 0.406250, time 203.369 samples/sec2022-12-01 17:40:50.126030: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 9407 of 10000
+2022-12-01 17:40:50.710858: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 16, total_step 260360, total loss is 13.35 , inference loss is 6.97, weight deacy loss is 6.38, training accuracy is 0.531250, time 189.725 samples/sec
+epoch 16, total_step 260380, total loss is 17.40 , inference loss is 11.02, weight deacy loss is 6.38, training accuracy is 0.437500, time 193.599 samples/sec
+epoch 16, total_step 260400, total loss is 14.99 , inference loss is 8.61, weight deacy loss is 6.38, training accuracy is 0.468750, time 8.509 samples/sec
+epoch 16, total_step 260420, total loss is 17.41 , inference loss is 11.03, weight deacy loss is 6.38, training accuracy is 0.250000, time 217.457 samples/sec
+epoch 16, total_step 260440, total loss is 12.21 , inference loss is 5.83, weight deacy loss is 6.38, training accuracy is 0.593750, time 187.627 samples/sec
+epoch 16, total_step 260460, total loss is 18.76 , inference loss is 12.37, weight deacy loss is 6.38, training accuracy is 0.468750, time 183.809 samples/sec
+epoch 16, total_step 260480, total loss is 12.22 , inference loss is 5.84, weight deacy loss is 6.38, training accuracy is 0.468750, time 174.690 samples/sec
+epoch 16, total_step 260500, total loss is 14.27 , inference loss is 7.89, weight deacy loss is 6.38, training accuracy is 0.500000, time 188.604 samples/sec
+epoch 16, total_step 260520, total loss is 14.68 , inference loss is 8.30, weight deacy loss is 6.38, training accuracy is 0.500000, time 196.788 samples/sec
+epoch 16, total_step 260540, total loss is 12.34 , inference loss is 5.96, weight deacy loss is 6.38, training accuracy is 0.468750, time 199.209 samples/sec
+epoch 16, total_step 260560, total loss is 16.71 , inference loss is 10.33, weight deacy loss is 6.38, training accuracy is 0.281250, time 200.547 samples/sec
+epoch 16, total_step 260580, total loss is 14.90 , inference loss is 8.51, weight deacy loss is 6.38, training accuracy is 0.437500, time 191.285 samples/sec
+epoch 16, total_step 260600, total loss is 15.31 , inference loss is 8.93, weight deacy loss is 6.38, training accuracy is 0.468750, time 200.939 samples/sec
+epoch 16, total_step 260620, total loss is 11.85 , inference loss is 5.47, weight deacy loss is 6.38, training accuracy is 0.437500, time 191.580 samples/sec
+End of epoch 16
+epoch 17, total_step 260640, total loss is 10.63 , inference loss is 4.25, weight deacy loss is 6.38, training accuracy is 0.531250, time 204.167 samples/sec
+epoch 17, total_step 260660, total loss is 9.39 , inference loss is 3.01, weight deacy loss is 6.38, training accuracy is 0.718750, time 211.810 samples/sec
+epoch 17, total_step 260680, total loss is 8.65 , inference loss is 2.27, weight deacy loss is 6.38, training accuracy is 0.812500, time 200.805 samples/sec
+epoch 17, total_step 260700, total loss is 8.18 , inference loss is 1.80, weight deacy loss is 6.38, training accuracy is 0.843750, time 9.315 samples/sec
+epoch 17, total_step 260720, total loss is 11.20 , inference loss is 4.82, weight deacy loss is 6.38, training accuracy is 0.718750, time 195.659 samples/sec
+epoch 17, total_step 260740, total loss is 8.05 , inference loss is 1.66, weight deacy loss is 6.38, training accuracy is 0.718750, time 207.924 samples/sec
+epoch 17, total_step 260760, total loss is 10.56 , inference loss is 4.18, weight deacy loss is 6.38, training accuracy is 0.843750, time 201.757 samples/sec
+epoch 17, total_step 260780, total loss is 10.40 , inference loss is 4.02, weight deacy loss is 6.38, training accuracy is 0.843750, time 209.306 samples/sec
+epoch 17, total_step 260800, total loss is 11.82 , inference loss is 5.44, weight deacy loss is 6.38, training accuracy is 0.687500, time 204.919 samples/sec
+epoch 17, total_step 260820, total loss is 8.42 , inference loss is 2.04, weight deacy loss is 6.38, training accuracy is 0.843750, time 228.009 samples/sec
+epoch 17, total_step 260840, total loss is 7.40 , inference loss is 1.01, weight deacy loss is 6.38, training accuracy is 0.812500, time 298.919 samples/sec
+epoch 17, total_step 260860, total loss is 14.79 , inference loss is 8.41, weight deacy loss is 6.38, training accuracy is 0.718750, time 299.371 samples/sec
+epoch 17, total_step 260880, total loss is 11.34 , inference loss is 4.96, weight deacy loss is 6.38, training accuracy is 0.687500, time 246.019 samples/sec
+epoch 17, total_step 260900, total loss is 11.77 , inference loss is 5.39, weight deacy loss is 6.38, training accuracy is 0.687500, time 211.508 samples/sec
+epoch 17, total_step 260920, total loss is 11.21 , inference loss is 4.83, weight deacy loss is 6.38, training accuracy is 0.593750, time 215.821 samples/sec
+epoch 17, total_step 260940, total loss is 7.77 , inference loss is 1.39, weight deacy loss is 6.38, training accuracy is 0.875000, time 214.851 samples/sec
+epoch 17, total_step 260960, total loss is 9.84 , inference loss is 3.46, weight deacy loss is 6.38, training accuracy is 0.718750, time 210.655 samples/sec
+epoch 17, total_step 260980, total loss is 16.83 , inference loss is 10.45, weight deacy loss is 6.38, training accuracy is 0.593750, time 211.612 samples/sec
+epoch 17, total_step 261000, total loss is 9.84 , inference loss is 3.46, weight deacy loss is 6.38, training accuracy is 0.625000, time 9.884 samples/sec
+epoch 17, total_step 261020, total loss is 11.26 , inference loss is 4.88, weight deacy loss is 6.38, training accuracy is 0.656250, time 199.470 samples/sec
+epoch 17, total_step 261040, total loss is 13.96 , inference loss is 7.58, weight deacy loss is 6.38, training accuracy is 0.531250, time 201.130 samples/sec
+epoch 17, total_step 261060, total loss is 12.58 , inference loss is 6.20, weight deacy loss is 6.38, training accuracy is 0.750000, time 198.426 samples/sec
+epoch 17, total_step 261080, total loss is 8.19 , inference loss is 1.81, weight deacy loss is 6.38, training accuracy is 0.781250, time 215.362 samples/sec
+epoch 17, total_step 261100, total loss is 9.22 , inference loss is 2.84, weight deacy loss is 6.38, training accuracy is 0.781250, time 198.204 samples/sec
+epoch 17, total_step 261120, total loss is 15.19 , inference loss is 8.80, weight deacy loss is 6.38, training accuracy is 0.531250, time 204.718 samples/sec
+epoch 17, total_step 261140, total loss is 7.88 , inference loss is 1.50, weight deacy loss is 6.38, training accuracy is 0.812500, time 107.549 samples/sec
+epoch 17, total_step 261160, total loss is 10.42 , inference loss is 4.03, weight deacy loss is 6.38, training accuracy is 0.656250, time 207.537 samples/sec
+epoch 17, total_step 261180, total loss is 11.01 , inference loss is 4.63, weight deacy loss is 6.38, training accuracy is 0.718750, time 205.077 samples/sec
+epoch 17, total_step 261200, total loss is 12.56 , inference loss is 6.17, weight deacy loss is 6.38, training accuracy is 0.593750, time 196.763 samples/sec
+epoch 17, total_step 261220, total loss is 10.09 , inference loss is 3.71, weight deacy loss is 6.38, training accuracy is 0.812500, time 222.695 samples/sec
+epoch 17, total_step 261240, total loss is 11.74 , inference loss is 5.35, weight deacy loss is 6.38, training accuracy is 0.656250, time 200.049 samples/sec
+epoch 17, total_step 261260, total loss is 17.60 , inference loss is 11.21, weight deacy loss is 6.38, training accuracy is 0.500000, time 201.017 samples/sec
+epoch 17, total_step 261280, total loss is 15.49 , inference loss is 9.11, weight deacy loss is 6.38, training accuracy is 0.593750, time 210.128 samples/sec
+epoch 17, total_step 261300, total loss is 13.87 , inference loss is 7.48, weight deacy loss is 6.38, training accuracy is 0.687500, time 12.450 samples/sec
+epoch 17, total_step 261320, total loss is 11.46 , inference loss is 5.08, weight deacy loss is 6.39, training accuracy is 0.593750, time 208.962 samples/sec
+epoch 17, total_step 261340, total loss is 9.57 , inference loss is 3.19, weight deacy loss is 6.39, training accuracy is 0.781250, time 224.783 samples/sec
+epoch 17, total_step 261360, total loss is 11.75 , inference loss is 5.36, weight deacy loss is 6.39, training accuracy is 0.687500, time 214.995 samples/sec
+epoch 17, total_step 261380, total loss is 8.35 , inference loss is 1.97, weight deacy loss is 6.39, training accuracy is 0.687500, time 220.433 samples/sec
+epoch 17, total_step 261400, total loss is 11.22 , inference loss is 4.84, weight deacy loss is 6.39, training accuracy is 0.687500, time 217.457 samples/sec
+epoch 17, total_step 261420, total loss is 8.42 , inference loss is 2.03, weight deacy loss is 6.39, training accuracy is 0.750000, time 174.824 samples/sec
+epoch 17, total_step 261440, total loss is 10.74 , inference loss is 4.35, weight deacy loss is 6.39, training accuracy is 0.625000, time 212.929 samples/sec
+epoch 17, total_step 261460, total loss is 11.17 , inference loss is 4.78, weight deacy loss is 6.39, training accuracy is 0.625000, time 201.182 samples/sec
+epoch 17, total_step 261480, total loss is 11.63 , inference loss is 5.24, weight deacy loss is 6.39, training accuracy is 0.687500, time 182.182 samples/sec
+epoch 17, total_step 261500, total loss is 9.63 , inference loss is 3.24, weight deacy loss is 6.39, training accuracy is 0.781250, time 205.422 samples/sec
+epoch 17, total_step 261520, total loss is 9.80 , inference loss is 3.41, weight deacy loss is 6.39, training accuracy is 0.593750, time 206.797 samples/sec
+epoch 17, total_step 261540, total loss is 9.70 , inference loss is 3.31, weight deacy loss is 6.39, training accuracy is 0.750000, time 215.791 samples/sec
+epoch 17, total_step 261560, total loss is 14.44 , inference loss is 8.05, weight deacy loss is 6.39, training accuracy is 0.500000, time 297.321 samples/sec
+epoch 17, total_step 261580, total loss is 12.91 , inference loss is 6.53, weight deacy loss is 6.39, training accuracy is 0.500000, time 210.439 samples/sec
+epoch 17, total_step 261600, total loss is 19.95 , inference loss is 13.56, weight deacy loss is 6.39, training accuracy is 0.406250, time 9.083 samples/sec
+epoch 17, total_step 261620, total loss is 15.41 , inference loss is 9.02, weight deacy loss is 6.39, training accuracy is 0.437500, time 207.487 samples/sec
+epoch 17, total_step 261640, total loss is 10.00 , inference loss is 3.61, weight deacy loss is 6.39, training accuracy is 0.562500, time 212.836 samples/sec
+epoch 17, total_step 261660, total loss is 12.59 , inference loss is 6.20, weight deacy loss is 6.39, training accuracy is 0.562500, time 213.927 samples/sec
+epoch 17, total_step 261680, total loss is 15.31 , inference loss is 8.92, weight deacy loss is 6.39, training accuracy is 0.468750, time 212.273 samples/sec
+epoch 17, total_step 261700, total loss is 11.06 , inference loss is 4.67, weight deacy loss is 6.39, training accuracy is 0.562500, time 219.985 samples/sec
+epoch 17, total_step 261720, total loss is 13.95 , inference loss is 7.56, weight deacy loss is 6.39, training accuracy is 0.562500, time 211.285 samples/sec
+epoch 17, total_step 261740, total loss is 8.27 , inference loss is 1.88, weight deacy loss is 6.39, training accuracy is 0.750000, time 209.373 samples/sec
+epoch 17, total_step 261760, total loss is 14.93 , inference loss is 8.53, weight deacy loss is 6.39, training accuracy is 0.656250, time 218.802 samples/sec
+epoch 17, total_step 261780, total loss is 12.46 , inference loss is 6.06, weight deacy loss is 6.39, training accuracy is 0.593750, time 206.381 samples/sec
+epoch 17, total_step 261800, total loss is 11.43 , inference loss is 5.04, weight deacy loss is 6.39, training accuracy is 0.625000, time 294.857 samples/sec
+epoch 17, total_step 261820, total loss is 15.75 , inference loss is 9.35, weight deacy loss is 6.39, training accuracy is 0.500000, time 300.156 samples/sec
+epoch 17, total_step 261840, total loss is 11.72 , inference loss is 5.33, weight deacy loss is 6.39, training accuracy is 0.593750, time 226.610 samples/sec
+epoch 17, total_step 261860, total loss is 10.44 , inference loss is 4.05, weight deacy loss is 6.39, training accuracy is 0.656250, time 225.562 samples/sec
+epoch 17, total_step 261880, total loss is 9.59 , inference loss is 3.19, weight deacy loss is 6.39, training accuracy is 0.718750, time 207.924 samples/sec
+epoch 17, total_step 261900, total loss is 12.51 , inference loss is 6.12, weight deacy loss is 6.39, training accuracy is 0.687500, time 12.975 samples/sec
+epoch 17, total_step 261920, total loss is 10.95 , inference loss is 4.56, weight deacy loss is 6.40, training accuracy is 0.781250, time 217.025 samples/sec
+epoch 17, total_step 261940, total loss is 10.53 , inference loss is 4.14, weight deacy loss is 6.40, training accuracy is 0.750000, time 212.211 samples/sec
+epoch 17, total_step 261960, total loss is 11.98 , inference loss is 5.58, weight deacy loss is 6.40, training accuracy is 0.718750, time 217.935 samples/sec
+epoch 17, total_step 261980, total loss is 9.14 , inference loss is 2.74, weight deacy loss is 6.40, training accuracy is 0.718750, time 229.705 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.893060999999964
+best_threshold_index 116 0.985
+best_threshold_index 120 0.9846296296296296
+best_threshold_index 116 0.9842592592592593
+best_threshold_index 116 0.9855555555555555
+best_threshold_index 116 0.9846296296296296
+best_threshold_index 116 0.9833333333333333
+best_threshold_index 116 0.9853703703703703
+best_threshold_index 120 0.9846296296296296
+best_threshold_index 120 0.9844444444444445
+best_threshold_index 116 0.9838888888888889
+[lfw][262000]XNorm: 18.990887
+[lfw][262000]Accuracy-Flip: 0.98383+-0.00563
+test accuracy is: 0.9838333333333333
+epoch 17, total_step 262000, total loss is 11.17 , inference loss is 4.77, weight deacy loss is 6.40, training accuracy is 0.656250, time 16.338 samples/sec
+epoch 17, total_step 262020, total loss is 12.38 , inference loss is 5.99, weight deacy loss is 6.40, training accuracy is 0.437500, time 222.134 samples/sec
+epoch 17, total_step 262040, total loss is 15.50 , inference loss is 9.10, weight deacy loss is 6.40, training accuracy is 0.656250, time 220.326 samples/sec
+epoch 17, total_step 262060, total loss is 11.68 , inference loss is 5.28, weight deacy loss is 6.40, training accuracy is 0.687500, time 226.516 samples/sec
+epoch 17, total_step 262080, total loss is 9.92 , inference loss is 3.52, weight deacy loss is 6.40, training accuracy is 0.656250, time 220.688 samples/sec
+epoch 17, total_step 262100, total loss is 10.75 , inference loss is 4.36, weight deacy loss is 6.40, training accuracy is 0.625000, time 232.410 samples/sec
+epoch 17, total_step 262120, total loss is 15.00 , inference loss is 8.60, weight deacy loss is 6.40, training accuracy is 0.625000, time 211.654 samples/sec
+epoch 17, total_step 262140, total loss is 10.91 , inference loss is 4.52, weight deacy loss is 6.40, training accuracy is 0.750000, time 218.973 samples/sec
+epoch 17, total_step 262160, total loss is 9.41 , inference loss is 3.01, weight deacy loss is 6.40, training accuracy is 0.656250, time 197.956 samples/sec
+epoch 17, total_step 262180, total loss is 12.21 , inference loss is 5.82, weight deacy loss is 6.40, training accuracy is 0.562500, time 199.473 samples/sec
+epoch 17, total_step 262200, total loss is 14.20 , inference loss is 7.80, weight deacy loss is 6.40, training accuracy is 0.562500, time 12.630 samples/sec
+epoch 17, total_step 262220, total loss is 9.75 , inference loss is 3.35, weight deacy loss is 6.40, training accuracy is 0.656250, time 226.757 samples/sec
+epoch 17, total_step 262240, total loss is 12.53 , inference loss is 6.13, weight deacy loss is 6.40, training accuracy is 0.687500, time 217.271 samples/sec
+epoch 17, total_step 262260, total loss is 13.42 , inference loss is 7.02, weight deacy loss is 6.40, training accuracy is 0.625000, time 230.786 samples/sec
+epoch 17, total_step 262280, total loss is 7.77 , inference loss is 1.37, weight deacy loss is 6.40, training accuracy is 0.812500, time 223.598 samples/sec
+epoch 17, total_step 262300, total loss is 14.91 , inference loss is 8.51, weight deacy loss is 6.40, training accuracy is 0.593750, time 213.625 samples/sec
+epoch 17, total_step 262320, total loss is 16.35 , inference loss is 9.95, weight deacy loss is 6.40, training accuracy is 0.531250, time 214.029 samples/sec
+epoch 17, total_step 262340, total loss is 11.11 , inference loss is 4.70, weight deacy loss is 6.40, training accuracy is 0.593750, time 218.123 samples/sec
+epoch 17, total_step 262360, total loss is 12.86 , inference loss is 6.46, weight deacy loss is 6.40, training accuracy is 0.625000, time 229.746 samples/sec
+epoch 17, total_step 262380, total loss is 11.67 , inference loss is 5.27, weight deacy loss is 6.40, training accuracy is 0.500000, time 218.886 samples/sec
+epoch 17, total_step 262400, total loss is 14.56 , inference loss is 8.16, weight deacy loss is 6.40, training accuracy is 0.593750, time 213.046 samples/sec
+epoch 17, total_step 262420, total loss is 12.39 , inference loss is 5.99, weight deacy loss is 6.40, training accuracy is 0.687500, time 220.204 samples/sec
+epoch 17, total_step 262440, total loss is 12.50 , inference loss is 6.10, weight deacy loss is 6.40, training accuracy is 0.687500, time 237.073 samples/sec
+epoch 17, total_step 262460, total loss is 13.28 , inference loss is 6.87, weight deacy loss is 6.40, training accuracy is 0.593750, time 299.169 samples/sec
+epoch 17, total_step 262480, total loss is 11.66 , inference loss is 5.26, weight deacy loss is 6.40, training accuracy is 0.625000, time 217.598 samples/sec
+epoch 17, total_step 262500, total loss is 11.99 , inference loss is 5.59, weight deacy loss is 6.40, training accuracy is 0.625000, time 13.289 samples/sec
+epoch 17, total_step 262520, total loss is 9.54 , inference loss is 3.13, weight deacy loss is 6.40, training accuracy is 0.687500, time 220.816 samples/sec
+epoch 17, total_step 262540, total loss is 10.47 , inference loss is 4.07, weight deacy loss is 6.40, training accuracy is 0.656250, time 218.632 samples/sec
+epoch 17, total_step 262560, total loss is 10.19 , inference loss is 3.79, weight deacy loss is 6.40, training accuracy is 0.531250, time 217.373 samples/sec
+epoch 17, total_step 262580, total loss is 10.43 , inference loss is 4.03, weight deacy loss is 6.40, training accuracy is 0.812500, time 225.633 samples/sec
+epoch 17, total_step 262600, total loss is 10.98 , inference loss is 4.58, weight deacy loss is 6.40, training accuracy is 0.656250, time 229.498 samples/sec
+epoch 17, total_step 262620, total loss is 11.48 , inference loss is 5.07, weight deacy loss is 6.40, training accuracy is 0.656250, time 214.394 samples/sec
+epoch 17, total_step 262640, total loss is 13.41 , inference loss is 7.01, weight deacy loss is 6.40, training accuracy is 0.468750, time 211.158 samples/sec
+epoch 17, total_step 262660, total loss is 12.30 , inference loss is 5.89, weight deacy loss is 6.40, training accuracy is 0.562500, time 212.257 samples/sec
+epoch 17, total_step 262680, total loss is 15.36 , inference loss is 8.96, weight deacy loss is 6.40, training accuracy is 0.593750, time 224.228 samples/sec
+epoch 17, total_step 262700, total loss is 11.30 , inference loss is 4.90, weight deacy loss is 6.40, training accuracy is 0.718750, time 215.224 samples/sec
+epoch 17, total_step 262720, total loss is 10.38 , inference loss is 3.98, weight deacy loss is 6.40, training accuracy is 0.593750, time 216.100 samples/sec
+epoch 17, total_step 262740, total loss is 13.76 , inference loss is 7.36, weight deacy loss is 6.40, training accuracy is 0.562500, time 210.867 samples/sec
+epoch 17, total_step 262760, total loss is 12.32 , inference loss is 5.92, weight deacy loss is 6.40, training accuracy is 0.687500, time 226.438 samples/sec
+epoch 17, total_step 262780, total loss is 10.54 , inference loss is 4.14, weight deacy loss is 6.40, training accuracy is 0.593750, time 298.177 samples/sec
+epoch 17, total_step 262800, total loss is 11.46 , inference loss is 5.06, weight deacy loss is 6.40, training accuracy is 0.656250, time 12.171 samples/sec
+epoch 17, total_step 262820, total loss is 11.26 , inference loss is 4.86, weight deacy loss is 6.40, training accuracy is 0.656250, time 226.368 samples/sec
+epoch 17, total_step 262840, total loss is 9.63 , inference loss is 3.23, weight deacy loss is 6.40, training accuracy is 0.718750, time 227.944 samples/sec
+epoch 17, total_step 262860, total loss is 8.65 , inference loss is 2.25, weight deacy loss is 6.40, training accuracy is 0.718750, time 299.803 samples/sec
+epoch 17, total_step 262880, total loss is 10.06 , inference loss is 3.66, weight deacy loss is 6.40, training accuracy is 0.718750, time 216.434 samples/sec
+epoch 17, total_step 262900, total loss is 11.29 , inference loss is 4.88, weight deacy loss is 6.40, training accuracy is 0.500000, time 230.694 samples/sec
+epoch 17, total_step 262920, total loss is 10.02 , inference loss is 3.62, weight deacy loss is 6.40, training accuracy is 0.750000, time 213.414 samples/sec
+epoch 17, total_step 262940, total loss is 8.42 , inference loss is 2.02, weight deacy loss is 6.40, training accuracy is 0.750000, time 216.283 samples/sec
+epoch 17, total_step 262960, total loss is 10.01 , inference loss is 3.61, weight deacy loss is 6.40, training accuracy is 0.750000, time 211.475 samples/sec
+epoch 17, total_step 262980, total loss is 10.03 , inference loss is 3.62, weight deacy loss is 6.40, training accuracy is 0.718750, time 224.081 samples/sec
+epoch 17, total_step 263000, total loss is 10.15 , inference loss is 3.74, weight deacy loss is 6.40, training accuracy is 0.562500, time 212.626 samples/sec
+epoch 17, total_step 263020, total loss is 10.07 , inference loss is 3.67, weight deacy loss is 6.40, training accuracy is 0.656250, time 218.435 samples/sec
+epoch 17, total_step 263040, total loss is 8.94 , inference loss is 2.54, weight deacy loss is 6.40, training accuracy is 0.718750, time 219.556 samples/sec
+epoch 17, total_step 263060, total loss is 10.28 , inference loss is 3.88, weight deacy loss is 6.40, training accuracy is 0.687500, time 213.756 samples/sec
+epoch 17, total_step 263080, total loss is 10.03 , inference loss is 3.63, weight deacy loss is 6.40, training accuracy is 0.656250, time 226.580 samples/sec
+epoch 17, total_step 263100, total loss is 10.22 , inference loss is 3.82, weight deacy loss is 6.40, training accuracy is 0.750000, time 12.320 samples/sec
+epoch 17, total_step 263120, total loss is 10.24 , inference loss is 3.84, weight deacy loss is 6.40, training accuracy is 0.687500, time 299.418 samples/sec
+epoch 17, total_step 263140, total loss is 9.82 , inference loss is 3.42, weight deacy loss is 6.40, training accuracy is 0.625000, time 222.159 samples/sec
+epoch 17, total_step 263160, total loss is 10.72 , inference loss is 4.32, weight deacy loss is 6.40, training accuracy is 0.562500, time 213.861 samples/sec
+epoch 17, total_step 263180, total loss is 8.94 , inference loss is 2.54, weight deacy loss is 6.40, training accuracy is 0.656250, time 220.182 samples/sec
+epoch 17, total_step 263200, total loss is 9.51 , inference loss is 3.11, weight deacy loss is 6.40, training accuracy is 0.750000, time 224.546 samples/sec
+epoch 17, total_step 263220, total loss is 9.87 , inference loss is 3.47, weight deacy loss is 6.40, training accuracy is 0.750000, time 224.525 samples/sec
+epoch 17, total_step 263240, total loss is 7.77 , inference loss is 1.37, weight deacy loss is 6.40, training accuracy is 0.687500, time 213.368 samples/sec
+epoch 17, total_step 263260, total loss is 13.02 , inference loss is 6.62, weight deacy loss is 6.40, training accuracy is 0.781250, time 228.162 samples/sec
+epoch 17, total_step 263280, total loss is 10.27 , inference loss is 3.87, weight deacy loss is 6.40, training accuracy is 0.687500, time 217.502 samples/sec
+epoch 17, total_step 263300, total loss is 11.13 , inference loss is 4.73, weight deacy loss is 6.40, training accuracy is 0.593750, time 232.092 samples/sec
+epoch 17, total_step 263320, total loss is 13.22 , inference loss is 6.82, weight deacy loss is 6.40, training accuracy is 0.687500, time 215.723 samples/sec
+epoch 17, total_step 263340, total loss is 9.89 , inference loss is 3.49, weight deacy loss is 6.40, training accuracy is 0.562500, time 226.561 samples/sec
+epoch 17, total_step 263360, total loss is 10.87 , inference loss is 4.47, weight deacy loss is 6.40, training accuracy is 0.656250, time 219.251 samples/sec
+epoch 17, total_step 263380, total loss is 8.50 , inference loss is 2.10, weight deacy loss is 6.40, training accuracy is 0.718750, time 213.387 samples/sec
+epoch 17, total_step 263400, total loss is 10.63 , inference loss is 4.23, weight deacy loss is 6.40, training accuracy is 0.562500, time 12.892 samples/sec
+epoch 17, total_step 263420, total loss is 9.94 , inference loss is 3.54, weight deacy loss is 6.40, training accuracy is 0.656250, time 216.374 samples/sec
+epoch 17, total_step 263440, total loss is 9.06 , inference loss is 2.66, weight deacy loss is 6.40, training accuracy is 0.750000, time 221.586 samples/sec
+epoch 17, total_step 263460, total loss is 7.00 , inference loss is 0.60, weight deacy loss is 6.40, training accuracy is 0.843750, time 225.747 samples/sec
+epoch 17, total_step 263480, total loss is 9.59 , inference loss is 3.19, weight deacy loss is 6.40, training accuracy is 0.781250, time 218.612 samples/sec
+epoch 17, total_step 263500, total loss is 9.86 , inference loss is 3.46, weight deacy loss is 6.40, training accuracy is 0.718750, time 217.391 samples/sec
+epoch 17, total_step 263520, total loss is 8.75 , inference loss is 2.35, weight deacy loss is 6.40, training accuracy is 0.687500, time 221.967 samples/sec
+epoch 17, total_step 263540, total loss is 11.99 , inference loss is 5.59, weight deacy loss is 6.40, training accuracy is 0.687500, time 211.124 samples/sec
+epoch 17, total_step 263560, total loss is 12.11 , inference loss is 5.71, weight deacy loss is 6.40, training accuracy is 0.562500, time 223.520 samples/sec
+epoch 17, total_step 263580, total loss is 11.90 , inference loss is 5.50, weight deacy loss is 6.40, training accuracy is 0.562500, time 228.394 samples/sec
+epoch 17, total_step 263600, total loss is 10.43 , inference loss is 4.03, weight deacy loss is 6.40, training accuracy is 0.718750, time 230.905 samples/sec
+epoch 17, total_step 263620, total loss is 8.69 , inference loss is 2.29, weight deacy loss is 6.40, training accuracy is 0.812500, time 221.093 samples/sec
+epoch 17, total_step 263640, total loss is 9.97 , inference loss is 3.57, weight deacy loss is 6.40, training accuracy is 0.718750, time 224.486 samples/sec
+epoch 17, total_step 263660, total loss is 10.65 , inference loss is 4.25, weight deacy loss is 6.40, training accuracy is 0.625000, time 213.851 samples/sec
+epoch 17, total_step 263680, total loss is 9.50 , inference loss is 3.10, weight deacy loss is 6.40, training accuracy is 0.625000, time 218.787 samples/sec
+epoch 17, total_step 263700, total loss is 9.58 , inference loss is 3.18, weight deacy loss is 6.40, training accuracy is 0.718750, time 13.557 samples/sec
+epoch 17, total_step 263720, total loss is 8.74 , inference loss is 2.34, weight deacy loss is 6.40, training accuracy is 0.812500, time 219.034 samples/sec
+epoch 17, total_step 263740, total loss is 10.56 , inference loss is 4.16, weight deacy loss is 6.40, training accuracy is 0.562500, time 229.638 samples/sec
+epoch 17, total_step 263760, total loss is 9.31 , inference loss is 2.91, weight deacy loss is 6.40, training accuracy is 0.656250, time 215.659 samples/sec
+epoch 17, total_step 263780, total loss is 11.83 , inference loss is 5.43, weight deacy loss is 6.40, training accuracy is 0.687500, time 229.014 samples/sec
+epoch 17, total_step 263800, total loss is 11.29 , inference loss is 4.89, weight deacy loss is 6.40, training accuracy is 0.687500, time 212.086 samples/sec
+epoch 17, total_step 263820, total loss is 13.18 , inference loss is 6.78, weight deacy loss is 6.40, training accuracy is 0.562500, time 214.999 samples/sec
+epoch 17, total_step 263840, total loss is 11.10 , inference loss is 4.70, weight deacy loss is 6.40, training accuracy is 0.656250, time 222.617 samples/sec
+epoch 17, total_step 263860, total loss is 15.00 , inference loss is 8.60, weight deacy loss is 6.40, training accuracy is 0.656250, time 211.164 samples/sec
+epoch 17, total_step 263880, total loss is 10.00 , inference loss is 3.60, weight deacy loss is 6.40, training accuracy is 0.625000, time 209.464 samples/sec
+epoch 17, total_step 263900, total loss is 15.54 , inference loss is 9.13, weight deacy loss is 6.40, training accuracy is 0.531250, time 217.391 samples/sec
+epoch 17, total_step 263920, total loss is 11.03 , inference loss is 4.63, weight deacy loss is 6.40, training accuracy is 0.562500, time 213.294 samples/sec
+epoch 17, total_step 263940, total loss is 17.34 , inference loss is 10.94, weight deacy loss is 6.40, training accuracy is 0.437500, time 213.163 samples/sec
+epoch 17, total_step 263960, total loss is 13.34 , inference loss is 6.94, weight deacy loss is 6.40, training accuracy is 0.593750, time 181.670 samples/sec
+epoch 17, total_step 263980, total loss is 9.30 , inference loss is 2.89, weight deacy loss is 6.40, training accuracy is 0.656250, time 203.660 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.984782000000017
+best_threshold_index 124 0.9875925925925926
+best_threshold_index 124 0.9872222222222222
+best_threshold_index 124 0.9881481481481481
+best_threshold_index 123 0.9885185185185185
+best_threshold_index 124 0.9872222222222222
+best_threshold_index 124 0.9874074074074074
+best_threshold_index 124 0.9887037037037038
+best_threshold_index 124 0.987037037037037
+best_threshold_index 124 0.9864814814814815
+best_threshold_index 124 0.9866666666666667
+[lfw][264000]XNorm: 22.130629
+[lfw][264000]Accuracy-Flip: 0.98717+-0.00691
+test accuracy is: 0.9871666666666664
+epoch 17, total_step 264000, total loss is 13.70 , inference loss is 7.30, weight deacy loss is 6.40, training accuracy is 0.500000, time 19.141 samples/sec
+epoch 17, total_step 264020, total loss is 11.91 , inference loss is 5.50, weight deacy loss is 6.40, training accuracy is 0.687500, time 299.421 samples/sec
+epoch 17, total_step 264040, total loss is 11.74 , inference loss is 5.34, weight deacy loss is 6.40, training accuracy is 0.468750, time 217.585 samples/sec
+epoch 17, total_step 264060, total loss is 14.82 , inference loss is 8.42, weight deacy loss is 6.40, training accuracy is 0.468750, time 212.948 samples/sec
+epoch 17, total_step 264080, total loss is 10.50 , inference loss is 4.10, weight deacy loss is 6.40, training accuracy is 0.500000, time 219.913 samples/sec
+epoch 17, total_step 264100, total loss is 13.63 , inference loss is 7.23, weight deacy loss is 6.40, training accuracy is 0.468750, time 224.830 samples/sec
+epoch 17, total_step 264120, total loss is 14.56 , inference loss is 8.15, weight deacy loss is 6.40, training accuracy is 0.437500, time 217.812 samples/sec
+epoch 17, total_step 264140, total loss is 11.12 , inference loss is 4.72, weight deacy loss is 6.40, training accuracy is 0.593750, time 235.066 samples/sec
+epoch 17, total_step 264160, total loss is 10.11 , inference loss is 3.71, weight deacy loss is 6.40, training accuracy is 0.625000, time 225.585 samples/sec
+epoch 17, total_step 264180, total loss is 13.41 , inference loss is 7.01, weight deacy loss is 6.40, training accuracy is 0.562500, time 219.449 samples/sec
+epoch 17, total_step 264200, total loss is 13.26 , inference loss is 6.86, weight deacy loss is 6.40, training accuracy is 0.656250, time 233.708 samples/sec
+epoch 17, total_step 264220, total loss is 11.93 , inference loss is 5.53, weight deacy loss is 6.40, training accuracy is 0.562500, time 231.369 samples/sec
+epoch 17, total_step 264240, total loss is 14.01 , inference loss is 7.61, weight deacy loss is 6.40, training accuracy is 0.625000, time 220.570 samples/sec
+epoch 17, total_step 264260, total loss is 10.75 , inference loss is 4.34, weight deacy loss is 6.40, training accuracy is 0.468750, time 220.592 samples/sec
+epoch 17, total_step 264280, total loss is 11.84 , inference loss is 5.44, weight deacy loss is 6.40, training accuracy is 0.437500, time 224.520 samples/sec
+epoch 17, total_step 264300, total loss is 10.73 , inference loss is 4.33, weight deacy loss is 6.40, training accuracy is 0.468750, time 12.869 samples/sec
+epoch 17, total_step 264320, total loss is 11.94 , inference loss is 5.54, weight deacy loss is 6.40, training accuracy is 0.531250, time 220.446 samples/sec
+epoch 17, total_step 264340, total loss is 12.64 , inference loss is 6.23, weight deacy loss is 6.40, training accuracy is 0.625000, time 211.864 samples/sec
+epoch 17, total_step 264360, total loss is 13.06 , inference loss is 6.66, weight deacy loss is 6.40, training accuracy is 0.593750, time 218.284 samples/sec
+epoch 17, total_step 264380, total loss is 14.37 , inference loss is 7.97, weight deacy loss is 6.40, training accuracy is 0.562500, time 219.328 samples/sec
+epoch 17, total_step 264400, total loss is 14.63 , inference loss is 8.23, weight deacy loss is 6.40, training accuracy is 0.625000, time 216.632 samples/sec
+epoch 17, total_step 264420, total loss is 13.71 , inference loss is 7.30, weight deacy loss is 6.40, training accuracy is 0.593750, time 215.085 samples/sec
+epoch 17, total_step 264440, total loss is 12.02 , inference loss is 5.61, weight deacy loss is 6.40, training accuracy is 0.562500, time 220.951 samples/sec
+epoch 17, total_step 264460, total loss is 8.19 , inference loss is 1.79, weight deacy loss is 6.40, training accuracy is 0.781250, time 226.542 samples/sec
+epoch 17, total_step 264480, total loss is 11.49 , inference loss is 5.09, weight deacy loss is 6.40, training accuracy is 0.468750, time 217.354 samples/sec
+epoch 17, total_step 264500, total loss is 14.07 , inference loss is 7.67, weight deacy loss is 6.40, training accuracy is 0.500000, time 214.586 samples/sec
+epoch 17, total_step 264520, total loss is 12.93 , inference loss is 6.53, weight deacy loss is 6.40, training accuracy is 0.593750, time 227.754 samples/sec
+epoch 17, total_step 264540, total loss is 13.17 , inference loss is 6.77, weight deacy loss is 6.40, training accuracy is 0.468750, time 217.425 samples/sec
+epoch 17, total_step 264560, total loss is 14.84 , inference loss is 8.43, weight deacy loss is 6.40, training accuracy is 0.437500, time 299.665 samples/sec
+epoch 17, total_step 264580, total loss is 10.38 , inference loss is 3.98, weight deacy loss is 6.40, training accuracy is 0.562500, time 299.823 samples/sec
+epoch 17, total_step 264600, total loss is 10.66 , inference loss is 4.26, weight deacy loss is 6.40, training accuracy is 0.531250, time 8.092 samples/sec
+epoch 17, total_step 264620, total loss is 10.88 , inference loss is 4.48, weight deacy loss is 6.40, training accuracy is 0.593750, time 202.658 samples/sec
+epoch 17, total_step 264640, total loss is 14.88 , inference loss is 8.48, weight deacy loss is 6.40, training accuracy is 0.562500, time 206.538 samples/sec
+epoch 17, total_step 264660, total loss is 10.34 , inference loss is 3.93, weight deacy loss is 6.40, training accuracy is 0.625000, time 218.458 samples/sec
+epoch 17, total_step 264680, total loss is 11.05 , inference loss is 4.65, weight deacy loss is 6.40, training accuracy is 0.562500, time 215.807 samples/sec
+epoch 17, total_step 264700, total loss is 10.18 , inference loss is 3.77, weight deacy loss is 6.40, training accuracy is 0.562500, time 202.086 samples/sec
+epoch 17, total_step 264720, total loss is 12.61 , inference loss is 6.21, weight deacy loss is 6.40, training accuracy is 0.375000, time 298.211 samples/sec
+epoch 17, total_step 264740, total loss is 12.82 , inference loss is 6.42, weight deacy loss is 6.40, training accuracy is 0.500000, time 299.670 samples/sec
+epoch 17, total_step 264760, total loss is 14.06 , inference loss is 7.65, weight deacy loss is 6.40, training accuracy is 0.562500, time 223.225 samples/sec
+epoch 17, total_step 264780, total loss is 11.42 , inference loss is 5.01, weight deacy loss is 6.40, training accuracy is 0.562500, time 215.778 samples/sec
+epoch 17, total_step 264800, total loss is 14.59 , inference loss is 8.19, weight deacy loss is 6.40, training accuracy is 0.468750, time 213.032 samples/sec
+epoch 17, total_step 264820, total loss is 16.06 , inference loss is 9.66, weight deacy loss is 6.40, training accuracy is 0.468750, time 212.335 samples/sec
+epoch 17, total_step 264840, total loss is 17.40 , inference loss is 10.99, weight deacy loss is 6.40, training accuracy is 0.500000, time 214.676 samples/sec
+epoch 17, total_step 264860, total loss is 10.26 , inference loss is 3.86, weight deacy loss is 6.40, training accuracy is 0.625000, time 213.565 samples/sec
+epoch 17, total_step 264880, total loss is 12.15 , inference loss is 5.75, weight deacy loss is 6.40, training accuracy is 0.468750, time 216.815 samples/sec
+epoch 17, total_step 264900, total loss is 12.22 , inference loss is 5.81, weight deacy loss is 6.40, training accuracy is 0.468750, time 6.785 samples/sec
+epoch 17, total_step 264920, total loss is 13.20 , inference loss is 6.79, weight deacy loss is 6.40, training accuracy is 0.406250, time 218.428 samples/sec
+epoch 17, total_step 264940, total loss is 15.15 , inference loss is 8.74, weight deacy loss is 6.40, training accuracy is 0.343750, time 203.440 samples/sec
+epoch 17, total_step 264960, total loss is 11.96 , inference loss is 5.56, weight deacy loss is 6.40, training accuracy is 0.687500, time 197.095 samples/sec
+epoch 17, total_step 264980, total loss is 14.07 , inference loss is 7.66, weight deacy loss is 6.40, training accuracy is 0.562500, time 299.740 samples/sec
+epoch 17, total_step 265000, total loss is 15.34 , inference loss is 8.94, weight deacy loss is 6.40, training accuracy is 0.500000, time 209.619 samples/sec
+epoch 17, total_step 265020, total loss is 12.35 , inference loss is 5.95, weight deacy loss is 6.40, training accuracy is 0.625000, time 208.998 samples/sec
+epoch 17, total_step 265040, total loss is 10.41 , inference loss is 4.00, weight deacy loss is 6.40, training accuracy is 0.687500, time 202.262 samples/sec
+epoch 17, total_step 265060, total loss is 17.42 , inference loss is 11.02, weight deacy loss is 6.40, training accuracy is 0.562500, time 194.951 samples/sec
+epoch 17, total_step 265080, total loss is 13.06 , inference loss is 6.66, weight deacy loss is 6.40, training accuracy is 0.468750, time 226.568 samples/sec
+epoch 17, total_step 265100, total loss is 11.21 , inference loss is 4.80, weight deacy loss is 6.40, training accuracy is 0.562500, time 194.158 samples/sec
+epoch 17, total_step 265120, total loss is 12.70 , inference loss is 6.29, weight deacy loss is 6.40, training accuracy is 0.562500, time 210.692 samples/sec
+epoch 17, total_step 265140, total loss is 11.30 , inference loss is 4.89, weight deacy loss is 6.40, training accuracy is 0.437500, time 207.465 samples/sec
+epoch 17, total_step 265160, total loss is 17.25 , inference loss is 10.84, weight deacy loss is 6.40, training accuracy is 0.375000, time 207.123 samples/sec
+epoch 17, total_step 265180, total loss is 15.26 , inference loss is 8.85, weight deacy loss is 6.40, training accuracy is 0.500000, time 212.405 samples/sec
+epoch 17, total_step 265200, total loss is 11.94 , inference loss is 5.53, weight deacy loss is 6.40, training accuracy is 0.531250, time 12.977 samples/sec
+epoch 17, total_step 265220, total loss is 12.25 , inference loss is 5.84, weight deacy loss is 6.41, training accuracy is 0.531250, time 214.620 samples/sec
+epoch 17, total_step 265240, total loss is 13.84 , inference loss is 7.44, weight deacy loss is 6.41, training accuracy is 0.406250, time 187.024 samples/sec
+epoch 17, total_step 265260, total loss is 13.25 , inference loss is 6.84, weight deacy loss is 6.41, training accuracy is 0.468750, time 202.515 samples/sec
+epoch 17, total_step 265280, total loss is 10.29 , inference loss is 3.88, weight deacy loss is 6.41, training accuracy is 0.500000, time 207.741 samples/sec
+epoch 17, total_step 265300, total loss is 14.41 , inference loss is 8.00, weight deacy loss is 6.41, training accuracy is 0.562500, time 201.728 samples/sec
+epoch 17, total_step 265320, total loss is 13.43 , inference loss is 7.03, weight deacy loss is 6.41, training accuracy is 0.625000, time 216.636 samples/sec
+epoch 17, total_step 265340, total loss is 17.34 , inference loss is 10.94, weight deacy loss is 6.41, training accuracy is 0.468750, time 199.434 samples/sec
+epoch 17, total_step 265360, total loss is 13.41 , inference loss is 7.01, weight deacy loss is 6.41, training accuracy is 0.437500, time 220.295 samples/sec
+epoch 17, total_step 265380, total loss is 14.94 , inference loss is 8.53, weight deacy loss is 6.41, training accuracy is 0.468750, time 197.740 samples/sec
+epoch 17, total_step 265400, total loss is 14.36 , inference loss is 7.95, weight deacy loss is 6.41, training accuracy is 0.437500, time 206.147 samples/sec
+epoch 17, total_step 265420, total loss is 13.06 , inference loss is 6.65, weight deacy loss is 6.41, training accuracy is 0.406250, time 182.208 samples/sec
+epoch 17, total_step 265440, total loss is 12.39 , inference loss is 5.99, weight deacy loss is 6.41, training accuracy is 0.500000, time 205.160 samples/sec
+epoch 17, total_step 265460, total loss is 15.89 , inference loss is 9.48, weight deacy loss is 6.41, training accuracy is 0.375000, time 204.252 samples/sec
+epoch 17, total_step 265480, total loss is 12.30 , inference loss is 5.90, weight deacy loss is 6.41, training accuracy is 0.437500, time 220.382 samples/sec
+epoch 17, total_step 265500, total loss is 12.12 , inference loss is 5.71, weight deacy loss is 6.41, training accuracy is 0.625000, time 13.039 samples/sec
+epoch 17, total_step 265520, total loss is 12.45 , inference loss is 6.04, weight deacy loss is 6.41, training accuracy is 0.593750, time 201.001 samples/sec
+epoch 17, total_step 265540, total loss is 15.10 , inference loss is 8.69, weight deacy loss is 6.41, training accuracy is 0.468750, time 230.707 samples/sec
+epoch 17, total_step 265560, total loss is 9.93 , inference loss is 3.52, weight deacy loss is 6.41, training accuracy is 0.500000, time 205.240 samples/sec
+epoch 17, total_step 265580, total loss is 9.91 , inference loss is 3.50, weight deacy loss is 6.41, training accuracy is 0.625000, time 224.059 samples/sec
+epoch 17, total_step 265600, total loss is 13.44 , inference loss is 7.03, weight deacy loss is 6.41, training accuracy is 0.500000, time 214.442 samples/sec
+epoch 17, total_step 265620, total loss is 13.63 , inference loss is 7.23, weight deacy loss is 6.41, training accuracy is 0.468750, time 186.133 samples/sec
+epoch 17, total_step 265640, total loss is 11.71 , inference loss is 5.30, weight deacy loss is 6.41, training accuracy is 0.718750, time 207.242 samples/sec
+epoch 17, total_step 265660, total loss is 13.65 , inference loss is 7.25, weight deacy loss is 6.41, training accuracy is 0.500000, time 192.377 samples/sec
+epoch 17, total_step 265680, total loss is 10.13 , inference loss is 3.72, weight deacy loss is 6.41, training accuracy is 0.687500, time 214.846 samples/sec
+epoch 17, total_step 265700, total loss is 12.13 , inference loss is 5.73, weight deacy loss is 6.41, training accuracy is 0.656250, time 194.086 samples/sec
+epoch 17, total_step 265720, total loss is 13.42 , inference loss is 7.02, weight deacy loss is 6.41, training accuracy is 0.500000, time 210.727 samples/sec
+epoch 17, total_step 265740, total loss is 15.84 , inference loss is 9.44, weight deacy loss is 6.41, training accuracy is 0.468750, time 215.874 samples/sec
+epoch 17, total_step 265760, total loss is 9.77 , inference loss is 3.37, weight deacy loss is 6.41, training accuracy is 0.687500, time 201.902 samples/sec
+epoch 17, total_step 265780, total loss is 11.02 , inference loss is 4.61, weight deacy loss is 6.41, training accuracy is 0.687500, time 202.807 samples/sec
+epoch 17, total_step 265800, total loss is 15.84 , inference loss is 9.43, weight deacy loss is 6.41, training accuracy is 0.468750, time 12.647 samples/sec
+epoch 17, total_step 265820, total loss is 10.85 , inference loss is 4.44, weight deacy loss is 6.41, training accuracy is 0.500000, time 215.131 samples/sec
+epoch 17, total_step 265840, total loss is 12.94 , inference loss is 6.54, weight deacy loss is 6.41, training accuracy is 0.593750, time 202.038 samples/sec
+epoch 17, total_step 265860, total loss is 10.28 , inference loss is 3.88, weight deacy loss is 6.41, training accuracy is 0.687500, time 204.880 samples/sec
+epoch 17, total_step 265880, total loss is 12.85 , inference loss is 6.45, weight deacy loss is 6.41, training accuracy is 0.531250, time 228.936 samples/sec
+epoch 17, total_step 265900, total loss is 17.89 , inference loss is 11.48, weight deacy loss is 6.41, training accuracy is 0.375000, time 210.198 samples/sec
+epoch 17, total_step 265920, total loss is 14.70 , inference loss is 8.30, weight deacy loss is 6.41, training accuracy is 0.500000, time 182.401 samples/sec
+epoch 17, total_step 265940, total loss is 10.32 , inference loss is 3.92, weight deacy loss is 6.41, training accuracy is 0.656250, time 215.310 samples/sec
+epoch 17, total_step 265960, total loss is 15.17 , inference loss is 8.76, weight deacy loss is 6.41, training accuracy is 0.468750, time 207.842 samples/sec
+epoch 17, total_step 265980, total loss is 7.61 , inference loss is 1.20, weight deacy loss is 6.41, training accuracy is 0.750000, time 208.286 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.241230000000016
+best_threshold_index 127 0.9887037037037038
+best_threshold_index 127 0.9883333333333333
+best_threshold_index 127 0.9885185185185185
+best_threshold_index 127 0.9883333333333333
+best_threshold_index 128 0.9887037037037038
+best_threshold_index 127 0.9874074074074074
+best_threshold_index 127 0.9881481481481481
+best_threshold_index 121 0.9874074074074074
+best_threshold_index 127 0.9874074074074074
+best_threshold_index 127 0.9872222222222222
+[lfw][266000]XNorm: 20.713881
+[lfw][266000]Accuracy-Flip: 0.98667+-0.00506
+test accuracy is: 0.9866666666666667
+epoch 17, total_step 266000, total loss is 8.89 , inference loss is 2.48, weight deacy loss is 6.41, training accuracy is 0.625000, time 17.896 samples/sec
+epoch 17, total_step 266020, total loss is 8.34 , inference loss is 1.93, weight deacy loss is 6.41, training accuracy is 0.750000, time 223.940 samples/sec
+epoch 17, total_step 266040, total loss is 11.06 , inference loss is 4.65, weight deacy loss is 6.41, training accuracy is 0.500000, time 215.129 samples/sec
+epoch 17, total_step 266060, total loss is 15.14 , inference loss is 8.74, weight deacy loss is 6.41, training accuracy is 0.562500, time 223.026 samples/sec
+epoch 17, total_step 266080, total loss is 9.99 , inference loss is 3.59, weight deacy loss is 6.41, training accuracy is 0.687500, time 227.102 samples/sec
+epoch 17, total_step 266100, total loss is 13.70 , inference loss is 7.29, weight deacy loss is 6.41, training accuracy is 0.500000, time 12.267 samples/sec
+epoch 17, total_step 266120, total loss is 11.63 , inference loss is 5.22, weight deacy loss is 6.41, training accuracy is 0.500000, time 296.508 samples/sec
+epoch 17, total_step 266140, total loss is 14.19 , inference loss is 7.79, weight deacy loss is 6.41, training accuracy is 0.531250, time 218.189 samples/sec
+epoch 17, total_step 266160, total loss is 14.73 , inference loss is 8.32, weight deacy loss is 6.41, training accuracy is 0.500000, time 211.643 samples/sec
+epoch 17, total_step 266180, total loss is 13.18 , inference loss is 6.78, weight deacy loss is 6.41, training accuracy is 0.625000, time 224.409 samples/sec
+epoch 17, total_step 266200, total loss is 10.09 , inference loss is 3.69, weight deacy loss is 6.41, training accuracy is 0.687500, time 214.285 samples/sec
+epoch 17, total_step 266220, total loss is 10.63 , inference loss is 4.22, weight deacy loss is 6.41, training accuracy is 0.625000, time 227.912 samples/sec
+epoch 17, total_step 266240, total loss is 9.78 , inference loss is 3.38, weight deacy loss is 6.41, training accuracy is 0.718750, time 218.688 samples/sec
+epoch 17, total_step 266260, total loss is 9.97 , inference loss is 3.57, weight deacy loss is 6.41, training accuracy is 0.562500, time 222.535 samples/sec
+epoch 17, total_step 266280, total loss is 12.13 , inference loss is 5.72, weight deacy loss is 6.41, training accuracy is 0.625000, time 299.935 samples/sec
+epoch 17, total_step 266300, total loss is 15.65 , inference loss is 9.24, weight deacy loss is 6.41, training accuracy is 0.437500, time 297.847 samples/sec
+epoch 17, total_step 266320, total loss is 11.71 , inference loss is 5.30, weight deacy loss is 6.41, training accuracy is 0.468750, time 220.537 samples/sec
+epoch 17, total_step 266340, total loss is 13.25 , inference loss is 6.84, weight deacy loss is 6.41, training accuracy is 0.625000, time 220.117 samples/sec
+epoch 17, total_step 266360, total loss is 11.81 , inference loss is 5.41, weight deacy loss is 6.41, training accuracy is 0.593750, time 212.855 samples/sec
+epoch 17, total_step 266380, total loss is 12.43 , inference loss is 6.02, weight deacy loss is 6.41, training accuracy is 0.562500, time 214.968 samples/sec
+epoch 17, total_step 266400, total loss is 13.05 , inference loss is 6.65, weight deacy loss is 6.41, training accuracy is 0.562500, time 13.131 samples/sec
+epoch 17, total_step 266420, total loss is 10.55 , inference loss is 4.14, weight deacy loss is 6.41, training accuracy is 0.531250, time 228.651 samples/sec
+epoch 17, total_step 266440, total loss is 13.03 , inference loss is 6.63, weight deacy loss is 6.41, training accuracy is 0.562500, time 221.640 samples/sec
+epoch 17, total_step 266460, total loss is 10.49 , inference loss is 4.09, weight deacy loss is 6.41, training accuracy is 0.562500, time 218.280 samples/sec
+epoch 17, total_step 266480, total loss is 10.54 , inference loss is 4.13, weight deacy loss is 6.41, training accuracy is 0.625000, time 207.390 samples/sec
+epoch 17, total_step 266500, total loss is 15.02 , inference loss is 8.62, weight deacy loss is 6.41, training accuracy is 0.593750, time 218.066 samples/sec
+epoch 17, total_step 266520, total loss is 11.37 , inference loss is 4.96, weight deacy loss is 6.41, training accuracy is 0.531250, time 213.473 samples/sec
+epoch 17, total_step 266540, total loss is 14.30 , inference loss is 7.90, weight deacy loss is 6.41, training accuracy is 0.406250, time 220.211 samples/sec
+epoch 17, total_step 266560, total loss is 17.02 , inference loss is 10.62, weight deacy loss is 6.41, training accuracy is 0.531250, time 300.076 samples/sec
+epoch 17, total_step 266580, total loss is 12.02 , inference loss is 5.61, weight deacy loss is 6.41, training accuracy is 0.468750, time 214.904 samples/sec
+epoch 17, total_step 266600, total loss is 11.88 , inference loss is 5.47, weight deacy loss is 6.41, training accuracy is 0.625000, time 227.408 samples/sec
+epoch 17, total_step 266620, total loss is 12.16 , inference loss is 5.76, weight deacy loss is 6.41, training accuracy is 0.437500, time 218.914 samples/sec
+epoch 17, total_step 266640, total loss is 13.68 , inference loss is 7.27, weight deacy loss is 6.41, training accuracy is 0.500000, time 218.406 samples/sec
+epoch 17, total_step 266660, total loss is 13.76 , inference loss is 7.36, weight deacy loss is 6.41, training accuracy is 0.437500, time 222.828 samples/sec
+epoch 17, total_step 266680, total loss is 8.39 , inference loss is 1.99, weight deacy loss is 6.41, training accuracy is 0.593750, time 217.136 samples/sec
+epoch 17, total_step 266700, total loss is 13.02 , inference loss is 6.61, weight deacy loss is 6.41, training accuracy is 0.468750, time 12.430 samples/sec
+epoch 17, total_step 266720, total loss is 11.02 , inference loss is 4.62, weight deacy loss is 6.41, training accuracy is 0.625000, time 215.096 samples/sec
+epoch 17, total_step 266740, total loss is 12.70 , inference loss is 6.30, weight deacy loss is 6.41, training accuracy is 0.625000, time 215.164 samples/sec
+epoch 17, total_step 266760, total loss is 10.26 , inference loss is 3.85, weight deacy loss is 6.41, training accuracy is 0.656250, time 225.357 samples/sec
+epoch 17, total_step 266780, total loss is 11.75 , inference loss is 5.35, weight deacy loss is 6.41, training accuracy is 0.593750, time 218.880 samples/sec
+epoch 17, total_step 266800, total loss is 14.92 , inference loss is 8.51, weight deacy loss is 6.41, training accuracy is 0.593750, time 227.310 samples/sec
+epoch 17, total_step 266820, total loss is 16.89 , inference loss is 10.48, weight deacy loss is 6.41, training accuracy is 0.375000, time 294.296 samples/sec
+epoch 17, total_step 266840, total loss is 11.79 , inference loss is 5.39, weight deacy loss is 6.41, training accuracy is 0.406250, time 197.228 samples/sec
+epoch 17, total_step 266860, total loss is 12.39 , inference loss is 5.99, weight deacy loss is 6.41, training accuracy is 0.562500, time 223.858 samples/sec
+epoch 17, total_step 266880, total loss is 12.53 , inference loss is 6.12, weight deacy loss is 6.41, training accuracy is 0.593750, time 226.406 samples/sec
+epoch 17, total_step 266900, total loss is 10.98 , inference loss is 4.58, weight deacy loss is 6.41, training accuracy is 0.531250, time 215.404 samples/sec
+epoch 17, total_step 266920, total loss is 12.03 , inference loss is 5.62, weight deacy loss is 6.41, training accuracy is 0.687500, time 214.802 samples/sec
+epoch 17, total_step 266940, total loss is 14.56 , inference loss is 8.15, weight deacy loss is 6.41, training accuracy is 0.500000, time 220.130 samples/sec
+epoch 17, total_step 266960, total loss is 10.52 , inference loss is 4.12, weight deacy loss is 6.41, training accuracy is 0.500000, time 228.686 samples/sec
+epoch 17, total_step 266980, total loss is 13.47 , inference loss is 7.07, weight deacy loss is 6.41, training accuracy is 0.625000, time 214.909 samples/sec
+epoch 17, total_step 267000, total loss is 12.03 , inference loss is 5.63, weight deacy loss is 6.41, training accuracy is 0.468750, time 12.344 samples/sec
+epoch 17, total_step 267020, total loss is 11.25 , inference loss is 4.84, weight deacy loss is 6.41, training accuracy is 0.500000, time 221.071 samples/sec
+epoch 17, total_step 267040, total loss is 12.90 , inference loss is 6.49, weight deacy loss is 6.41, training accuracy is 0.531250, time 228.313 samples/sec
+epoch 17, total_step 267060, total loss is 9.85 , inference loss is 3.45, weight deacy loss is 6.41, training accuracy is 0.562500, time 216.716 samples/sec
+epoch 17, total_step 267080, total loss is 14.82 , inference loss is 8.42, weight deacy loss is 6.41, training accuracy is 0.468750, time 196.034 samples/sec
+epoch 17, total_step 267100, total loss is 14.07 , inference loss is 7.66, weight deacy loss is 6.41, training accuracy is 0.375000, time 210.456 samples/sec
+epoch 17, total_step 267120, total loss is 12.16 , inference loss is 5.76, weight deacy loss is 6.41, training accuracy is 0.437500, time 224.248 samples/sec
+epoch 17, total_step 267140, total loss is 11.66 , inference loss is 5.26, weight deacy loss is 6.41, training accuracy is 0.531250, time 218.891 samples/sec
+epoch 17, total_step 267160, total loss is 11.23 , inference loss is 4.83, weight deacy loss is 6.41, training accuracy is 0.468750, time 217.817 samples/sec
+epoch 17, total_step 267180, total loss is 12.77 , inference loss is 6.36, weight deacy loss is 6.41, training accuracy is 0.593750, time 226.132 samples/sec
+epoch 17, total_step 267200, total loss is 11.70 , inference loss is 5.30, weight deacy loss is 6.41, training accuracy is 0.437500, time 208.671 samples/sec
+epoch 17, total_step 267220, total loss is 14.06 , inference loss is 7.65, weight deacy loss is 6.41, training accuracy is 0.500000, time 227.635 samples/sec
+epoch 17, total_step 267240, total loss is 9.95 , inference loss is 3.54, weight deacy loss is 6.41, training accuracy is 0.687500, time 220.112 samples/sec
+epoch 17, total_step 267260, total loss is 11.29 , inference loss is 4.88, weight deacy loss is 6.41, training accuracy is 0.593750, time 226.211 samples/sec
+epoch 17, total_step 267280, total loss is 16.65 , inference loss is 10.25, weight deacy loss is 6.41, training accuracy is 0.500000, time 215.798 samples/sec
+epoch 17, total_step 267300, total loss is 11.32 , inference loss is 4.91, weight deacy loss is 6.41, training accuracy is 0.531250, time 13.625 samples/sec
+epoch 17, total_step 267320, total loss is 12.56 , inference loss is 6.16, weight deacy loss is 6.41, training accuracy is 0.375000, time 214.772 samples/sec
+epoch 17, total_step 267340, total loss is 14.37 , inference loss is 7.97, weight deacy loss is 6.41, training accuracy is 0.531250, time 219.602 samples/sec
+epoch 17, total_step 267360, total loss is 11.70 , inference loss is 5.30, weight deacy loss is 6.41, training accuracy is 0.593750, time 198.230 samples/sec
+epoch 17, total_step 267380, total loss is 14.29 , inference loss is 7.89, weight deacy loss is 6.41, training accuracy is 0.500000, time 217.628 samples/sec
+epoch 17, total_step 267400, total loss is 13.12 , inference loss is 6.72, weight deacy loss is 6.41, training accuracy is 0.531250, time 218.632 samples/sec
+epoch 17, total_step 267420, total loss is 12.37 , inference loss is 5.96, weight deacy loss is 6.41, training accuracy is 0.593750, time 215.922 samples/sec
+epoch 17, total_step 267440, total loss is 10.57 , inference loss is 4.17, weight deacy loss is 6.41, training accuracy is 0.625000, time 221.230 samples/sec
+epoch 17, total_step 267460, total loss is 12.31 , inference loss is 5.90, weight deacy loss is 6.41, training accuracy is 0.406250, time 214.106 samples/sec
+epoch 17, total_step 267480, total loss is 12.94 , inference loss is 6.53, weight deacy loss is 6.41, training accuracy is 0.468750, time 223.045 samples/sec
+epoch 17, total_step 267500, total loss is 16.30 , inference loss is 9.89, weight deacy loss is 6.41, training accuracy is 0.437500, time 221.391 samples/sec
+epoch 17, total_step 267520, total loss is 12.41 , inference loss is 6.00, weight deacy loss is 6.41, training accuracy is 0.593750, time 214.077 samples/sec
+epoch 17, total_step 267540, total loss is 12.24 , inference loss is 5.84, weight deacy loss is 6.41, training accuracy is 0.687500, time 220.038 samples/sec
+epoch 17, total_step 267560, total loss is 12.00 , inference loss is 5.59, weight deacy loss is 6.41, training accuracy is 0.562500, time 212.204 samples/sec
+epoch 17, total_step 267580, total loss is 14.96 , inference loss is 8.55, weight deacy loss is 6.41, training accuracy is 0.468750, time 215.702 samples/sec
+epoch 17, total_step 267600, total loss is 9.35 , inference loss is 2.94, weight deacy loss is 6.41, training accuracy is 0.656250, time 12.137 samples/sec
+epoch 17, total_step 267620, total loss is 12.00 , inference loss is 5.60, weight deacy loss is 6.41, training accuracy is 0.437500, time 245.527 samples/sec
+epoch 17, total_step 267640, total loss is 11.84 , inference loss is 5.44, weight deacy loss is 6.41, training accuracy is 0.593750, time 222.171 samples/sec
+epoch 17, total_step 267660, total loss is 13.40 , inference loss is 6.99, weight deacy loss is 6.41, training accuracy is 0.625000, time 221.458 samples/sec
+epoch 17, total_step 267680, total loss is 12.62 , inference loss is 6.22, weight deacy loss is 6.41, training accuracy is 0.343750, time 219.436 samples/sec
+epoch 17, total_step 267700, total loss is 14.33 , inference loss is 7.93, weight deacy loss is 6.41, training accuracy is 0.500000, time 225.232 samples/sec
+epoch 17, total_step 267720, total loss is 8.97 , inference loss is 2.57, weight deacy loss is 6.41, training accuracy is 0.687500, time 211.969 samples/sec
+epoch 17, total_step 267740, total loss is 11.30 , inference loss is 4.90, weight deacy loss is 6.41, training accuracy is 0.500000, time 216.798 samples/sec
+epoch 17, total_step 267760, total loss is 11.14 , inference loss is 4.74, weight deacy loss is 6.41, training accuracy is 0.593750, time 220.328 samples/sec
+epoch 17, total_step 267780, total loss is 13.55 , inference loss is 7.14, weight deacy loss is 6.41, training accuracy is 0.593750, time 198.424 samples/sec
+epoch 17, total_step 267800, total loss is 13.70 , inference loss is 7.29, weight deacy loss is 6.41, training accuracy is 0.656250, time 214.040 samples/sec
+epoch 17, total_step 267820, total loss is 12.83 , inference loss is 6.42, weight deacy loss is 6.41, training accuracy is 0.593750, time 206.996 samples/sec
+epoch 17, total_step 267840, total loss is 8.93 , inference loss is 2.53, weight deacy loss is 6.41, training accuracy is 0.656250, time 212.928 samples/sec
+epoch 17, total_step 267860, total loss is 13.09 , inference loss is 6.68, weight deacy loss is 6.41, training accuracy is 0.562500, time 217.111 samples/sec
+epoch 17, total_step 267880, total loss is 12.51 , inference loss is 6.10, weight deacy loss is 6.41, training accuracy is 0.500000, time 209.145 samples/sec
+epoch 17, total_step 267900, total loss is 8.24 , inference loss is 1.84, weight deacy loss is 6.41, training accuracy is 0.687500, time 11.322 samples/sec
+epoch 17, total_step 267920, total loss is 14.79 , inference loss is 8.38, weight deacy loss is 6.41, training accuracy is 0.375000, time 213.171 samples/sec
+epoch 17, total_step 267940, total loss is 11.43 , inference loss is 5.02, weight deacy loss is 6.41, training accuracy is 0.562500, time 201.544 samples/sec
+epoch 17, total_step 267960, total loss is 10.97 , inference loss is 4.56, weight deacy loss is 6.41, training accuracy is 0.562500, time 216.428 samples/sec
+epoch 17, total_step 267980, total loss is 14.51 , inference loss is 8.11, weight deacy loss is 6.41, training accuracy is 0.562500, time 221.993 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.45081800000001
+best_threshold_index 119 0.9888888888888889
+best_threshold_index 121 0.9887037037037038
+best_threshold_index 119 0.9881481481481481
+best_threshold_index 119 0.9887037037037038
+best_threshold_index 119 0.9888888888888889
+best_threshold_index 119 0.9877777777777778
+best_threshold_index 121 0.9892592592592593
+best_threshold_index 119 0.9879629629629629
+best_threshold_index 119 0.9877777777777778
+best_threshold_index 119 0.9875925925925926
+[lfw][268000]XNorm: 21.725108
+[lfw][268000]Accuracy-Flip: 0.98800+-0.00499
+test accuracy is: 0.9879999999999999
+epoch 17, total_step 268000, total loss is 12.76 , inference loss is 6.35, weight deacy loss is 6.41, training accuracy is 0.500000, time 16.960 samples/sec
+epoch 17, total_step 268020, total loss is 13.05 , inference loss is 6.64, weight deacy loss is 6.41, training accuracy is 0.500000, time 224.629 samples/sec
+epoch 17, total_step 268040, total loss is 11.36 , inference loss is 4.95, weight deacy loss is 6.41, training accuracy is 0.531250, time 213.227 samples/sec
+epoch 17, total_step 268060, total loss is 14.36 , inference loss is 7.95, weight deacy loss is 6.41, training accuracy is 0.625000, time 228.895 samples/sec
+epoch 17, total_step 268080, total loss is 13.78 , inference loss is 7.38, weight deacy loss is 6.41, training accuracy is 0.531250, time 211.965 samples/sec
+epoch 17, total_step 268100, total loss is 11.20 , inference loss is 4.79, weight deacy loss is 6.41, training accuracy is 0.593750, time 214.685 samples/sec
+epoch 17, total_step 268120, total loss is 12.15 , inference loss is 5.74, weight deacy loss is 6.41, training accuracy is 0.468750, time 299.084 samples/sec
+epoch 17, total_step 268140, total loss is 10.20 , inference loss is 3.79, weight deacy loss is 6.41, training accuracy is 0.656250, time 207.058 samples/sec
+epoch 17, total_step 268160, total loss is 14.23 , inference loss is 7.83, weight deacy loss is 6.41, training accuracy is 0.437500, time 211.148 samples/sec
+epoch 17, total_step 268180, total loss is 14.08 , inference loss is 7.67, weight deacy loss is 6.41, training accuracy is 0.500000, time 206.805 samples/sec
+epoch 17, total_step 268200, total loss is 15.75 , inference loss is 9.35, weight deacy loss is 6.41, training accuracy is 0.531250, time 8.933 samples/sec
+epoch 17, total_step 268220, total loss is 11.97 , inference loss is 5.56, weight deacy loss is 6.41, training accuracy is 0.531250, time 205.206 samples/sec
+epoch 17, total_step 268240, total loss is 13.77 , inference loss is 7.36, weight deacy loss is 6.41, training accuracy is 0.531250, time 205.677 samples/sec
+epoch 17, total_step 268260, total loss is 8.58 , inference loss is 2.18, weight deacy loss is 6.41, training accuracy is 0.781250, time 196.189 samples/sec
+epoch 17, total_step 268280, total loss is 13.67 , inference loss is 7.27, weight deacy loss is 6.41, training accuracy is 0.468750, time 226.941 samples/sec
+epoch 17, total_step 268300, total loss is 12.60 , inference loss is 6.20, weight deacy loss is 6.41, training accuracy is 0.625000, time 198.172 samples/sec
+epoch 17, total_step 268320, total loss is 9.08 , inference loss is 2.67, weight deacy loss is 6.40, training accuracy is 0.812500, time 217.022 samples/sec
+epoch 17, total_step 268340, total loss is 12.81 , inference loss is 6.40, weight deacy loss is 6.40, training accuracy is 0.500000, time 300.231 samples/sec
+epoch 17, total_step 268360, total loss is 11.30 , inference loss is 4.90, weight deacy loss is 6.40, training accuracy is 0.593750, time 209.222 samples/sec
+epoch 17, total_step 268380, total loss is 12.30 , inference loss is 5.89, weight deacy loss is 6.40, training accuracy is 0.500000, time 217.700 samples/sec
+epoch 17, total_step 268400, total loss is 11.77 , inference loss is 5.36, weight deacy loss is 6.40, training accuracy is 0.687500, time 208.580 samples/sec
+epoch 17, total_step 268420, total loss is 9.76 , inference loss is 3.35, weight deacy loss is 6.40, training accuracy is 0.718750, time 217.825 samples/sec
+epoch 17, total_step 268440, total loss is 12.58 , inference loss is 6.17, weight deacy loss is 6.40, training accuracy is 0.593750, time 209.867 samples/sec
+epoch 17, total_step 268460, total loss is 11.85 , inference loss is 5.44, weight deacy loss is 6.40, training accuracy is 0.531250, time 197.415 samples/sec
+epoch 17, total_step 268480, total loss is 12.87 , inference loss is 6.46, weight deacy loss is 6.40, training accuracy is 0.562500, time 192.007 samples/sec
+epoch 17, total_step 268500, total loss is 12.63 , inference loss is 6.22, weight deacy loss is 6.40, training accuracy is 0.468750, time 8.526 samples/sec
+epoch 17, total_step 268520, total loss is 12.44 , inference loss is 6.03, weight deacy loss is 6.40, training accuracy is 0.468750, time 206.125 samples/sec
+epoch 17, total_step 268540, total loss is 11.23 , inference loss is 4.83, weight deacy loss is 6.40, training accuracy is 0.656250, time 218.525 samples/sec
+epoch 17, total_step 268560, total loss is 11.08 , inference loss is 4.68, weight deacy loss is 6.40, training accuracy is 0.656250, time 300.222 samples/sec
+epoch 17, total_step 268580, total loss is 11.36 , inference loss is 4.96, weight deacy loss is 6.40, training accuracy is 0.562500, time 198.385 samples/sec
+epoch 17, total_step 268600, total loss is 12.24 , inference loss is 5.84, weight deacy loss is 6.40, training accuracy is 0.625000, time 203.630 samples/sec
+epoch 17, total_step 268620, total loss is 12.48 , inference loss is 6.08, weight deacy loss is 6.40, training accuracy is 0.468750, time 214.341 samples/sec
+epoch 17, total_step 268640, total loss is 16.00 , inference loss is 9.60, weight deacy loss is 6.40, training accuracy is 0.531250, time 210.745 samples/sec
+epoch 17, total_step 268660, total loss is 14.91 , inference loss is 8.51, weight deacy loss is 6.40, training accuracy is 0.562500, time 220.873 samples/sec
+epoch 17, total_step 268680, total loss is 14.33 , inference loss is 7.93, weight deacy loss is 6.40, training accuracy is 0.593750, time 216.229 samples/sec
+epoch 17, total_step 268700, total loss is 12.51 , inference loss is 6.11, weight deacy loss is 6.40, training accuracy is 0.562500, time 208.276 samples/sec
+epoch 17, total_step 268720, total loss is 15.26 , inference loss is 8.86, weight deacy loss is 6.40, training accuracy is 0.437500, time 207.496 samples/sec
+epoch 17, total_step 268740, total loss is 13.26 , inference loss is 6.85, weight deacy loss is 6.40, training accuracy is 0.531250, time 209.166 samples/sec
+epoch 17, total_step 268760, total loss is 12.47 , inference loss is 6.07, weight deacy loss is 6.40, training accuracy is 0.562500, time 215.767 samples/sec
+epoch 17, total_step 268780, total loss is 10.07 , inference loss is 3.67, weight deacy loss is 6.40, training accuracy is 0.625000, time 211.084 samples/sec
+epoch 17, total_step 268800, total loss is 14.28 , inference loss is 7.88, weight deacy loss is 6.40, training accuracy is 0.437500, time 12.197 samples/sec
+epoch 17, total_step 268820, total loss is 14.76 , inference loss is 8.36, weight deacy loss is 6.40, training accuracy is 0.500000, time 297.757 samples/sec
+epoch 17, total_step 268840, total loss is 10.74 , inference loss is 4.33, weight deacy loss is 6.40, training accuracy is 0.531250, time 300.425 samples/sec
+epoch 17, total_step 268860, total loss is 12.27 , inference loss is 5.86, weight deacy loss is 6.40, training accuracy is 0.531250, time 198.069 samples/sec
+epoch 17, total_step 268880, total loss is 14.40 , inference loss is 8.00, weight deacy loss is 6.40, training accuracy is 0.437500, time 193.420 samples/sec
+epoch 17, total_step 268900, total loss is 12.82 , inference loss is 6.42, weight deacy loss is 6.40, training accuracy is 0.500000, time 197.355 samples/sec
+epoch 17, total_step 268920, total loss is 11.18 , inference loss is 4.78, weight deacy loss is 6.40, training accuracy is 0.437500, time 198.389 samples/sec
+epoch 17, total_step 268940, total loss is 12.47 , inference loss is 6.06, weight deacy loss is 6.41, training accuracy is 0.656250, time 212.880 samples/sec
+epoch 17, total_step 268960, total loss is 15.70 , inference loss is 9.29, weight deacy loss is 6.41, training accuracy is 0.500000, time 209.853 samples/sec
+epoch 17, total_step 268980, total loss is 11.82 , inference loss is 5.41, weight deacy loss is 6.41, training accuracy is 0.562500, time 208.660 samples/sec
+epoch 17, total_step 269000, total loss is 13.61 , inference loss is 7.20, weight deacy loss is 6.41, training accuracy is 0.437500, time 199.949 samples/sec
+epoch 17, total_step 269020, total loss is 15.41 , inference loss is 9.00, weight deacy loss is 6.41, training accuracy is 0.468750, time 204.719 samples/sec
+epoch 17, total_step 269040, total loss is 15.95 , inference loss is 9.55, weight deacy loss is 6.41, training accuracy is 0.531250, time 208.911 samples/sec
+epoch 17, total_step 269060, total loss is 11.79 , inference loss is 5.38, weight deacy loss is 6.41, training accuracy is 0.562500, time 213.898 samples/sec
+epoch 17, total_step 269080, total loss is 15.81 , inference loss is 9.41, weight deacy loss is 6.41, training accuracy is 0.406250, time 219.999 samples/sec
+epoch 17, total_step 269100, total loss is 14.23 , inference loss is 7.83, weight deacy loss is 6.41, training accuracy is 0.531250, time 12.718 samples/sec
+epoch 17, total_step 269120, total loss is 14.98 , inference loss is 8.57, weight deacy loss is 6.41, training accuracy is 0.531250, time 199.565 samples/sec
+epoch 17, total_step 269140, total loss is 14.41 , inference loss is 8.01, weight deacy loss is 6.41, training accuracy is 0.468750, time 229.269 samples/sec
+epoch 17, total_step 269160, total loss is 17.94 , inference loss is 11.54, weight deacy loss is 6.41, training accuracy is 0.500000, time 195.384 samples/sec
+epoch 17, total_step 269180, total loss is 10.89 , inference loss is 4.49, weight deacy loss is 6.41, training accuracy is 0.656250, time 222.928 samples/sec
+epoch 17, total_step 269200, total loss is 16.80 , inference loss is 10.39, weight deacy loss is 6.41, training accuracy is 0.375000, time 205.419 samples/sec
+epoch 17, total_step 269220, total loss is 9.84 , inference loss is 3.44, weight deacy loss is 6.41, training accuracy is 0.531250, time 207.997 samples/sec
+epoch 17, total_step 269240, total loss is 12.35 , inference loss is 5.95, weight deacy loss is 6.41, training accuracy is 0.562500, time 228.792 samples/sec
+epoch 17, total_step 269260, total loss is 13.82 , inference loss is 7.41, weight deacy loss is 6.41, training accuracy is 0.468750, time 222.511 samples/sec
+epoch 17, total_step 269280, total loss is 11.54 , inference loss is 5.14, weight deacy loss is 6.41, training accuracy is 0.593750, time 212.805 samples/sec
+epoch 17, total_step 269300, total loss is 10.69 , inference loss is 4.28, weight deacy loss is 6.41, training accuracy is 0.625000, time 215.408 samples/sec
+epoch 17, total_step 269320, total loss is 14.03 , inference loss is 7.62, weight deacy loss is 6.41, training accuracy is 0.625000, time 201.933 samples/sec
+epoch 17, total_step 269340, total loss is 9.96 , inference loss is 3.55, weight deacy loss is 6.41, training accuracy is 0.656250, time 213.985 samples/sec
+epoch 17, total_step 269360, total loss is 16.02 , inference loss is 9.61, weight deacy loss is 6.41, training accuracy is 0.343750, time 216.724 samples/sec
+epoch 17, total_step 269380, total loss is 12.62 , inference loss is 6.21, weight deacy loss is 6.41, training accuracy is 0.593750, time 212.655 samples/sec
+epoch 17, total_step 269400, total loss is 11.84 , inference loss is 5.44, weight deacy loss is 6.41, training accuracy is 0.562500, time 14.468 samples/sec
+epoch 17, total_step 269420, total loss is 11.92 , inference loss is 5.52, weight deacy loss is 6.41, training accuracy is 0.500000, time 300.647 samples/sec
+epoch 17, total_step 269440, total loss is 10.38 , inference loss is 3.97, weight deacy loss is 6.41, training accuracy is 0.656250, time 197.052 samples/sec
+epoch 17, total_step 269460, total loss is 10.56 , inference loss is 4.15, weight deacy loss is 6.41, training accuracy is 0.593750, time 196.676 samples/sec
+epoch 17, total_step 269480, total loss is 13.24 , inference loss is 6.84, weight deacy loss is 6.41, training accuracy is 0.437500, time 226.017 samples/sec
+epoch 17, total_step 269500, total loss is 12.75 , inference loss is 6.34, weight deacy loss is 6.41, training accuracy is 0.593750, time 219.363 samples/sec
+epoch 17, total_step 269520, total loss is 10.39 , inference loss is 3.98, weight deacy loss is 6.41, training accuracy is 0.593750, time 231.775 samples/sec
+epoch 17, total_step 269540, total loss is 16.65 , inference loss is 10.25, weight deacy loss is 6.41, training accuracy is 0.343750, time 231.545 samples/sec
+epoch 17, total_step 269560, total loss is 15.14 , inference loss is 8.73, weight deacy loss is 6.41, training accuracy is 0.531250, time 198.135 samples/sec
+epoch 17, total_step 269580, total loss is 16.32 , inference loss is 9.91, weight deacy loss is 6.41, training accuracy is 0.437500, time 205.795 samples/sec
+epoch 17, total_step 269600, total loss is 11.12 , inference loss is 4.71, weight deacy loss is 6.41, training accuracy is 0.500000, time 202.184 samples/sec
+epoch 17, total_step 269620, total loss is 11.76 , inference loss is 5.35, weight deacy loss is 6.41, training accuracy is 0.593750, time 227.494 samples/sec
+epoch 17, total_step 269640, total loss is 11.83 , inference loss is 5.43, weight deacy loss is 6.41, training accuracy is 0.500000, time 296.419 samples/sec
+epoch 17, total_step 269660, total loss is 14.47 , inference loss is 8.06, weight deacy loss is 6.41, training accuracy is 0.531250, time 206.822 samples/sec
+epoch 17, total_step 269680, total loss is 11.37 , inference loss is 4.96, weight deacy loss is 6.41, training accuracy is 0.593750, time 213.925 samples/sec
+epoch 17, total_step 269700, total loss is 11.50 , inference loss is 5.09, weight deacy loss is 6.41, training accuracy is 0.531250, time 9.751 samples/sec
+epoch 17, total_step 269720, total loss is 12.22 , inference loss is 5.81, weight deacy loss is 6.41, training accuracy is 0.625000, time 191.470 samples/sec
+epoch 17, total_step 269740, total loss is 12.64 , inference loss is 6.23, weight deacy loss is 6.41, training accuracy is 0.468750, time 217.939 samples/sec
+epoch 17, total_step 269760, total loss is 9.72 , inference loss is 3.32, weight deacy loss is 6.41, training accuracy is 0.656250, time 217.251 samples/sec
+epoch 17, total_step 269780, total loss is 12.45 , inference loss is 6.04, weight deacy loss is 6.41, training accuracy is 0.500000, time 198.276 samples/sec
+epoch 17, total_step 269800, total loss is 14.49 , inference loss is 8.08, weight deacy loss is 6.41, training accuracy is 0.500000, time 192.801 samples/sec
+epoch 17, total_step 269820, total loss is 13.67 , inference loss is 7.27, weight deacy loss is 6.41, training accuracy is 0.468750, time 214.179 samples/sec
+epoch 17, total_step 269840, total loss is 12.36 , inference loss is 5.95, weight deacy loss is 6.41, training accuracy is 0.625000, time 176.376 samples/sec
+epoch 17, total_step 269860, total loss is 11.11 , inference loss is 4.71, weight deacy loss is 6.41, training accuracy is 0.625000, time 299.088 samples/sec
+epoch 17, total_step 269880, total loss is 11.88 , inference loss is 5.47, weight deacy loss is 6.41, training accuracy is 0.468750, time 215.031 samples/sec
+epoch 17, total_step 269900, total loss is 11.57 , inference loss is 5.16, weight deacy loss is 6.41, training accuracy is 0.531250, time 195.653 samples/sec
+epoch 17, total_step 269920, total loss is 14.00 , inference loss is 7.59, weight deacy loss is 6.41, training accuracy is 0.562500, time 224.330 samples/sec
+epoch 17, total_step 269940, total loss is 12.47 , inference loss is 6.06, weight deacy loss is 6.41, training accuracy is 0.500000, time 215.524 samples/sec
+epoch 17, total_step 269960, total loss is 13.75 , inference loss is 7.34, weight deacy loss is 6.41, training accuracy is 0.531250, time 210.910 samples/sec
+epoch 17, total_step 269980, total loss is 11.95 , inference loss is 5.54, weight deacy loss is 6.41, training accuracy is 0.531250, time 223.487 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.555210999999996
+best_threshold_index 115 0.9877777777777778
+best_threshold_index 115 0.9879629629629629
+best_threshold_index 115 0.9881481481481481
+best_threshold_index 115 0.9883333333333333
+best_threshold_index 115 0.9883333333333333
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9883333333333333
+best_threshold_index 115 0.987037037037037
+best_threshold_index 115 0.9874074074074074
+best_threshold_index 115 0.9875925925925926
+[lfw][270000]XNorm: 19.484533
+[lfw][270000]Accuracy-Flip: 0.98783+-0.00395
+test accuracy is: 0.9878333333333333
+epoch 17, total_step 270000, total loss is 14.79 , inference loss is 8.38, weight deacy loss is 6.41, training accuracy is 0.468750, time 16.233 samples/sec
+epoch 17, total_step 270020, total loss is 13.94 , inference loss is 7.53, weight deacy loss is 6.41, training accuracy is 0.437500, time 211.472 samples/sec
+epoch 17, total_step 270040, total loss is 10.42 , inference loss is 4.01, weight deacy loss is 6.41, training accuracy is 0.625000, time 229.847 samples/sec
+epoch 17, total_step 270060, total loss is 13.30 , inference loss is 6.90, weight deacy loss is 6.41, training accuracy is 0.593750, time 225.106 samples/sec
+epoch 17, total_step 270080, total loss is 11.83 , inference loss is 5.42, weight deacy loss is 6.41, training accuracy is 0.531250, time 220.969 samples/sec
+epoch 17, total_step 270100, total loss is 12.61 , inference loss is 6.20, weight deacy loss is 6.41, training accuracy is 0.500000, time 202.783 samples/sec
+epoch 17, total_step 270120, total loss is 15.87 , inference loss is 9.46, weight deacy loss is 6.41, training accuracy is 0.531250, time 212.740 samples/sec
+epoch 17, total_step 270140, total loss is 11.17 , inference loss is 4.76, weight deacy loss is 6.41, training accuracy is 0.468750, time 211.579 samples/sec
+epoch 17, total_step 270160, total loss is 15.21 , inference loss is 8.80, weight deacy loss is 6.41, training accuracy is 0.312500, time 215.587 samples/sec
+epoch 17, total_step 270180, total loss is 12.82 , inference loss is 6.41, weight deacy loss is 6.41, training accuracy is 0.531250, time 218.943 samples/sec
+epoch 17, total_step 270200, total loss is 10.70 , inference loss is 4.29, weight deacy loss is 6.41, training accuracy is 0.687500, time 179.188 samples/sec
+epoch 17, total_step 270220, total loss is 13.86 , inference loss is 7.45, weight deacy loss is 6.41, training accuracy is 0.562500, time 299.498 samples/sec
+epoch 17, total_step 270240, total loss is 12.12 , inference loss is 5.71, weight deacy loss is 6.41, training accuracy is 0.500000, time 213.816 samples/sec
+epoch 17, total_step 270260, total loss is 16.80 , inference loss is 10.39, weight deacy loss is 6.41, training accuracy is 0.437500, time 230.988 samples/sec
+epoch 17, total_step 270280, total loss is 13.49 , inference loss is 7.08, weight deacy loss is 6.41, training accuracy is 0.500000, time 217.218 samples/sec
+epoch 17, total_step 270300, total loss is 13.55 , inference loss is 7.14, weight deacy loss is 6.41, training accuracy is 0.500000, time 12.047 samples/sec
+epoch 17, total_step 270320, total loss is 16.64 , inference loss is 10.23, weight deacy loss is 6.41, training accuracy is 0.437500, time 224.145 samples/sec
+epoch 17, total_step 270340, total loss is 15.54 , inference loss is 9.13, weight deacy loss is 6.41, training accuracy is 0.531250, time 211.730 samples/sec
+epoch 17, total_step 270360, total loss is 10.46 , inference loss is 4.05, weight deacy loss is 6.41, training accuracy is 0.593750, time 200.862 samples/sec
+epoch 17, total_step 270380, total loss is 11.20 , inference loss is 4.79, weight deacy loss is 6.41, training accuracy is 0.750000, time 211.690 samples/sec
+epoch 17, total_step 270400, total loss is 13.35 , inference loss is 6.94, weight deacy loss is 6.41, training accuracy is 0.593750, time 193.549 samples/sec
+epoch 17, total_step 270420, total loss is 10.22 , inference loss is 3.81, weight deacy loss is 6.41, training accuracy is 0.687500, time 194.624 samples/sec
+epoch 17, total_step 270440, total loss is 12.25 , inference loss is 5.84, weight deacy loss is 6.41, training accuracy is 0.562500, time 192.695 samples/sec
+epoch 17, total_step 270460, total loss is 12.22 , inference loss is 5.81, weight deacy loss is 6.41, training accuracy is 0.656250, time 298.108 samples/sec
+epoch 17, total_step 270480, total loss is 14.58 , inference loss is 8.17, weight deacy loss is 6.41, training accuracy is 0.312500, time 296.862 samples/sec
+epoch 17, total_step 270500, total loss is 16.76 , inference loss is 10.35, weight deacy loss is 6.41, training accuracy is 0.468750, time 210.005 samples/sec
+epoch 17, total_step 270520, total loss is 13.51 , inference loss is 7.10, weight deacy loss is 6.41, training accuracy is 0.468750, time 214.414 samples/sec
+epoch 17, total_step 270540, total loss is 12.69 , inference loss is 6.28, weight deacy loss is 6.41, training accuracy is 0.562500, time 211.886 samples/sec
+epoch 17, total_step 270560, total loss is 11.76 , inference loss is 5.35, weight deacy loss is 6.41, training accuracy is 0.562500, time 221.085 samples/sec
+epoch 17, total_step 270580, total loss is 11.53 , inference loss is 5.12, weight deacy loss is 6.41, training accuracy is 0.500000, time 195.757 samples/sec
+epoch 17, total_step 270600, total loss is 11.26 , inference loss is 4.85, weight deacy loss is 6.41, training accuracy is 0.593750, time 8.940 samples/sec
+epoch 17, total_step 270620, total loss is 9.82 , inference loss is 3.41, weight deacy loss is 6.41, training accuracy is 0.625000, time 219.993 samples/sec
+epoch 17, total_step 270640, total loss is 14.44 , inference loss is 8.03, weight deacy loss is 6.41, training accuracy is 0.468750, time 207.794 samples/sec
+epoch 17, total_step 270660, total loss is 9.49 , inference loss is 3.08, weight deacy loss is 6.41, training accuracy is 0.531250, time 206.462 samples/sec
+epoch 17, total_step 270680, total loss is 9.65 , inference loss is 3.24, weight deacy loss is 6.41, training accuracy is 0.531250, time 198.506 samples/sec
+epoch 17, total_step 270700, total loss is 11.05 , inference loss is 4.65, weight deacy loss is 6.41, training accuracy is 0.718750, time 298.708 samples/sec
+epoch 17, total_step 270720, total loss is 10.75 , inference loss is 4.34, weight deacy loss is 6.41, training accuracy is 0.562500, time 299.262 samples/sec
+epoch 17, total_step 270740, total loss is 11.99 , inference loss is 5.58, weight deacy loss is 6.41, training accuracy is 0.468750, time 212.036 samples/sec
+epoch 17, total_step 270760, total loss is 11.37 , inference loss is 4.96, weight deacy loss is 6.41, training accuracy is 0.656250, time 219.195 samples/sec
+epoch 17, total_step 270780, total loss is 12.18 , inference loss is 5.78, weight deacy loss is 6.41, training accuracy is 0.531250, time 210.175 samples/sec
+epoch 17, total_step 270800, total loss is 18.35 , inference loss is 11.95, weight deacy loss is 6.41, training accuracy is 0.468750, time 215.012 samples/sec
+epoch 17, total_step 270820, total loss is 8.83 , inference loss is 2.42, weight deacy loss is 6.41, training accuracy is 0.562500, time 212.943 samples/sec
+epoch 17, total_step 270840, total loss is 11.82 , inference loss is 5.41, weight deacy loss is 6.41, training accuracy is 0.531250, time 220.337 samples/sec
+epoch 17, total_step 270860, total loss is 12.17 , inference loss is 5.76, weight deacy loss is 6.41, training accuracy is 0.500000, time 196.887 samples/sec
+epoch 17, total_step 270880, total loss is 10.19 , inference loss is 3.78, weight deacy loss is 6.41, training accuracy is 0.562500, time 218.800 samples/sec
+epoch 17, total_step 270900, total loss is 9.00 , inference loss is 2.59, weight deacy loss is 6.41, training accuracy is 0.656250, time 8.345 samples/sec
+epoch 17, total_step 270920, total loss is 14.06 , inference loss is 7.65, weight deacy loss is 6.41, training accuracy is 0.562500, time 192.167 samples/sec
+epoch 17, total_step 270940, total loss is 12.67 , inference loss is 6.27, weight deacy loss is 6.41, training accuracy is 0.687500, time 297.521 samples/sec
+epoch 17, total_step 270960, total loss is 10.52 , inference loss is 4.11, weight deacy loss is 6.41, training accuracy is 0.562500, time 299.006 samples/sec
+epoch 17, total_step 270980, total loss is 12.87 , inference loss is 6.46, weight deacy loss is 6.41, training accuracy is 0.593750, time 205.912 samples/sec
+epoch 17, total_step 271000, total loss is 9.89 , inference loss is 3.48, weight deacy loss is 6.41, training accuracy is 0.687500, time 211.517 samples/sec
+epoch 17, total_step 271020, total loss is 11.82 , inference loss is 5.41, weight deacy loss is 6.41, training accuracy is 0.625000, time 215.201 samples/sec
+epoch 17, total_step 271040, total loss is 10.91 , inference loss is 4.50, weight deacy loss is 6.41, training accuracy is 0.593750, time 214.466 samples/sec
+epoch 17, total_step 271060, total loss is 8.82 , inference loss is 2.41, weight deacy loss is 6.41, training accuracy is 0.812500, time 212.551 samples/sec
+epoch 17, total_step 271080, total loss is 10.37 , inference loss is 3.96, weight deacy loss is 6.41, training accuracy is 0.593750, time 216.521 samples/sec
+epoch 17, total_step 271100, total loss is 11.61 , inference loss is 5.20, weight deacy loss is 6.41, training accuracy is 0.656250, time 214.014 samples/sec
+epoch 17, total_step 271120, total loss is 11.56 , inference loss is 5.15, weight deacy loss is 6.41, training accuracy is 0.656250, time 298.815 samples/sec
+epoch 17, total_step 271140, total loss is 16.94 , inference loss is 10.54, weight deacy loss is 6.41, training accuracy is 0.343750, time 297.091 samples/sec
+epoch 17, total_step 271160, total loss is 10.03 , inference loss is 3.62, weight deacy loss is 6.41, training accuracy is 0.656250, time 298.912 samples/sec
+epoch 17, total_step 271180, total loss is 13.96 , inference loss is 7.55, weight deacy loss is 6.41, training accuracy is 0.500000, time 222.048 samples/sec
+epoch 17, total_step 271200, total loss is 14.42 , inference loss is 8.02, weight deacy loss is 6.41, training accuracy is 0.468750, time 12.494 samples/sec
+epoch 17, total_step 271220, total loss is 14.99 , inference loss is 8.58, weight deacy loss is 6.41, training accuracy is 0.437500, time 226.486 samples/sec
+epoch 17, total_step 271240, total loss is 13.01 , inference loss is 6.60, weight deacy loss is 6.41, training accuracy is 0.468750, time 225.413 samples/sec
+epoch 17, total_step 271260, total loss is 13.99 , inference loss is 7.59, weight deacy loss is 6.41, training accuracy is 0.437500, time 220.795 samples/sec
+epoch 17, total_step 271280, total loss is 10.80 , inference loss is 4.39, weight deacy loss is 6.41, training accuracy is 0.593750, time 213.985 samples/sec
+epoch 17, total_step 271300, total loss is 14.20 , inference loss is 7.79, weight deacy loss is 6.41, training accuracy is 0.406250, time 216.270 samples/sec
+epoch 17, total_step 271320, total loss is 12.59 , inference loss is 6.18, weight deacy loss is 6.41, training accuracy is 0.562500, time 213.695 samples/sec
+epoch 17, total_step 271340, total loss is 12.24 , inference loss is 5.84, weight deacy loss is 6.41, training accuracy is 0.468750, time 210.275 samples/sec
+epoch 17, total_step 271360, total loss is 9.25 , inference loss is 2.84, weight deacy loss is 6.41, training accuracy is 0.625000, time 211.864 samples/sec
+epoch 17, total_step 271380, total loss is 11.29 , inference loss is 4.88, weight deacy loss is 6.41, training accuracy is 0.531250, time 216.399 samples/sec
+epoch 17, total_step 271400, total loss is 17.65 , inference loss is 11.24, weight deacy loss is 6.41, training accuracy is 0.437500, time 217.905 samples/sec
+epoch 17, total_step 271420, total loss is 11.63 , inference loss is 5.22, weight deacy loss is 6.41, training accuracy is 0.656250, time 297.901 samples/sec
+epoch 17, total_step 271440, total loss is 12.64 , inference loss is 6.23, weight deacy loss is 6.41, training accuracy is 0.687500, time 220.537 samples/sec
+epoch 17, total_step 271460, total loss is 11.86 , inference loss is 5.45, weight deacy loss is 6.41, training accuracy is 0.593750, time 222.760 samples/sec
+epoch 17, total_step 271480, total loss is 12.41 , inference loss is 6.00, weight deacy loss is 6.41, training accuracy is 0.562500, time 217.121 samples/sec
+epoch 17, total_step 271500, total loss is 11.43 , inference loss is 5.02, weight deacy loss is 6.41, training accuracy is 0.656250, time 12.025 samples/sec
+epoch 17, total_step 271520, total loss is 15.02 , inference loss is 8.62, weight deacy loss is 6.41, training accuracy is 0.468750, time 225.059 samples/sec
+epoch 17, total_step 271540, total loss is 11.07 , inference loss is 4.66, weight deacy loss is 6.41, training accuracy is 0.500000, time 224.218 samples/sec
+epoch 17, total_step 271560, total loss is 12.00 , inference loss is 5.59, weight deacy loss is 6.41, training accuracy is 0.562500, time 221.417 samples/sec
+epoch 17, total_step 271580, total loss is 12.22 , inference loss is 5.81, weight deacy loss is 6.41, training accuracy is 0.500000, time 229.949 samples/sec
+epoch 17, total_step 271600, total loss is 13.57 , inference loss is 7.16, weight deacy loss is 6.41, training accuracy is 0.625000, time 226.042 samples/sec
+epoch 17, total_step 271620, total loss is 12.32 , inference loss is 5.91, weight deacy loss is 6.41, training accuracy is 0.625000, time 220.089 samples/sec
+epoch 17, total_step 271640, total loss is 12.73 , inference loss is 6.32, weight deacy loss is 6.41, training accuracy is 0.500000, time 224.835 samples/sec
+epoch 17, total_step 271660, total loss is 13.54 , inference loss is 7.13, weight deacy loss is 6.41, training accuracy is 0.625000, time 213.050 samples/sec
+epoch 17, total_step 271680, total loss is 13.48 , inference loss is 7.08, weight deacy loss is 6.41, training accuracy is 0.375000, time 296.820 samples/sec
+epoch 17, total_step 271700, total loss is 12.37 , inference loss is 5.96, weight deacy loss is 6.41, training accuracy is 0.531250, time 218.430 samples/sec
+epoch 17, total_step 271720, total loss is 15.57 , inference loss is 9.17, weight deacy loss is 6.41, training accuracy is 0.437500, time 219.029 samples/sec
+epoch 17, total_step 271740, total loss is 11.84 , inference loss is 5.43, weight deacy loss is 6.41, training accuracy is 0.531250, time 217.929 samples/sec
+epoch 17, total_step 271760, total loss is 11.04 , inference loss is 4.63, weight deacy loss is 6.41, training accuracy is 0.500000, time 219.695 samples/sec
+epoch 17, total_step 271780, total loss is 13.54 , inference loss is 7.13, weight deacy loss is 6.41, training accuracy is 0.406250, time 233.969 samples/sec
+epoch 17, total_step 271800, total loss is 13.42 , inference loss is 7.01, weight deacy loss is 6.41, training accuracy is 0.468750, time 12.497 samples/sec
+epoch 17, total_step 271820, total loss is 12.71 , inference loss is 6.30, weight deacy loss is 6.41, training accuracy is 0.500000, time 219.601 samples/sec
+epoch 17, total_step 271840, total loss is 12.06 , inference loss is 5.65, weight deacy loss is 6.41, training accuracy is 0.593750, time 222.965 samples/sec
+epoch 17, total_step 271860, total loss is 16.03 , inference loss is 9.62, weight deacy loss is 6.41, training accuracy is 0.375000, time 222.024 samples/sec
+epoch 17, total_step 271880, total loss is 10.02 , inference loss is 3.61, weight deacy loss is 6.41, training accuracy is 0.562500, time 225.570 samples/sec
+epoch 17, total_step 271900, total loss is 10.57 , inference loss is 4.17, weight deacy loss is 6.41, training accuracy is 0.593750, time 232.975 samples/sec
+epoch 17, total_step 271920, total loss is 12.81 , inference loss is 6.40, weight deacy loss is 6.41, training accuracy is 0.562500, time 236.602 samples/sec
+epoch 17, total_step 271940, total loss is 10.04 , inference loss is 3.63, weight deacy loss is 6.41, training accuracy is 0.687500, time 295.424 samples/sec
+epoch 17, total_step 271960, total loss is 13.03 , inference loss is 6.62, weight deacy loss is 6.41, training accuracy is 0.500000, time 209.528 samples/sec
+epoch 17, total_step 271980, total loss is 15.22 , inference loss is 8.81, weight deacy loss is 6.41, training accuracy is 0.437500, time 215.946 samples/sec
+testing verification..
+(12000, 512)
+infer time 26.032024999999994
+best_threshold_index 114 0.9842592592592593
+best_threshold_index 116 0.9835185185185186
+best_threshold_index 116 0.9838888888888889
+best_threshold_index 116 0.9857407407407407
+best_threshold_index 116 0.985
+best_threshold_index 116 0.9825925925925926
+best_threshold_index 116 0.9842592592592593
+best_threshold_index 116 0.9848148148148148
+best_threshold_index 116 0.9833333333333333
+best_threshold_index 116 0.9842592592592593
+[lfw][272000]XNorm: 21.595662
+[lfw][272000]Accuracy-Flip: 0.98383+-0.00778
+test accuracy is: 0.9838333333333333
+epoch 17, total_step 272000, total loss is 14.12 , inference loss is 7.71, weight deacy loss is 6.41, training accuracy is 0.406250, time 14.606 samples/sec
+epoch 17, total_step 272020, total loss is 14.46 , inference loss is 8.05, weight deacy loss is 6.41, training accuracy is 0.437500, time 297.629 samples/sec
+epoch 17, total_step 272040, total loss is 13.39 , inference loss is 6.99, weight deacy loss is 6.41, training accuracy is 0.593750, time 212.519 samples/sec
+epoch 17, total_step 272060, total loss is 13.82 , inference loss is 7.41, weight deacy loss is 6.41, training accuracy is 0.437500, time 219.565 samples/sec
+epoch 17, total_step 272080, total loss is 12.96 , inference loss is 6.55, weight deacy loss is 6.41, training accuracy is 0.625000, time 215.196 samples/sec
+epoch 17, total_step 272100, total loss is 10.92 , inference loss is 4.51, weight deacy loss is 6.41, training accuracy is 0.687500, time 12.977 samples/sec
+epoch 17, total_step 272120, total loss is 14.53 , inference loss is 8.12, weight deacy loss is 6.41, training accuracy is 0.468750, time 230.579 samples/sec
+epoch 17, total_step 272140, total loss is 12.02 , inference loss is 5.62, weight deacy loss is 6.41, training accuracy is 0.625000, time 176.079 samples/sec
+epoch 17, total_step 272160, total loss is 13.20 , inference loss is 6.79, weight deacy loss is 6.41, training accuracy is 0.500000, time 219.725 samples/sec
+epoch 17, total_step 272180, total loss is 12.98 , inference loss is 6.57, weight deacy loss is 6.41, training accuracy is 0.468750, time 216.732 samples/sec
+epoch 17, total_step 272200, total loss is 12.21 , inference loss is 5.81, weight deacy loss is 6.41, training accuracy is 0.406250, time 225.580 samples/sec
+epoch 17, total_step 272220, total loss is 10.53 , inference loss is 4.12, weight deacy loss is 6.41, training accuracy is 0.593750, time 233.470 samples/sec
+epoch 17, total_step 272240, total loss is 9.91 , inference loss is 3.50, weight deacy loss is 6.41, training accuracy is 0.562500, time 205.055 samples/sec
+epoch 17, total_step 272260, total loss is 15.80 , inference loss is 9.40, weight deacy loss is 6.41, training accuracy is 0.406250, time 240.998 samples/sec
+epoch 17, total_step 272280, total loss is 14.14 , inference loss is 7.73, weight deacy loss is 6.41, training accuracy is 0.343750, time 298.701 samples/sec
+epoch 17, total_step 272300, total loss is 11.18 , inference loss is 4.77, weight deacy loss is 6.41, training accuracy is 0.656250, time 211.039 samples/sec
+epoch 17, total_step 272320, total loss is 13.52 , inference loss is 7.11, weight deacy loss is 6.41, training accuracy is 0.531250, time 222.046 samples/sec
+epoch 17, total_step 272340, total loss is 13.31 , inference loss is 6.90, weight deacy loss is 6.41, training accuracy is 0.406250, time 212.212 samples/sec
+epoch 17, total_step 272360, total loss is 11.45 , inference loss is 5.04, weight deacy loss is 6.41, training accuracy is 0.625000, time 212.623 samples/sec
+epoch 17, total_step 272380, total loss is 17.52 , inference loss is 11.11, weight deacy loss is 6.41, training accuracy is 0.343750, time 212.693 samples/sec
+epoch 17, total_step 272400, total loss is 13.05 , inference loss is 6.64, weight deacy loss is 6.41, training accuracy is 0.531250, time 12.223 samples/sec
+epoch 17, total_step 272420, total loss is 12.30 , inference loss is 5.89, weight deacy loss is 6.41, training accuracy is 0.500000, time 217.746 samples/sec
+epoch 17, total_step 272440, total loss is 16.22 , inference loss is 9.82, weight deacy loss is 6.41, training accuracy is 0.562500, time 215.044 samples/sec
+epoch 17, total_step 272460, total loss is 9.94 , inference loss is 3.53, weight deacy loss is 6.41, training accuracy is 0.593750, time 223.191 samples/sec
+epoch 17, total_step 272480, total loss is 9.61 , inference loss is 3.20, weight deacy loss is 6.41, training accuracy is 0.531250, time 211.512 samples/sec
+epoch 17, total_step 272500, total loss is 13.36 , inference loss is 6.95, weight deacy loss is 6.41, training accuracy is 0.437500, time 217.604 samples/sec
+epoch 17, total_step 272520, total loss is 15.19 , inference loss is 8.78, weight deacy loss is 6.41, training accuracy is 0.437500, time 215.793 samples/sec
+epoch 17, total_step 272540, total loss is 10.45 , inference loss is 4.04, weight deacy loss is 6.41, training accuracy is 0.593750, time 295.533 samples/sec
+epoch 17, total_step 272560, total loss is 13.68 , inference loss is 7.27, weight deacy loss is 6.41, training accuracy is 0.562500, time 228.758 samples/sec
+epoch 17, total_step 272580, total loss is 12.44 , inference loss is 6.03, weight deacy loss is 6.41, training accuracy is 0.500000, time 220.888 samples/sec
+epoch 17, total_step 272600, total loss is 10.21 , inference loss is 3.80, weight deacy loss is 6.41, training accuracy is 0.593750, time 216.889 samples/sec
+epoch 17, total_step 272620, total loss is 14.63 , inference loss is 8.22, weight deacy loss is 6.41, training accuracy is 0.437500, time 217.206 samples/sec
+epoch 17, total_step 272640, total loss is 12.52 , inference loss is 6.11, weight deacy loss is 6.41, training accuracy is 0.531250, time 229.651 samples/sec
+epoch 17, total_step 272660, total loss is 13.74 , inference loss is 7.33, weight deacy loss is 6.41, training accuracy is 0.531250, time 219.616 samples/sec
+epoch 17, total_step 272680, total loss is 14.25 , inference loss is 7.84, weight deacy loss is 6.41, training accuracy is 0.281250, time 220.045 samples/sec
+epoch 17, total_step 272700, total loss is 15.18 , inference loss is 8.78, weight deacy loss is 6.41, training accuracy is 0.406250, time 12.608 samples/sec
+epoch 17, total_step 272720, total loss is 12.98 , inference loss is 6.57, weight deacy loss is 6.41, training accuracy is 0.562500, time 210.441 samples/sec
+epoch 17, total_step 272740, total loss is 13.40 , inference loss is 6.99, weight deacy loss is 6.41, training accuracy is 0.531250, time 217.092 samples/sec
+epoch 17, total_step 272760, total loss is 15.70 , inference loss is 9.29, weight deacy loss is 6.41, training accuracy is 0.562500, time 216.236 samples/sec
+epoch 17, total_step 272780, total loss is 13.96 , inference loss is 7.55, weight deacy loss is 6.41, training accuracy is 0.312500, time 215.233 samples/sec
+epoch 17, total_step 272800, total loss is 11.80 , inference loss is 5.39, weight deacy loss is 6.41, training accuracy is 0.375000, time 217.103 samples/sec
+epoch 17, total_step 272820, total loss is 12.96 , inference loss is 6.55, weight deacy loss is 6.41, training accuracy is 0.593750, time 226.409 samples/sec
+epoch 17, total_step 272840, total loss is 13.65 , inference loss is 7.25, weight deacy loss is 6.41, training accuracy is 0.562500, time 298.693 samples/sec
+epoch 17, total_step 272860, total loss is 10.08 , inference loss is 3.67, weight deacy loss is 6.41, training accuracy is 0.625000, time 296.252 samples/sec
+epoch 17, total_step 272880, total loss is 15.32 , inference loss is 8.92, weight deacy loss is 6.41, training accuracy is 0.531250, time 298.103 samples/sec
+epoch 17, total_step 272900, total loss is 18.13 , inference loss is 11.72, weight deacy loss is 6.41, training accuracy is 0.500000, time 177.126 samples/sec
+epoch 17, total_step 272920, total loss is 11.96 , inference loss is 5.56, weight deacy loss is 6.41, training accuracy is 0.531250, time 215.157 samples/sec
+epoch 17, total_step 272940, total loss is 14.68 , inference loss is 8.27, weight deacy loss is 6.41, training accuracy is 0.500000, time 218.449 samples/sec
+epoch 17, total_step 272960, total loss is 12.77 , inference loss is 6.36, weight deacy loss is 6.41, training accuracy is 0.531250, time 213.238 samples/sec
+epoch 17, total_step 272980, total loss is 12.16 , inference loss is 5.75, weight deacy loss is 6.41, training accuracy is 0.562500, time 212.405 samples/sec
+epoch 17, total_step 273000, total loss is 14.95 , inference loss is 8.54, weight deacy loss is 6.41, training accuracy is 0.312500, time 11.979 samples/sec
+epoch 17, total_step 273020, total loss is 13.89 , inference loss is 7.49, weight deacy loss is 6.41, training accuracy is 0.406250, time 210.775 samples/sec
+epoch 17, total_step 273040, total loss is 10.56 , inference loss is 4.15, weight deacy loss is 6.41, training accuracy is 0.562500, time 216.435 samples/sec
+epoch 17, total_step 273060, total loss is 18.75 , inference loss is 12.34, weight deacy loss is 6.41, training accuracy is 0.406250, time 214.938 samples/sec
+epoch 17, total_step 273080, total loss is 12.42 , inference loss is 6.01, weight deacy loss is 6.41, training accuracy is 0.437500, time 207.586 samples/sec
+epoch 17, total_step 273100, total loss is 12.21 , inference loss is 5.80, weight deacy loss is 6.41, training accuracy is 0.375000, time 213.037 samples/sec
+epoch 17, total_step 273120, total loss is 15.88 , inference loss is 9.47, weight deacy loss is 6.41, training accuracy is 0.437500, time 215.589 samples/sec
+epoch 17, total_step 273140, total loss is 10.32 , inference loss is 3.91, weight deacy loss is 6.41, training accuracy is 0.593750, time 216.522 samples/sec
+epoch 17, total_step 273160, total loss is 10.27 , inference loss is 3.86, weight deacy loss is 6.41, training accuracy is 0.562500, time 215.365 samples/sec
+epoch 17, total_step 273180, total loss is 14.54 , inference loss is 8.13, weight deacy loss is 6.41, training accuracy is 0.625000, time 213.372 samples/sec
+epoch 17, total_step 273200, total loss is 12.52 , inference loss is 6.11, weight deacy loss is 6.41, training accuracy is 0.562500, time 211.835 samples/sec
+epoch 17, total_step 273220, total loss is 15.09 , inference loss is 8.68, weight deacy loss is 6.41, training accuracy is 0.437500, time 216.334 samples/sec
+epoch 17, total_step 273240, total loss is 14.15 , inference loss is 7.74, weight deacy loss is 6.41, training accuracy is 0.500000, time 206.915 samples/sec
+epoch 17, total_step 273260, total loss is 16.20 , inference loss is 9.79, weight deacy loss is 6.41, training accuracy is 0.437500, time 300.124 samples/sec
+epoch 17, total_step 273280, total loss is 12.39 , inference loss is 5.99, weight deacy loss is 6.41, training accuracy is 0.531250, time 222.033 samples/sec
+epoch 17, total_step 273300, total loss is 15.57 , inference loss is 9.17, weight deacy loss is 6.41, training accuracy is 0.406250, time 12.083 samples/sec
+epoch 17, total_step 273320, total loss is 11.74 , inference loss is 5.33, weight deacy loss is 6.41, training accuracy is 0.500000, time 220.788 samples/sec
+epoch 17, total_step 273340, total loss is 12.61 , inference loss is 6.20, weight deacy loss is 6.41, training accuracy is 0.468750, time 217.175 samples/sec
+epoch 17, total_step 273360, total loss is 13.29 , inference loss is 6.88, weight deacy loss is 6.41, training accuracy is 0.312500, time 213.136 samples/sec
+epoch 17, total_step 273380, total loss is 14.03 , inference loss is 7.63, weight deacy loss is 6.41, training accuracy is 0.343750, time 222.374 samples/sec
+epoch 17, total_step 273400, total loss is 14.19 , inference loss is 7.78, weight deacy loss is 6.41, training accuracy is 0.468750, time 210.015 samples/sec
+epoch 17, total_step 273420, total loss is 13.01 , inference loss is 6.60, weight deacy loss is 6.41, training accuracy is 0.406250, time 210.639 samples/sec
+epoch 17, total_step 273440, total loss is 16.85 , inference loss is 10.45, weight deacy loss is 6.41, training accuracy is 0.375000, time 212.573 samples/sec
+epoch 17, total_step 273460, total loss is 11.20 , inference loss is 4.79, weight deacy loss is 6.41, training accuracy is 0.593750, time 223.829 samples/sec
+epoch 17, total_step 273480, total loss is 12.56 , inference loss is 6.15, weight deacy loss is 6.41, training accuracy is 0.437500, time 221.862 samples/sec
+epoch 17, total_step 273500, total loss is 12.86 , inference loss is 6.45, weight deacy loss is 6.41, training accuracy is 0.500000, time 186.356 samples/sec
+epoch 17, total_step 273520, total loss is 13.13 , inference loss is 6.72, weight deacy loss is 6.41, training accuracy is 0.406250, time 296.778 samples/sec
+epoch 17, total_step 273540, total loss is 14.98 , inference loss is 8.57, weight deacy loss is 6.41, training accuracy is 0.656250, time 217.318 samples/sec
+epoch 17, total_step 273560, total loss is 11.89 , inference loss is 5.49, weight deacy loss is 6.41, training accuracy is 0.437500, time 206.789 samples/sec
+epoch 17, total_step 273580, total loss is 12.63 , inference loss is 6.22, weight deacy loss is 6.41, training accuracy is 0.406250, time 207.923 samples/sec
+epoch 17, total_step 273600, total loss is 12.55 , inference loss is 6.15, weight deacy loss is 6.41, training accuracy is 0.500000, time 8.812 samples/sec
+epoch 17, total_step 273620, total loss is 11.18 , inference loss is 4.77, weight deacy loss is 6.41, training accuracy is 0.312500, time 211.724 samples/sec
+epoch 17, total_step 273640, total loss is 12.42 , inference loss is 6.01, weight deacy loss is 6.41, training accuracy is 0.406250, time 217.218 samples/sec
+epoch 17, total_step 273660, total loss is 15.14 , inference loss is 8.73, weight deacy loss is 6.41, training accuracy is 0.406250, time 211.270 samples/sec
+epoch 17, total_step 273680, total loss is 12.80 , inference loss is 6.40, weight deacy loss is 6.41, training accuracy is 0.562500, time 219.262 samples/sec
+epoch 17, total_step 273700, total loss is 20.22 , inference loss is 13.82, weight deacy loss is 6.41, training accuracy is 0.343750, time 217.252 samples/sec
+epoch 17, total_step 273720, total loss is 14.97 , inference loss is 8.56, weight deacy loss is 6.41, training accuracy is 0.375000, time 298.994 samples/sec
+epoch 17, total_step 273740, total loss is 18.95 , inference loss is 12.54, weight deacy loss is 6.41, training accuracy is 0.406250, time 219.441 samples/sec
+epoch 17, total_step 273760, total loss is 19.26 , inference loss is 12.85, weight deacy loss is 6.41, training accuracy is 0.218750, time 217.651 samples/sec
+epoch 17, total_step 273780, total loss is 14.99 , inference loss is 8.58, weight deacy loss is 6.41, training accuracy is 0.437500, time 211.299 samples/sec
+epoch 17, total_step 273800, total loss is 16.52 , inference loss is 10.11, weight deacy loss is 6.41, training accuracy is 0.218750, time 213.641 samples/sec
+epoch 17, total_step 273820, total loss is 13.01 , inference loss is 6.61, weight deacy loss is 6.41, training accuracy is 0.281250, time 217.709 samples/sec
+epoch 17, total_step 273840, total loss is 12.45 , inference loss is 6.04, weight deacy loss is 6.41, training accuracy is 0.531250, time 215.893 samples/sec
+epoch 17, total_step 273860, total loss is 15.71 , inference loss is 9.30, weight deacy loss is 6.41, training accuracy is 0.500000, time 219.709 samples/sec
+epoch 17, total_step 273880, total loss is 14.24 , inference loss is 7.83, weight deacy loss is 6.41, training accuracy is 0.593750, time 213.164 samples/sec
+epoch 17, total_step 273900, total loss is 19.23 , inference loss is 12.82, weight deacy loss is 6.41, training accuracy is 0.250000, time 11.809 samples/sec
+epoch 17, total_step 273920, total loss is 14.09 , inference loss is 7.68, weight deacy loss is 6.41, training accuracy is 0.531250, time 232.015 samples/sec
+epoch 17, total_step 273940, total loss is 13.67 , inference loss is 7.26, weight deacy loss is 6.41, training accuracy is 0.437500, time 224.361 samples/sec
+epoch 17, total_step 273960, total loss is 13.22 , inference loss is 6.81, weight deacy loss is 6.41, training accuracy is 0.406250, time 214.291 samples/sec
+epoch 17, total_step 273980, total loss is 15.30 , inference loss is 8.89, weight deacy loss is 6.41, training accuracy is 0.375000, time 297.431 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.828824999999973
+best_threshold_index 112 0.9835185185185186
+best_threshold_index 112 0.9822222222222222
+best_threshold_index 113 0.9833333333333333
+best_threshold_index 112 0.9848148148148148
+best_threshold_index 112 0.9837037037037037
+best_threshold_index 112 0.9818518518518519
+best_threshold_index 112 0.9825925925925926
+best_threshold_index 112 0.9822222222222222
+best_threshold_index 112 0.980925925925926
+best_threshold_index 112 0.9816666666666667
+[lfw][274000]XNorm: 21.992819
+[lfw][274000]Accuracy-Flip: 0.98233+-0.00995
+test accuracy is: 0.9823333333333334
+epoch 17, total_step 274000, total loss is 18.28 , inference loss is 11.87, weight deacy loss is 6.41, training accuracy is 0.218750, time 15.366 samples/sec
+epoch 17, total_step 274020, total loss is 14.58 , inference loss is 8.18, weight deacy loss is 6.41, training accuracy is 0.500000, time 198.537 samples/sec
+epoch 17, total_step 274040, total loss is 16.25 , inference loss is 9.84, weight deacy loss is 6.41, training accuracy is 0.312500, time 202.761 samples/sec
+epoch 17, total_step 274060, total loss is 13.46 , inference loss is 7.05, weight deacy loss is 6.41, training accuracy is 0.531250, time 295.245 samples/sec
+epoch 17, total_step 274080, total loss is 14.87 , inference loss is 8.46, weight deacy loss is 6.41, training accuracy is 0.468750, time 296.618 samples/sec
+epoch 17, total_step 274100, total loss is 12.58 , inference loss is 6.17, weight deacy loss is 6.41, training accuracy is 0.562500, time 205.962 samples/sec
+epoch 17, total_step 274120, total loss is 11.62 , inference loss is 5.21, weight deacy loss is 6.41, training accuracy is 0.375000, time 231.597 samples/sec
+epoch 17, total_step 274140, total loss is 11.26 , inference loss is 4.85, weight deacy loss is 6.41, training accuracy is 0.500000, time 201.562 samples/sec
+epoch 17, total_step 274160, total loss is 12.69 , inference loss is 6.28, weight deacy loss is 6.41, training accuracy is 0.437500, time 205.176 samples/sec
+epoch 17, total_step 274180, total loss is 10.09 , inference loss is 3.68, weight deacy loss is 6.41, training accuracy is 0.625000, time 218.180 samples/sec
+epoch 17, total_step 274200, total loss is 12.97 , inference loss is 6.56, weight deacy loss is 6.41, training accuracy is 0.500000, time 12.189 samples/sec
+epoch 17, total_step 274220, total loss is 12.48 , inference loss is 6.08, weight deacy loss is 6.41, training accuracy is 0.437500, time 231.197 samples/sec
+epoch 17, total_step 274240, total loss is 12.58 , inference loss is 6.17, weight deacy loss is 6.41, training accuracy is 0.468750, time 211.000 samples/sec
+epoch 17, total_step 274260, total loss is 15.34 , inference loss is 8.93, weight deacy loss is 6.41, training accuracy is 0.500000, time 223.857 samples/sec
+epoch 17, total_step 274280, total loss is 11.88 , inference loss is 5.47, weight deacy loss is 6.41, training accuracy is 0.531250, time 208.301 samples/sec
+epoch 17, total_step 274300, total loss is 13.74 , inference loss is 7.34, weight deacy loss is 6.41, training accuracy is 0.500000, time 210.270 samples/sec
+epoch 17, total_step 274320, total loss is 10.19 , inference loss is 3.78, weight deacy loss is 6.41, training accuracy is 0.500000, time 224.632 samples/sec
+epoch 17, total_step 274340, total loss is 13.83 , inference loss is 7.42, weight deacy loss is 6.41, training accuracy is 0.343750, time 297.138 samples/sec
+epoch 17, total_step 274360, total loss is 13.45 , inference loss is 7.04, weight deacy loss is 6.41, training accuracy is 0.406250, time 239.837 samples/sec
+epoch 17, total_step 274380, total loss is 16.05 , inference loss is 9.64, weight deacy loss is 6.41, training accuracy is 0.437500, time 227.027 samples/sec
+epoch 17, total_step 274400, total loss is 14.95 , inference loss is 8.54, weight deacy loss is 6.41, training accuracy is 0.312500, time 214.877 samples/sec
+epoch 17, total_step 274420, total loss is 13.39 , inference loss is 6.99, weight deacy loss is 6.41, training accuracy is 0.625000, time 220.169 samples/sec
+epoch 17, total_step 274440, total loss is 14.12 , inference loss is 7.71, weight deacy loss is 6.41, training accuracy is 0.468750, time 233.045 samples/sec
+epoch 17, total_step 274460, total loss is 14.25 , inference loss is 7.84, weight deacy loss is 6.41, training accuracy is 0.593750, time 214.804 samples/sec
+epoch 17, total_step 274480, total loss is 15.12 , inference loss is 8.71, weight deacy loss is 6.41, training accuracy is 0.437500, time 225.885 samples/sec
+epoch 17, total_step 274500, total loss is 11.16 , inference loss is 4.76, weight deacy loss is 6.41, training accuracy is 0.500000, time 8.672 samples/sec
+epoch 17, total_step 274520, total loss is 11.09 , inference loss is 4.68, weight deacy loss is 6.41, training accuracy is 0.500000, time 294.748 samples/sec
+epoch 17, total_step 274540, total loss is 10.75 , inference loss is 4.34, weight deacy loss is 6.41, training accuracy is 0.625000, time 301.912 samples/sec
+epoch 17, total_step 274560, total loss is 14.85 , inference loss is 8.45, weight deacy loss is 6.41, training accuracy is 0.406250, time 300.485 samples/sec
+epoch 17, total_step 274580, total loss is 13.22 , inference loss is 6.81, weight deacy loss is 6.41, training accuracy is 0.593750, time 300.293 samples/sec
+epoch 17, total_step 274600, total loss is 19.01 , inference loss is 12.60, weight deacy loss is 6.41, training accuracy is 0.312500, time 298.307 samples/sec
+epoch 17, total_step 274620, total loss is 14.70 , inference loss is 8.29, weight deacy loss is 6.41, training accuracy is 0.406250, time 296.641 samples/sec
+epoch 17, total_step 274640, total loss is 17.56 , inference loss is 11.15, weight deacy loss is 6.41, training accuracy is 0.406250, time 300.038 samples/sec
+epoch 17, total_step 274660, total loss is 15.41 , inference loss is 9.00, weight deacy loss is 6.41, training accuracy is 0.406250, time 203.781 samples/sec
+epoch 17, total_step 274680, total loss is 13.02 , inference loss is 6.62, weight deacy loss is 6.41, training accuracy is 0.500000, time 205.533 samples/sec
+epoch 17, total_step 274700, total loss is 14.04 , inference loss is 7.63, weight deacy loss is 6.41, training accuracy is 0.375000, time 207.993 samples/sec
+epoch 17, total_step 274720, total loss is 11.94 , inference loss is 5.53, weight deacy loss is 6.41, training accuracy is 0.562500, time 215.741 samples/sec
+epoch 17, total_step 274740, total loss is 13.98 , inference loss is 7.57, weight deacy loss is 6.41, training accuracy is 0.437500, time 228.019 samples/sec
+epoch 17, total_step 274760, total loss is 13.79 , inference loss is 7.38, weight deacy loss is 6.41, training accuracy is 0.437500, time 210.133 samples/sec
+epoch 17, total_step 274780, total loss is 18.26 , inference loss is 11.85, weight deacy loss is 6.41, training accuracy is 0.406250, time 199.479 samples/sec
+epoch 17, total_step 274800, total loss is 12.71 , inference loss is 6.31, weight deacy loss is 6.41, training accuracy is 0.437500, time 8.794 samples/sec
+epoch 17, total_step 274820, total loss is 14.69 , inference loss is 8.29, weight deacy loss is 6.41, training accuracy is 0.312500, time 216.478 samples/sec
+epoch 17, total_step 274840, total loss is 10.66 , inference loss is 4.26, weight deacy loss is 6.41, training accuracy is 0.468750, time 226.121 samples/sec
+epoch 17, total_step 274860, total loss is 13.57 , inference loss is 7.17, weight deacy loss is 6.41, training accuracy is 0.343750, time 238.111 samples/sec
+epoch 17, total_step 274880, total loss is 15.80 , inference loss is 9.39, weight deacy loss is 6.41, training accuracy is 0.343750, time 226.467 samples/sec
+epoch 17, total_step 274900, total loss is 14.17 , inference loss is 7.76, weight deacy loss is 6.41, training accuracy is 0.343750, time 212.404 samples/sec
+epoch 17, total_step 274920, total loss is 12.17 , inference loss is 5.76, weight deacy loss is 6.41, training accuracy is 0.468750, time 204.310 samples/sec
+epoch 17, total_step 274940, total loss is 17.02 , inference loss is 10.61, weight deacy loss is 6.41, training accuracy is 0.312500, time 195.676 samples/sec
+epoch 17, total_step 274960, total loss is 18.97 , inference loss is 12.57, weight deacy loss is 6.41, training accuracy is 0.343750, time 196.500 samples/sec
+epoch 17, total_step 274980, total loss is 14.80 , inference loss is 8.40, weight deacy loss is 6.41, training accuracy is 0.375000, time 196.603 samples/sec
+epoch 17, total_step 275000, total loss is 12.07 , inference loss is 5.66, weight deacy loss is 6.41, training accuracy is 0.562500, time 217.078 samples/sec
+epoch 17, total_step 275020, total loss is 11.47 , inference loss is 5.06, weight deacy loss is 6.41, training accuracy is 0.500000, time 197.962 samples/sec
+epoch 17, total_step 275040, total loss is 17.46 , inference loss is 11.05, weight deacy loss is 6.41, training accuracy is 0.375000, time 217.647 samples/sec
+epoch 17, total_step 275060, total loss is 15.03 , inference loss is 8.63, weight deacy loss is 6.41, training accuracy is 0.343750, time 207.403 samples/sec
+epoch 17, total_step 275080, total loss is 15.40 , inference loss is 9.00, weight deacy loss is 6.41, training accuracy is 0.312500, time 296.145 samples/sec
+epoch 17, total_step 275100, total loss is 14.28 , inference loss is 7.87, weight deacy loss is 6.41, training accuracy is 0.437500, time 8.874 samples/sec
+epoch 17, total_step 275120, total loss is 14.29 , inference loss is 7.88, weight deacy loss is 6.41, training accuracy is 0.406250, time 202.044 samples/sec
+epoch 17, total_step 275140, total loss is 14.98 , inference loss is 8.57, weight deacy loss is 6.41, training accuracy is 0.281250, time 221.008 samples/sec
+epoch 17, total_step 275160, total loss is 10.28 , inference loss is 3.87, weight deacy loss is 6.41, training accuracy is 0.531250, time 197.711 samples/sec
+epoch 17, total_step 275180, total loss is 18.24 , inference loss is 11.83, weight deacy loss is 6.41, training accuracy is 0.406250, time 215.375 samples/sec
+epoch 17, total_step 275200, total loss is 13.51 , inference loss is 7.10, weight deacy loss is 6.41, training accuracy is 0.375000, time 204.397 samples/sec
+epoch 17, total_step 275220, total loss is 14.44 , inference loss is 8.03, weight deacy loss is 6.41, training accuracy is 0.468750, time 207.973 samples/sec
+epoch 17, total_step 275240, total loss is 12.29 , inference loss is 5.89, weight deacy loss is 6.41, training accuracy is 0.562500, time 196.305 samples/sec
+epoch 17, total_step 275260, total loss is 12.14 , inference loss is 5.73, weight deacy loss is 6.41, training accuracy is 0.468750, time 209.378 samples/sec
+epoch 17, total_step 275280, total loss is 16.68 , inference loss is 10.28, weight deacy loss is 6.41, training accuracy is 0.500000, time 208.924 samples/sec
+epoch 17, total_step 275300, total loss is 15.79 , inference loss is 9.39, weight deacy loss is 6.41, training accuracy is 0.406250, time 233.342 samples/sec
+epoch 17, total_step 275320, total loss is 20.25 , inference loss is 13.84, weight deacy loss is 6.41, training accuracy is 0.281250, time 296.943 samples/sec
+epoch 17, total_step 275340, total loss is 15.18 , inference loss is 8.78, weight deacy loss is 6.41, training accuracy is 0.375000, time 224.406 samples/sec
+epoch 17, total_step 275360, total loss is 13.20 , inference loss is 6.79, weight deacy loss is 6.41, training accuracy is 0.406250, time 211.038 samples/sec
+epoch 17, total_step 275380, total loss is 11.64 , inference loss is 5.23, weight deacy loss is 6.41, training accuracy is 0.531250, time 221.561 samples/sec
+epoch 17, total_step 275400, total loss is 14.59 , inference loss is 8.18, weight deacy loss is 6.41, training accuracy is 0.437500, time 11.993 samples/sec
+epoch 17, total_step 275420, total loss is 13.16 , inference loss is 6.75, weight deacy loss is 6.41, training accuracy is 0.562500, time 223.903 samples/sec
+epoch 17, total_step 275440, total loss is 14.20 , inference loss is 7.79, weight deacy loss is 6.41, training accuracy is 0.375000, time 214.068 samples/sec
+epoch 17, total_step 275460, total loss is 16.49 , inference loss is 10.08, weight deacy loss is 6.41, training accuracy is 0.343750, time 217.216 samples/sec
+epoch 17, total_step 275480, total loss is 13.11 , inference loss is 6.70, weight deacy loss is 6.41, training accuracy is 0.531250, time 224.583 samples/sec2022-12-01 18:43:17.498392: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:145] Filling up shuffle buffer (this may take a while): 7067 of 10000
+2022-12-01 18:43:21.693740: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:195] Shuffle buffer filled.
+
+epoch 17, total_step 275500, total loss is 13.60 , inference loss is 7.20, weight deacy loss is 6.41, training accuracy is 0.468750, time 223.593 samples/sec
+epoch 17, total_step 275520, total loss is 16.31 , inference loss is 9.90, weight deacy loss is 6.41, training accuracy is 0.375000, time 218.644 samples/sec
+epoch 17, total_step 275540, total loss is 13.79 , inference loss is 7.38, weight deacy loss is 6.41, training accuracy is 0.406250, time 217.986 samples/sec
+epoch 17, total_step 275560, total loss is 15.17 , inference loss is 8.76, weight deacy loss is 6.41, training accuracy is 0.312500, time 226.460 samples/sec
+epoch 17, total_step 275580, total loss is 13.53 , inference loss is 7.12, weight deacy loss is 6.41, training accuracy is 0.531250, time 298.809 samples/sec
+epoch 17, total_step 275600, total loss is 13.75 , inference loss is 7.34, weight deacy loss is 6.41, training accuracy is 0.468750, time 215.070 samples/sec
+epoch 17, total_step 275620, total loss is 15.20 , inference loss is 8.79, weight deacy loss is 6.41, training accuracy is 0.343750, time 216.674 samples/sec
+epoch 17, total_step 275640, total loss is 13.59 , inference loss is 7.18, weight deacy loss is 6.41, training accuracy is 0.343750, time 217.216 samples/sec
+epoch 17, total_step 275660, total loss is 12.82 , inference loss is 6.41, weight deacy loss is 6.41, training accuracy is 0.468750, time 193.356 samples/sec
+epoch 17, total_step 275680, total loss is 15.00 , inference loss is 8.59, weight deacy loss is 6.41, training accuracy is 0.312500, time 211.853 samples/sec
+epoch 17, total_step 275700, total loss is 11.75 , inference loss is 5.35, weight deacy loss is 6.41, training accuracy is 0.312500, time 12.357 samples/sec
+epoch 17, total_step 275720, total loss is 19.44 , inference loss is 13.04, weight deacy loss is 6.41, training accuracy is 0.218750, time 194.302 samples/sec
+epoch 17, total_step 275740, total loss is 14.12 , inference loss is 7.71, weight deacy loss is 6.41, training accuracy is 0.500000, time 192.601 samples/sec
+epoch 17, total_step 275760, total loss is 16.94 , inference loss is 10.53, weight deacy loss is 6.41, training accuracy is 0.406250, time 225.100 samples/sec
+epoch 17, total_step 275780, total loss is 13.40 , inference loss is 7.00, weight deacy loss is 6.41, training accuracy is 0.531250, time 175.248 samples/sec
+epoch 17, total_step 275800, total loss is 13.82 , inference loss is 7.41, weight deacy loss is 6.41, training accuracy is 0.281250, time 191.875 samples/sec
+epoch 17, total_step 275820, total loss is 12.08 , inference loss is 5.67, weight deacy loss is 6.41, training accuracy is 0.437500, time 226.345 samples/sec
+epoch 17, total_step 275840, total loss is 11.45 , inference loss is 5.05, weight deacy loss is 6.41, training accuracy is 0.562500, time 176.902 samples/sec
+epoch 17, total_step 275860, total loss is 16.50 , inference loss is 10.10, weight deacy loss is 6.41, training accuracy is 0.406250, time 196.485 samples/sec
+epoch 17, total_step 275880, total loss is 9.35 , inference loss is 2.94, weight deacy loss is 6.41, training accuracy is 0.593750, time 195.323 samples/sec
+epoch 17, total_step 275900, total loss is 18.77 , inference loss is 12.37, weight deacy loss is 6.41, training accuracy is 0.343750, time 299.098 samples/sec
+epoch 17, total_step 275920, total loss is 11.67 , inference loss is 5.26, weight deacy loss is 6.41, training accuracy is 0.437500, time 198.397 samples/sec
+epoch 17, total_step 275940, total loss is 16.40 , inference loss is 10.00, weight deacy loss is 6.41, training accuracy is 0.250000, time 209.909 samples/sec
+End of epoch 17
+epoch 18, total_step 275960, total loss is 22.53 , inference loss is 16.13, weight deacy loss is 6.41, training accuracy is 0.000000, time 210.295 samples/sec
+epoch 18, total_step 275980, total loss is 13.51 , inference loss is 7.10, weight deacy loss is 6.41, training accuracy is 0.593750, time 226.382 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.229265999999985
+best_threshold_index 100 0.9835185185185186
+best_threshold_index 100 0.9827777777777778
+best_threshold_index 99 0.9827777777777778
+best_threshold_index 99 0.9840740740740741
+best_threshold_index 100 0.9827777777777778
+best_threshold_index 100 0.9822222222222222
+best_threshold_index 100 0.9833333333333333
+best_threshold_index 99 0.9829629629629629
+best_threshold_index 100 0.9822222222222222
+best_threshold_index 100 0.9818518518518519
+[lfw][276000]XNorm: 19.663508
+[lfw][276000]Accuracy-Flip: 0.98217+-0.00606
+test accuracy is: 0.9821666666666667
+epoch 18, total_step 276000, total loss is 9.61 , inference loss is 3.20, weight deacy loss is 6.41, training accuracy is 0.718750, time 15.802 samples/sec
+epoch 18, total_step 276020, total loss is 10.45 , inference loss is 4.04, weight deacy loss is 6.41, training accuracy is 0.687500, time 218.228 samples/sec
+epoch 18, total_step 276040, total loss is 8.75 , inference loss is 2.35, weight deacy loss is 6.41, training accuracy is 0.843750, time 219.747 samples/sec
+epoch 18, total_step 276060, total loss is 11.88 , inference loss is 5.47, weight deacy loss is 6.41, training accuracy is 0.687500, time 219.772 samples/sec
+epoch 18, total_step 276080, total loss is 6.60 , inference loss is 0.20, weight deacy loss is 6.41, training accuracy is 0.968750, time 226.989 samples/sec
+epoch 18, total_step 276100, total loss is 9.91 , inference loss is 3.51, weight deacy loss is 6.41, training accuracy is 0.750000, time 237.130 samples/sec
+epoch 18, total_step 276120, total loss is 7.56 , inference loss is 1.16, weight deacy loss is 6.41, training accuracy is 0.812500, time 218.091 samples/sec
+epoch 18, total_step 276140, total loss is 8.55 , inference loss is 2.15, weight deacy loss is 6.40, training accuracy is 0.875000, time 219.087 samples/sec
+epoch 18, total_step 276160, total loss is 14.10 , inference loss is 7.70, weight deacy loss is 6.40, training accuracy is 0.781250, time 222.922 samples/sec
+epoch 18, total_step 276180, total loss is 9.39 , inference loss is 2.98, weight deacy loss is 6.40, training accuracy is 0.812500, time 293.989 samples/sec
+epoch 18, total_step 276200, total loss is 8.86 , inference loss is 2.46, weight deacy loss is 6.40, training accuracy is 0.812500, time 210.064 samples/sec
+epoch 18, total_step 276220, total loss is 11.01 , inference loss is 4.61, weight deacy loss is 6.40, training accuracy is 0.718750, time 208.941 samples/sec
+epoch 18, total_step 276240, total loss is 11.65 , inference loss is 5.25, weight deacy loss is 6.40, training accuracy is 0.750000, time 222.893 samples/sec
+epoch 18, total_step 276260, total loss is 10.44 , inference loss is 4.03, weight deacy loss is 6.40, training accuracy is 0.750000, time 213.000 samples/sec
+epoch 18, total_step 276280, total loss is 12.10 , inference loss is 5.70, weight deacy loss is 6.40, training accuracy is 0.687500, time 213.691 samples/sec
+epoch 18, total_step 276300, total loss is 14.68 , inference loss is 8.28, weight deacy loss is 6.40, training accuracy is 0.593750, time 12.758 samples/sec
+epoch 18, total_step 276320, total loss is 9.19 , inference loss is 2.79, weight deacy loss is 6.40, training accuracy is 0.750000, time 299.136 samples/sec
+epoch 18, total_step 276340, total loss is 8.84 , inference loss is 2.44, weight deacy loss is 6.41, training accuracy is 0.812500, time 222.370 samples/sec
+epoch 18, total_step 276360, total loss is 12.72 , inference loss is 6.32, weight deacy loss is 6.41, training accuracy is 0.750000, time 219.946 samples/sec
+epoch 18, total_step 276380, total loss is 8.94 , inference loss is 2.54, weight deacy loss is 6.41, training accuracy is 0.781250, time 225.432 samples/sec
+epoch 18, total_step 276400, total loss is 10.48 , inference loss is 4.07, weight deacy loss is 6.41, training accuracy is 0.843750, time 223.431 samples/sec
+epoch 18, total_step 276420, total loss is 10.51 , inference loss is 4.11, weight deacy loss is 6.41, training accuracy is 0.593750, time 219.787 samples/sec
+epoch 18, total_step 276440, total loss is 12.47 , inference loss is 6.06, weight deacy loss is 6.41, training accuracy is 0.750000, time 240.632 samples/sec
+epoch 18, total_step 276460, total loss is 8.23 , inference loss is 1.82, weight deacy loss is 6.41, training accuracy is 0.718750, time 223.107 samples/sec
+epoch 18, total_step 276480, total loss is 6.99 , inference loss is 0.59, weight deacy loss is 6.41, training accuracy is 0.812500, time 231.606 samples/sec
+epoch 18, total_step 276500, total loss is 15.19 , inference loss is 8.78, weight deacy loss is 6.41, training accuracy is 0.625000, time 216.295 samples/sec
+epoch 18, total_step 276520, total loss is 8.14 , inference loss is 1.73, weight deacy loss is 6.41, training accuracy is 0.750000, time 216.971 samples/sec
+epoch 18, total_step 276540, total loss is 11.72 , inference loss is 5.31, weight deacy loss is 6.41, training accuracy is 0.750000, time 213.466 samples/sec
+epoch 18, total_step 276560, total loss is 13.50 , inference loss is 7.09, weight deacy loss is 6.41, training accuracy is 0.625000, time 222.628 samples/sec
+epoch 18, total_step 276580, total loss is 9.33 , inference loss is 2.92, weight deacy loss is 6.41, training accuracy is 0.718750, time 222.946 samples/sec
+epoch 18, total_step 276600, total loss is 9.29 , inference loss is 2.88, weight deacy loss is 6.41, training accuracy is 0.750000, time 12.616 samples/sec
+epoch 18, total_step 276620, total loss is 12.82 , inference loss is 6.41, weight deacy loss is 6.41, training accuracy is 0.625000, time 223.931 samples/sec
+epoch 18, total_step 276640, total loss is 10.36 , inference loss is 3.95, weight deacy loss is 6.41, training accuracy is 0.875000, time 221.052 samples/sec
+epoch 18, total_step 276660, total loss is 12.23 , inference loss is 5.82, weight deacy loss is 6.41, training accuracy is 0.656250, time 212.067 samples/sec
+epoch 18, total_step 276680, total loss is 12.58 , inference loss is 6.17, weight deacy loss is 6.41, training accuracy is 0.687500, time 224.348 samples/sec
+epoch 18, total_step 276700, total loss is 8.47 , inference loss is 2.06, weight deacy loss is 6.41, training accuracy is 0.843750, time 218.945 samples/sec
+epoch 18, total_step 276720, total loss is 12.40 , inference loss is 5.99, weight deacy loss is 6.41, training accuracy is 0.750000, time 223.272 samples/sec
+epoch 18, total_step 276740, total loss is 9.43 , inference loss is 3.02, weight deacy loss is 6.41, training accuracy is 0.812500, time 225.949 samples/sec
+epoch 18, total_step 276760, total loss is 11.37 , inference loss is 4.96, weight deacy loss is 6.41, training accuracy is 0.562500, time 220.156 samples/sec
+epoch 18, total_step 276780, total loss is 11.87 , inference loss is 5.46, weight deacy loss is 6.41, training accuracy is 0.718750, time 224.036 samples/sec
+epoch 18, total_step 276800, total loss is 12.53 , inference loss is 6.12, weight deacy loss is 6.41, training accuracy is 0.687500, time 223.456 samples/sec
+epoch 18, total_step 276820, total loss is 12.94 , inference loss is 6.53, weight deacy loss is 6.41, training accuracy is 0.437500, time 219.342 samples/sec
+epoch 18, total_step 276840, total loss is 10.08 , inference loss is 3.67, weight deacy loss is 6.41, training accuracy is 0.718750, time 224.163 samples/sec
+epoch 18, total_step 276860, total loss is 10.19 , inference loss is 3.78, weight deacy loss is 6.41, training accuracy is 0.625000, time 228.523 samples/sec
+epoch 18, total_step 276880, total loss is 10.65 , inference loss is 4.23, weight deacy loss is 6.41, training accuracy is 0.687500, time 224.958 samples/sec
+epoch 18, total_step 276900, total loss is 12.24 , inference loss is 5.82, weight deacy loss is 6.41, training accuracy is 0.625000, time 11.665 samples/sec
+epoch 18, total_step 276920, total loss is 12.09 , inference loss is 5.68, weight deacy loss is 6.41, training accuracy is 0.562500, time 229.428 samples/sec
+epoch 18, total_step 276940, total loss is 8.63 , inference loss is 2.22, weight deacy loss is 6.41, training accuracy is 0.843750, time 224.625 samples/sec
+epoch 18, total_step 276960, total loss is 11.84 , inference loss is 5.43, weight deacy loss is 6.41, training accuracy is 0.531250, time 226.805 samples/sec
+epoch 18, total_step 276980, total loss is 14.24 , inference loss is 7.83, weight deacy loss is 6.41, training accuracy is 0.562500, time 227.463 samples/sec
+epoch 18, total_step 277000, total loss is 9.33 , inference loss is 2.92, weight deacy loss is 6.42, training accuracy is 0.656250, time 225.319 samples/sec
+epoch 18, total_step 277020, total loss is 14.94 , inference loss is 8.52, weight deacy loss is 6.42, training accuracy is 0.531250, time 235.550 samples/sec
+epoch 18, total_step 277040, total loss is 13.03 , inference loss is 6.61, weight deacy loss is 6.42, training accuracy is 0.625000, time 223.495 samples/sec
+epoch 18, total_step 277060, total loss is 11.82 , inference loss is 5.40, weight deacy loss is 6.42, training accuracy is 0.656250, time 213.422 samples/sec
+epoch 18, total_step 277080, total loss is 14.12 , inference loss is 7.70, weight deacy loss is 6.42, training accuracy is 0.468750, time 224.228 samples/sec
+epoch 18, total_step 277100, total loss is 12.78 , inference loss is 6.36, weight deacy loss is 6.42, training accuracy is 0.625000, time 226.756 samples/sec
+epoch 18, total_step 277120, total loss is 16.19 , inference loss is 9.78, weight deacy loss is 6.42, training accuracy is 0.656250, time 209.091 samples/sec
+epoch 18, total_step 277140, total loss is 13.21 , inference loss is 6.79, weight deacy loss is 6.42, training accuracy is 0.562500, time 211.266 samples/sec
+epoch 18, total_step 277160, total loss is 12.45 , inference loss is 6.03, weight deacy loss is 6.42, training accuracy is 0.656250, time 298.790 samples/sec
+epoch 18, total_step 277180, total loss is 8.82 , inference loss is 2.40, weight deacy loss is 6.42, training accuracy is 0.562500, time 229.986 samples/sec
+epoch 18, total_step 277200, total loss is 14.03 , inference loss is 7.62, weight deacy loss is 6.42, training accuracy is 0.656250, time 12.146 samples/sec
+epoch 18, total_step 277220, total loss is 11.79 , inference loss is 5.37, weight deacy loss is 6.42, training accuracy is 0.562500, time 223.385 samples/sec
+epoch 18, total_step 277240, total loss is 12.80 , inference loss is 6.39, weight deacy loss is 6.42, training accuracy is 0.625000, time 225.372 samples/sec
+epoch 18, total_step 277260, total loss is 12.69 , inference loss is 6.27, weight deacy loss is 6.42, training accuracy is 0.562500, time 229.889 samples/sec
+epoch 18, total_step 277280, total loss is 13.50 , inference loss is 7.08, weight deacy loss is 6.42, training accuracy is 0.500000, time 218.959 samples/sec
+epoch 18, total_step 277300, total loss is 9.63 , inference loss is 3.21, weight deacy loss is 6.42, training accuracy is 0.625000, time 216.059 samples/sec
+epoch 18, total_step 277320, total loss is 10.43 , inference loss is 4.01, weight deacy loss is 6.42, training accuracy is 0.562500, time 217.883 samples/sec
+epoch 18, total_step 277340, total loss is 11.37 , inference loss is 4.95, weight deacy loss is 6.42, training accuracy is 0.562500, time 216.286 samples/sec
+epoch 18, total_step 277360, total loss is 11.05 , inference loss is 4.63, weight deacy loss is 6.42, training accuracy is 0.562500, time 228.951 samples/sec
+epoch 18, total_step 277380, total loss is 11.10 , inference loss is 4.68, weight deacy loss is 6.42, training accuracy is 0.562500, time 217.099 samples/sec
+epoch 18, total_step 277400, total loss is 10.55 , inference loss is 4.13, weight deacy loss is 6.42, training accuracy is 0.656250, time 222.725 samples/sec
+epoch 18, total_step 277420, total loss is 11.66 , inference loss is 5.24, weight deacy loss is 6.42, training accuracy is 0.500000, time 298.744 samples/sec
+epoch 18, total_step 277440, total loss is 12.58 , inference loss is 6.16, weight deacy loss is 6.42, training accuracy is 0.593750, time 225.325 samples/sec
+epoch 18, total_step 277460, total loss is 12.49 , inference loss is 6.07, weight deacy loss is 6.42, training accuracy is 0.593750, time 211.558 samples/sec
+epoch 18, total_step 277480, total loss is 13.33 , inference loss is 6.90, weight deacy loss is 6.42, training accuracy is 0.593750, time 223.337 samples/sec
+epoch 18, total_step 277500, total loss is 10.55 , inference loss is 4.12, weight deacy loss is 6.42, training accuracy is 0.718750, time 12.727 samples/sec
+epoch 18, total_step 277520, total loss is 13.11 , inference loss is 6.69, weight deacy loss is 6.42, training accuracy is 0.750000, time 244.111 samples/sec
+epoch 18, total_step 277540, total loss is 12.83 , inference loss is 6.41, weight deacy loss is 6.42, training accuracy is 0.656250, time 228.173 samples/sec
+epoch 18, total_step 277560, total loss is 13.02 , inference loss is 6.59, weight deacy loss is 6.42, training accuracy is 0.593750, time 220.383 samples/sec
+epoch 18, total_step 277580, total loss is 10.24 , inference loss is 3.82, weight deacy loss is 6.42, training accuracy is 0.625000, time 228.047 samples/sec
+epoch 18, total_step 277600, total loss is 14.66 , inference loss is 8.24, weight deacy loss is 6.42, training accuracy is 0.562500, time 216.884 samples/sec
+epoch 18, total_step 277620, total loss is 12.35 , inference loss is 5.93, weight deacy loss is 6.42, training accuracy is 0.687500, time 188.600 samples/sec
+epoch 18, total_step 277640, total loss is 11.42 , inference loss is 5.00, weight deacy loss is 6.42, training accuracy is 0.593750, time 223.874 samples/sec
+epoch 18, total_step 277660, total loss is 12.70 , inference loss is 6.28, weight deacy loss is 6.42, training accuracy is 0.562500, time 225.143 samples/sec
+epoch 18, total_step 277680, total loss is 12.56 , inference loss is 6.13, weight deacy loss is 6.42, training accuracy is 0.718750, time 233.606 samples/sec
+epoch 18, total_step 277700, total loss is 9.18 , inference loss is 2.75, weight deacy loss is 6.43, training accuracy is 0.687500, time 237.954 samples/sec
+epoch 18, total_step 277720, total loss is 12.18 , inference loss is 5.76, weight deacy loss is 6.43, training accuracy is 0.656250, time 227.970 samples/sec
+epoch 18, total_step 277740, total loss is 10.18 , inference loss is 3.76, weight deacy loss is 6.43, training accuracy is 0.656250, time 216.904 samples/sec
+epoch 18, total_step 277760, total loss is 11.65 , inference loss is 5.23, weight deacy loss is 6.43, training accuracy is 0.468750, time 220.427 samples/sec
+epoch 18, total_step 277780, total loss is 8.88 , inference loss is 2.46, weight deacy loss is 6.43, training accuracy is 0.656250, time 225.399 samples/sec
+epoch 18, total_step 277800, total loss is 19.11 , inference loss is 12.68, weight deacy loss is 6.43, training accuracy is 0.500000, time 12.357 samples/sec
+epoch 18, total_step 277820, total loss is 12.11 , inference loss is 5.69, weight deacy loss is 6.43, training accuracy is 0.406250, time 200.844 samples/sec
+epoch 18, total_step 277840, total loss is 15.64 , inference loss is 9.21, weight deacy loss is 6.43, training accuracy is 0.531250, time 217.300 samples/sec
+epoch 18, total_step 277860, total loss is 12.85 , inference loss is 6.43, weight deacy loss is 6.43, training accuracy is 0.562500, time 236.690 samples/sec
+epoch 18, total_step 277880, total loss is 9.62 , inference loss is 3.19, weight deacy loss is 6.43, training accuracy is 0.687500, time 224.945 samples/sec
+epoch 18, total_step 277900, total loss is 15.68 , inference loss is 9.25, weight deacy loss is 6.43, training accuracy is 0.625000, time 226.003 samples/sec
+epoch 18, total_step 277920, total loss is 10.65 , inference loss is 4.22, weight deacy loss is 6.43, training accuracy is 0.562500, time 216.219 samples/sec
+epoch 18, total_step 277940, total loss is 8.03 , inference loss is 1.60, weight deacy loss is 6.43, training accuracy is 0.843750, time 183.721 samples/sec
+epoch 18, total_step 277960, total loss is 9.13 , inference loss is 2.70, weight deacy loss is 6.43, training accuracy is 0.812500, time 228.555 samples/sec
+epoch 18, total_step 277980, total loss is 12.04 , inference loss is 5.61, weight deacy loss is 6.43, training accuracy is 0.656250, time 218.123 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.353662000000003
+best_threshold_index 120 0.9864814814814815
+best_threshold_index 120 0.9861111111111112
+best_threshold_index 120 0.9855555555555555
+best_threshold_index 120 0.9859259259259259
+best_threshold_index 120 0.9861111111111112
+best_threshold_index 120 0.9848148148148148
+best_threshold_index 120 0.9855555555555555
+best_threshold_index 120 0.9853703703703703
+best_threshold_index 120 0.9842592592592593
+best_threshold_index 120 0.9848148148148148
+[lfw][278000]XNorm: 21.163878
+[lfw][278000]Accuracy-Flip: 0.98550+-0.00597
+test accuracy is: 0.9855
+epoch 18, total_step 278000, total loss is 11.62 , inference loss is 5.19, weight deacy loss is 6.43, training accuracy is 0.625000, time 18.161 samples/sec
+epoch 18, total_step 278020, total loss is 10.94 , inference loss is 4.52, weight deacy loss is 6.43, training accuracy is 0.656250, time 229.200 samples/sec
+epoch 18, total_step 278040, total loss is 10.49 , inference loss is 4.06, weight deacy loss is 6.43, training accuracy is 0.687500, time 212.139 samples/sec
+epoch 18, total_step 278060, total loss is 12.90 , inference loss is 6.47, weight deacy loss is 6.43, training accuracy is 0.562500, time 217.573 samples/sec
+epoch 18, total_step 278080, total loss is 10.17 , inference loss is 3.74, weight deacy loss is 6.43, training accuracy is 0.562500, time 224.106 samples/sec
+epoch 18, total_step 278100, total loss is 11.90 , inference loss is 5.47, weight deacy loss is 6.43, training accuracy is 0.562500, time 13.686 samples/sec
+epoch 18, total_step 278120, total loss is 8.32 , inference loss is 1.90, weight deacy loss is 6.43, training accuracy is 0.750000, time 219.523 samples/sec
+epoch 18, total_step 278140, total loss is 10.00 , inference loss is 3.57, weight deacy loss is 6.43, training accuracy is 0.718750, time 219.071 samples/sec
+epoch 18, total_step 278160, total loss is 11.15 , inference loss is 4.72, weight deacy loss is 6.43, training accuracy is 0.593750, time 217.932 samples/sec
+epoch 18, total_step 278180, total loss is 12.41 , inference loss is 5.98, weight deacy loss is 6.43, training accuracy is 0.656250, time 238.061 samples/sec
+epoch 18, total_step 278200, total loss is 13.78 , inference loss is 7.35, weight deacy loss is 6.43, training accuracy is 0.656250, time 298.735 samples/sec
+epoch 18, total_step 278220, total loss is 12.39 , inference loss is 5.97, weight deacy loss is 6.43, training accuracy is 0.593750, time 235.656 samples/sec
+epoch 18, total_step 278240, total loss is 11.37 , inference loss is 4.94, weight deacy loss is 6.43, training accuracy is 0.750000, time 218.082 samples/sec
+epoch 18, total_step 278260, total loss is 8.41 , inference loss is 1.99, weight deacy loss is 6.43, training accuracy is 0.781250, time 224.539 samples/sec
+epoch 18, total_step 278280, total loss is 10.17 , inference loss is 3.74, weight deacy loss is 6.43, training accuracy is 0.500000, time 197.883 samples/sec
+epoch 18, total_step 278300, total loss is 11.96 , inference loss is 5.53, weight deacy loss is 6.43, training accuracy is 0.562500, time 216.990 samples/sec
+epoch 18, total_step 278320, total loss is 9.49 , inference loss is 3.06, weight deacy loss is 6.43, training accuracy is 0.750000, time 215.210 samples/sec
+epoch 18, total_step 278340, total loss is 10.88 , inference loss is 4.45, weight deacy loss is 6.43, training accuracy is 0.718750, time 221.598 samples/sec
+epoch 18, total_step 278360, total loss is 10.01 , inference loss is 3.58, weight deacy loss is 6.43, training accuracy is 0.687500, time 233.179 samples/sec
+epoch 18, total_step 278380, total loss is 10.72 , inference loss is 4.29, weight deacy loss is 6.43, training accuracy is 0.625000, time 227.327 samples/sec
+epoch 18, total_step 278400, total loss is 12.08 , inference loss is 5.65, weight deacy loss is 6.43, training accuracy is 0.687500, time 12.401 samples/sec
+epoch 18, total_step 278420, total loss is 11.50 , inference loss is 5.07, weight deacy loss is 6.43, training accuracy is 0.781250, time 227.342 samples/sec
+epoch 18, total_step 278440, total loss is 11.52 , inference loss is 5.09, weight deacy loss is 6.43, training accuracy is 0.687500, time 226.467 samples/sec
+epoch 18, total_step 278460, total loss is 8.60 , inference loss is 2.17, weight deacy loss is 6.43, training accuracy is 0.687500, time 298.289 samples/sec
+epoch 18, total_step 278480, total loss is 7.37 , inference loss is 0.95, weight deacy loss is 6.43, training accuracy is 0.812500, time 217.801 samples/sec
+epoch 18, total_step 278500, total loss is 10.81 , inference loss is 4.39, weight deacy loss is 6.43, training accuracy is 0.750000, time 226.267 samples/sec
+epoch 18, total_step 278520, total loss is 9.65 , inference loss is 3.22, weight deacy loss is 6.43, training accuracy is 0.687500, time 223.645 samples/sec
+epoch 18, total_step 278540, total loss is 9.47 , inference loss is 3.04, weight deacy loss is 6.43, training accuracy is 0.718750, time 217.011 samples/sec
+epoch 18, total_step 278560, total loss is 9.31 , inference loss is 2.89, weight deacy loss is 6.43, training accuracy is 0.718750, time 222.646 samples/sec
+epoch 18, total_step 278580, total loss is 8.46 , inference loss is 2.04, weight deacy loss is 6.43, training accuracy is 0.781250, time 215.129 samples/sec
+epoch 18, total_step 278600, total loss is 9.22 , inference loss is 2.79, weight deacy loss is 6.43, training accuracy is 0.718750, time 223.517 samples/sec
+epoch 18, total_step 278620, total loss is 9.64 , inference loss is 3.21, weight deacy loss is 6.43, training accuracy is 0.687500, time 221.082 samples/sec
+epoch 18, total_step 278640, total loss is 10.35 , inference loss is 3.93, weight deacy loss is 6.43, training accuracy is 0.750000, time 221.318 samples/sec
+epoch 18, total_step 278660, total loss is 11.34 , inference loss is 4.92, weight deacy loss is 6.43, training accuracy is 0.781250, time 223.797 samples/sec
+epoch 18, total_step 278680, total loss is 8.99 , inference loss is 2.56, weight deacy loss is 6.43, training accuracy is 0.562500, time 237.022 samples/sec
+epoch 18, total_step 278700, total loss is 10.02 , inference loss is 3.60, weight deacy loss is 6.43, training accuracy is 0.562500, time 12.033 samples/sec
+epoch 18, total_step 278720, total loss is 8.03 , inference loss is 1.61, weight deacy loss is 6.43, training accuracy is 0.718750, time 299.034 samples/sec
+epoch 18, total_step 278740, total loss is 10.96 , inference loss is 4.54, weight deacy loss is 6.43, training accuracy is 0.562500, time 218.082 samples/sec
+epoch 18, total_step 278760, total loss is 9.08 , inference loss is 2.66, weight deacy loss is 6.43, training accuracy is 0.687500, time 227.138 samples/sec
+epoch 18, total_step 278780, total loss is 8.58 , inference loss is 2.15, weight deacy loss is 6.43, training accuracy is 0.812500, time 223.177 samples/sec
+epoch 18, total_step 278800, total loss is 8.65 , inference loss is 2.22, weight deacy loss is 6.43, training accuracy is 0.718750, time 223.273 samples/sec
+epoch 18, total_step 278820, total loss is 8.37 , inference loss is 1.94, weight deacy loss is 6.43, training accuracy is 0.781250, time 216.738 samples/sec
+epoch 18, total_step 278840, total loss is 8.29 , inference loss is 1.87, weight deacy loss is 6.43, training accuracy is 0.718750, time 225.210 samples/sec
+epoch 18, total_step 278860, total loss is 11.45 , inference loss is 5.02, weight deacy loss is 6.42, training accuracy is 0.625000, time 174.943 samples/sec
+epoch 18, total_step 278880, total loss is 8.87 , inference loss is 2.45, weight deacy loss is 6.42, training accuracy is 0.687500, time 226.632 samples/sec
+epoch 18, total_step 278900, total loss is 12.23 , inference loss is 5.80, weight deacy loss is 6.42, training accuracy is 0.625000, time 217.027 samples/sec
+epoch 18, total_step 278920, total loss is 7.83 , inference loss is 1.40, weight deacy loss is 6.42, training accuracy is 0.781250, time 238.762 samples/sec
+epoch 18, total_step 278940, total loss is 10.64 , inference loss is 4.22, weight deacy loss is 6.42, training accuracy is 0.531250, time 223.346 samples/sec
+epoch 18, total_step 278960, total loss is 10.90 , inference loss is 4.47, weight deacy loss is 6.42, training accuracy is 0.687500, time 225.596 samples/sec
+epoch 18, total_step 278980, total loss is 9.15 , inference loss is 2.72, weight deacy loss is 6.42, training accuracy is 0.500000, time 225.381 samples/sec
+epoch 18, total_step 279000, total loss is 10.25 , inference loss is 3.83, weight deacy loss is 6.42, training accuracy is 0.750000, time 13.265 samples/sec
+epoch 18, total_step 279020, total loss is 8.66 , inference loss is 2.23, weight deacy loss is 6.42, training accuracy is 0.781250, time 227.018 samples/sec
+epoch 18, total_step 279040, total loss is 10.10 , inference loss is 3.67, weight deacy loss is 6.42, training accuracy is 0.656250, time 223.122 samples/sec
+epoch 18, total_step 279060, total loss is 9.75 , inference loss is 3.32, weight deacy loss is 6.42, training accuracy is 0.875000, time 226.131 samples/sec
+epoch 18, total_step 279080, total loss is 11.77 , inference loss is 5.34, weight deacy loss is 6.42, training accuracy is 0.718750, time 221.083 samples/sec
+epoch 18, total_step 279100, total loss is 12.52 , inference loss is 6.09, weight deacy loss is 6.42, training accuracy is 0.562500, time 226.020 samples/sec
+epoch 18, total_step 279120, total loss is 12.34 , inference loss is 5.92, weight deacy loss is 6.42, training accuracy is 0.656250, time 217.144 samples/sec
+epoch 18, total_step 279140, total loss is 10.12 , inference loss is 3.69, weight deacy loss is 6.42, training accuracy is 0.593750, time 222.798 samples/sec
+epoch 18, total_step 279160, total loss is 11.21 , inference loss is 4.79, weight deacy loss is 6.43, training accuracy is 0.562500, time 225.281 samples/sec
+epoch 18, total_step 279180, total loss is 11.22 , inference loss is 4.80, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.454 samples/sec
+epoch 18, total_step 279200, total loss is 14.07 , inference loss is 7.65, weight deacy loss is 6.43, training accuracy is 0.468750, time 224.421 samples/sec
+epoch 18, total_step 279220, total loss is 11.62 , inference loss is 5.19, weight deacy loss is 6.43, training accuracy is 0.593750, time 229.876 samples/sec
+epoch 18, total_step 279240, total loss is 11.28 , inference loss is 4.85, weight deacy loss is 6.43, training accuracy is 0.656250, time 245.986 samples/sec
+epoch 18, total_step 279260, total loss is 10.83 , inference loss is 4.40, weight deacy loss is 6.43, training accuracy is 0.562500, time 237.678 samples/sec
+epoch 18, total_step 279280, total loss is 14.20 , inference loss is 7.77, weight deacy loss is 6.43, training accuracy is 0.562500, time 224.786 samples/sec
+epoch 18, total_step 279300, total loss is 10.00 , inference loss is 3.58, weight deacy loss is 6.43, training accuracy is 0.687500, time 11.882 samples/sec
+epoch 18, total_step 279320, total loss is 11.67 , inference loss is 5.25, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.794 samples/sec
+epoch 18, total_step 279340, total loss is 9.31 , inference loss is 2.89, weight deacy loss is 6.43, training accuracy is 0.750000, time 221.854 samples/sec
+epoch 18, total_step 279360, total loss is 12.30 , inference loss is 5.87, weight deacy loss is 6.43, training accuracy is 0.625000, time 223.044 samples/sec
+epoch 18, total_step 279380, total loss is 11.08 , inference loss is 4.66, weight deacy loss is 6.43, training accuracy is 0.562500, time 215.604 samples/sec
+epoch 18, total_step 279400, total loss is 10.96 , inference loss is 4.53, weight deacy loss is 6.43, training accuracy is 0.437500, time 224.865 samples/sec
+epoch 18, total_step 279420, total loss is 15.48 , inference loss is 9.05, weight deacy loss is 6.43, training accuracy is 0.531250, time 210.500 samples/sec
+epoch 18, total_step 279440, total loss is 10.96 , inference loss is 4.54, weight deacy loss is 6.43, training accuracy is 0.625000, time 217.745 samples/sec
+epoch 18, total_step 279460, total loss is 12.63 , inference loss is 6.20, weight deacy loss is 6.43, training accuracy is 0.468750, time 220.310 samples/sec
+epoch 18, total_step 279480, total loss is 11.04 , inference loss is 4.61, weight deacy loss is 6.43, training accuracy is 0.625000, time 225.218 samples/sec
+epoch 18, total_step 279500, total loss is 14.43 , inference loss is 8.00, weight deacy loss is 6.43, training accuracy is 0.375000, time 226.358 samples/sec
+epoch 18, total_step 279520, total loss is 8.85 , inference loss is 2.42, weight deacy loss is 6.43, training accuracy is 0.656250, time 224.748 samples/sec
+epoch 18, total_step 279540, total loss is 11.90 , inference loss is 5.48, weight deacy loss is 6.43, training accuracy is 0.656250, time 226.259 samples/sec
+epoch 18, total_step 279560, total loss is 12.17 , inference loss is 5.74, weight deacy loss is 6.43, training accuracy is 0.625000, time 299.496 samples/sec
+epoch 18, total_step 279580, total loss is 10.22 , inference loss is 3.79, weight deacy loss is 6.43, training accuracy is 0.625000, time 218.853 samples/sec
+epoch 18, total_step 279600, total loss is 11.54 , inference loss is 5.11, weight deacy loss is 6.43, training accuracy is 0.593750, time 12.039 samples/sec
+epoch 18, total_step 279620, total loss is 17.39 , inference loss is 10.96, weight deacy loss is 6.43, training accuracy is 0.531250, time 220.005 samples/sec
+epoch 18, total_step 279640, total loss is 15.38 , inference loss is 8.95, weight deacy loss is 6.43, training accuracy is 0.656250, time 224.788 samples/sec
+epoch 18, total_step 279660, total loss is 11.81 , inference loss is 5.38, weight deacy loss is 6.43, training accuracy is 0.531250, time 219.356 samples/sec
+epoch 18, total_step 279680, total loss is 15.13 , inference loss is 8.70, weight deacy loss is 6.43, training accuracy is 0.625000, time 226.625 samples/sec
+epoch 18, total_step 279700, total loss is 14.39 , inference loss is 7.96, weight deacy loss is 6.43, training accuracy is 0.531250, time 224.875 samples/sec
+epoch 18, total_step 279720, total loss is 11.45 , inference loss is 5.02, weight deacy loss is 6.43, training accuracy is 0.625000, time 228.877 samples/sec
+epoch 18, total_step 279740, total loss is 10.97 , inference loss is 4.54, weight deacy loss is 6.43, training accuracy is 0.656250, time 225.869 samples/sec
+epoch 18, total_step 279760, total loss is 11.71 , inference loss is 5.29, weight deacy loss is 6.43, training accuracy is 0.625000, time 194.902 samples/sec
+epoch 18, total_step 279780, total loss is 11.04 , inference loss is 4.62, weight deacy loss is 6.43, training accuracy is 0.531250, time 230.466 samples/sec
+epoch 18, total_step 279800, total loss is 13.70 , inference loss is 7.27, weight deacy loss is 6.43, training accuracy is 0.562500, time 228.567 samples/sec
+epoch 18, total_step 279820, total loss is 11.51 , inference loss is 5.08, weight deacy loss is 6.43, training accuracy is 0.593750, time 298.439 samples/sec
+epoch 18, total_step 279840, total loss is 15.98 , inference loss is 9.56, weight deacy loss is 6.43, training accuracy is 0.437500, time 217.445 samples/sec
+epoch 18, total_step 279860, total loss is 16.91 , inference loss is 10.48, weight deacy loss is 6.43, training accuracy is 0.625000, time 221.145 samples/sec
+epoch 18, total_step 279880, total loss is 15.18 , inference loss is 8.75, weight deacy loss is 6.43, training accuracy is 0.500000, time 225.502 samples/sec
+epoch 18, total_step 279900, total loss is 10.24 , inference loss is 3.81, weight deacy loss is 6.43, training accuracy is 0.531250, time 12.617 samples/sec
+epoch 18, total_step 279920, total loss is 12.82 , inference loss is 6.39, weight deacy loss is 6.43, training accuracy is 0.625000, time 220.588 samples/sec
+epoch 18, total_step 279940, total loss is 15.20 , inference loss is 8.77, weight deacy loss is 6.43, training accuracy is 0.593750, time 214.713 samples/sec
+epoch 18, total_step 279960, total loss is 14.86 , inference loss is 8.43, weight deacy loss is 6.43, training accuracy is 0.531250, time 213.364 samples/sec
+epoch 18, total_step 279980, total loss is 12.80 , inference loss is 6.37, weight deacy loss is 6.43, training accuracy is 0.500000, time 219.999 samples/sec
+testing verification..
+(12000, 512)
+infer time 23.99685099999998
+best_threshold_index 126 0.9872222222222222
+best_threshold_index 126 0.9859259259259259
+best_threshold_index 126 0.9864814814814815
+best_threshold_index 126 0.9874074074074074
+best_threshold_index 126 0.9857407407407407
+best_threshold_index 126 0.9859259259259259
+best_threshold_index 126 0.9868518518518519
+best_threshold_index 126 0.9855555555555555
+best_threshold_index 126 0.985
+best_threshold_index 126 0.9855555555555555
+[lfw][280000]XNorm: 20.391007
+[lfw][280000]Accuracy-Flip: 0.98617+-0.00675
+test accuracy is: 0.9861666666666666
+epoch 18, total_step 280000, total loss is 13.46 , inference loss is 7.03, weight deacy loss is 6.43, training accuracy is 0.500000, time 16.822 samples/sec
+epoch 18, total_step 280020, total loss is 13.48 , inference loss is 7.05, weight deacy loss is 6.43, training accuracy is 0.531250, time 217.752 samples/sec
+epoch 18, total_step 280040, total loss is 13.34 , inference loss is 6.91, weight deacy loss is 6.43, training accuracy is 0.531250, time 217.584 samples/sec
+epoch 18, total_step 280060, total loss is 10.86 , inference loss is 4.43, weight deacy loss is 6.43, training accuracy is 0.656250, time 219.372 samples/sec
+epoch 18, total_step 280080, total loss is 9.71 , inference loss is 3.29, weight deacy loss is 6.43, training accuracy is 0.687500, time 218.109 samples/sec
+epoch 18, total_step 280100, total loss is 9.14 , inference loss is 2.71, weight deacy loss is 6.43, training accuracy is 0.687500, time 218.006 samples/sec
+epoch 18, total_step 280120, total loss is 16.13 , inference loss is 9.70, weight deacy loss is 6.43, training accuracy is 0.468750, time 226.996 samples/sec
+epoch 18, total_step 280140, total loss is 9.57 , inference loss is 3.14, weight deacy loss is 6.43, training accuracy is 0.625000, time 225.343 samples/sec
+epoch 18, total_step 280160, total loss is 9.05 , inference loss is 2.62, weight deacy loss is 6.43, training accuracy is 0.687500, time 235.487 samples/sec
+epoch 18, total_step 280180, total loss is 9.97 , inference loss is 3.54, weight deacy loss is 6.43, training accuracy is 0.656250, time 221.664 samples/sec
+epoch 18, total_step 280200, total loss is 13.54 , inference loss is 7.12, weight deacy loss is 6.43, training accuracy is 0.593750, time 12.213 samples/sec
+epoch 18, total_step 280220, total loss is 9.27 , inference loss is 2.84, weight deacy loss is 6.43, training accuracy is 0.687500, time 221.694 samples/sec
+epoch 18, total_step 280240, total loss is 11.79 , inference loss is 5.36, weight deacy loss is 6.43, training accuracy is 0.562500, time 238.413 samples/sec
+epoch 18, total_step 280260, total loss is 9.21 , inference loss is 2.78, weight deacy loss is 6.43, training accuracy is 0.718750, time 299.297 samples/sec
+epoch 18, total_step 280280, total loss is 12.85 , inference loss is 6.42, weight deacy loss is 6.43, training accuracy is 0.718750, time 215.844 samples/sec
+epoch 18, total_step 280300, total loss is 11.91 , inference loss is 5.48, weight deacy loss is 6.43, training accuracy is 0.531250, time 229.966 samples/sec
+epoch 18, total_step 280320, total loss is 14.94 , inference loss is 8.51, weight deacy loss is 6.43, training accuracy is 0.500000, time 218.540 samples/sec
+epoch 18, total_step 280340, total loss is 14.01 , inference loss is 7.58, weight deacy loss is 6.43, training accuracy is 0.500000, time 221.491 samples/sec
+epoch 18, total_step 280360, total loss is 10.79 , inference loss is 4.36, weight deacy loss is 6.43, training accuracy is 0.656250, time 227.518 samples/sec
+epoch 18, total_step 280380, total loss is 13.87 , inference loss is 7.44, weight deacy loss is 6.43, training accuracy is 0.343750, time 221.985 samples/sec
+epoch 18, total_step 280400, total loss is 14.30 , inference loss is 7.87, weight deacy loss is 6.43, training accuracy is 0.593750, time 215.566 samples/sec
+epoch 18, total_step 280420, total loss is 13.46 , inference loss is 7.03, weight deacy loss is 6.43, training accuracy is 0.500000, time 228.905 samples/sec
+epoch 18, total_step 280440, total loss is 13.42 , inference loss is 6.99, weight deacy loss is 6.43, training accuracy is 0.593750, time 220.548 samples/sec
+epoch 18, total_step 280460, total loss is 10.37 , inference loss is 3.94, weight deacy loss is 6.43, training accuracy is 0.750000, time 217.553 samples/sec
+epoch 18, total_step 280480, total loss is 13.47 , inference loss is 7.04, weight deacy loss is 6.43, training accuracy is 0.562500, time 226.744 samples/sec
+epoch 18, total_step 280500, total loss is 11.16 , inference loss is 4.73, weight deacy loss is 6.43, training accuracy is 0.562500, time 12.004 samples/sec
+epoch 18, total_step 280520, total loss is 13.79 , inference loss is 7.36, weight deacy loss is 6.43, training accuracy is 0.625000, time 299.443 samples/sec
+epoch 18, total_step 280540, total loss is 12.55 , inference loss is 6.12, weight deacy loss is 6.43, training accuracy is 0.531250, time 216.409 samples/sec
+epoch 18, total_step 280560, total loss is 11.19 , inference loss is 4.76, weight deacy loss is 6.43, training accuracy is 0.593750, time 222.949 samples/sec
+epoch 18, total_step 280580, total loss is 13.29 , inference loss is 6.86, weight deacy loss is 6.43, training accuracy is 0.531250, time 209.561 samples/sec
+epoch 18, total_step 280600, total loss is 10.73 , inference loss is 4.30, weight deacy loss is 6.43, training accuracy is 0.593750, time 227.807 samples/sec
+epoch 18, total_step 280620, total loss is 10.23 , inference loss is 3.80, weight deacy loss is 6.43, training accuracy is 0.750000, time 216.470 samples/sec
+epoch 18, total_step 280640, total loss is 11.99 , inference loss is 5.56, weight deacy loss is 6.43, training accuracy is 0.562500, time 212.276 samples/sec
+epoch 18, total_step 280660, total loss is 11.72 , inference loss is 5.29, weight deacy loss is 6.43, training accuracy is 0.531250, time 221.506 samples/sec
+epoch 18, total_step 280680, total loss is 8.14 , inference loss is 1.71, weight deacy loss is 6.43, training accuracy is 0.812500, time 178.707 samples/sec
+epoch 18, total_step 280700, total loss is 15.11 , inference loss is 8.68, weight deacy loss is 6.43, training accuracy is 0.500000, time 234.383 samples/sec
+epoch 18, total_step 280720, total loss is 11.17 , inference loss is 4.74, weight deacy loss is 6.43, training accuracy is 0.468750, time 228.736 samples/sec
+epoch 18, total_step 280740, total loss is 8.69 , inference loss is 2.26, weight deacy loss is 6.43, training accuracy is 0.687500, time 225.579 samples/sec
+epoch 18, total_step 280760, total loss is 12.24 , inference loss is 5.81, weight deacy loss is 6.43, training accuracy is 0.468750, time 225.061 samples/sec
+epoch 18, total_step 280780, total loss is 10.30 , inference loss is 3.87, weight deacy loss is 6.43, training accuracy is 0.593750, time 227.389 samples/sec
+epoch 18, total_step 280800, total loss is 17.88 , inference loss is 11.45, weight deacy loss is 6.43, training accuracy is 0.375000, time 13.463 samples/sec
+epoch 18, total_step 280820, total loss is 12.29 , inference loss is 5.86, weight deacy loss is 6.43, training accuracy is 0.500000, time 224.350 samples/sec
+epoch 18, total_step 280840, total loss is 11.45 , inference loss is 5.02, weight deacy loss is 6.43, training accuracy is 0.562500, time 214.537 samples/sec
+epoch 18, total_step 280860, total loss is 12.69 , inference loss is 6.25, weight deacy loss is 6.43, training accuracy is 0.531250, time 220.081 samples/sec
+epoch 18, total_step 280880, total loss is 15.70 , inference loss is 9.27, weight deacy loss is 6.43, training accuracy is 0.500000, time 218.611 samples/sec
+epoch 18, total_step 280900, total loss is 10.35 , inference loss is 3.92, weight deacy loss is 6.43, training accuracy is 0.593750, time 196.700 samples/sec
+epoch 18, total_step 280920, total loss is 17.67 , inference loss is 11.24, weight deacy loss is 6.43, training accuracy is 0.375000, time 221.946 samples/sec
+epoch 18, total_step 280940, total loss is 12.16 , inference loss is 5.73, weight deacy loss is 6.43, training accuracy is 0.562500, time 188.052 samples/sec
+epoch 18, total_step 280960, total loss is 10.30 , inference loss is 3.86, weight deacy loss is 6.43, training accuracy is 0.656250, time 237.126 samples/sec
+epoch 18, total_step 280980, total loss is 17.59 , inference loss is 11.16, weight deacy loss is 6.43, training accuracy is 0.375000, time 222.160 samples/sec
+epoch 18, total_step 281000, total loss is 12.18 , inference loss is 5.75, weight deacy loss is 6.43, training accuracy is 0.656250, time 229.583 samples/sec
+epoch 18, total_step 281020, total loss is 12.75 , inference loss is 6.32, weight deacy loss is 6.43, training accuracy is 0.593750, time 220.723 samples/sec
+epoch 18, total_step 281040, total loss is 13.90 , inference loss is 7.47, weight deacy loss is 6.43, training accuracy is 0.656250, time 199.092 samples/sec
+epoch 18, total_step 281060, total loss is 12.36 , inference loss is 5.93, weight deacy loss is 6.43, training accuracy is 0.531250, time 218.818 samples/sec
+epoch 18, total_step 281080, total loss is 11.65 , inference loss is 5.21, weight deacy loss is 6.43, training accuracy is 0.625000, time 224.185 samples/sec
+epoch 18, total_step 281100, total loss is 9.72 , inference loss is 3.29, weight deacy loss is 6.43, training accuracy is 0.718750, time 12.238 samples/sec
+epoch 18, total_step 281120, total loss is 12.25 , inference loss is 5.82, weight deacy loss is 6.43, training accuracy is 0.593750, time 223.022 samples/sec
+epoch 18, total_step 281140, total loss is 11.26 , inference loss is 4.83, weight deacy loss is 6.43, training accuracy is 0.656250, time 218.994 samples/sec
+epoch 18, total_step 281160, total loss is 12.60 , inference loss is 6.17, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.579 samples/sec
+epoch 18, total_step 281180, total loss is 13.71 , inference loss is 7.28, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.797 samples/sec
+epoch 18, total_step 281200, total loss is 12.85 , inference loss is 6.42, weight deacy loss is 6.43, training accuracy is 0.437500, time 226.003 samples/sec
+epoch 18, total_step 281220, total loss is 13.26 , inference loss is 6.83, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.240 samples/sec
+epoch 18, total_step 281240, total loss is 15.65 , inference loss is 9.22, weight deacy loss is 6.43, training accuracy is 0.500000, time 188.819 samples/sec
+epoch 18, total_step 281260, total loss is 11.79 , inference loss is 5.36, weight deacy loss is 6.43, training accuracy is 0.781250, time 226.474 samples/sec
+epoch 18, total_step 281280, total loss is 11.57 , inference loss is 5.14, weight deacy loss is 6.43, training accuracy is 0.593750, time 226.991 samples/sec
+epoch 18, total_step 281300, total loss is 10.67 , inference loss is 4.23, weight deacy loss is 6.43, training accuracy is 0.781250, time 230.864 samples/sec
+epoch 18, total_step 281320, total loss is 9.67 , inference loss is 3.24, weight deacy loss is 6.43, training accuracy is 0.750000, time 230.438 samples/sec
+epoch 18, total_step 281340, total loss is 12.91 , inference loss is 6.48, weight deacy loss is 6.43, training accuracy is 0.406250, time 225.395 samples/sec
+epoch 18, total_step 281360, total loss is 12.40 , inference loss is 5.97, weight deacy loss is 6.43, training accuracy is 0.656250, time 300.014 samples/sec
+epoch 18, total_step 281380, total loss is 10.77 , inference loss is 4.33, weight deacy loss is 6.43, training accuracy is 0.687500, time 213.104 samples/sec
+epoch 18, total_step 281400, total loss is 10.36 , inference loss is 3.93, weight deacy loss is 6.43, training accuracy is 0.656250, time 11.932 samples/sec
+epoch 18, total_step 281420, total loss is 13.75 , inference loss is 7.32, weight deacy loss is 6.43, training accuracy is 0.500000, time 229.247 samples/sec
+epoch 18, total_step 281440, total loss is 14.11 , inference loss is 7.68, weight deacy loss is 6.43, training accuracy is 0.531250, time 224.591 samples/sec
+epoch 18, total_step 281460, total loss is 10.59 , inference loss is 4.15, weight deacy loss is 6.43, training accuracy is 0.687500, time 220.719 samples/sec
+epoch 18, total_step 281480, total loss is 13.36 , inference loss is 6.93, weight deacy loss is 6.43, training accuracy is 0.531250, time 225.925 samples/sec
+epoch 18, total_step 281500, total loss is 14.53 , inference loss is 8.10, weight deacy loss is 6.43, training accuracy is 0.437500, time 204.430 samples/sec
+epoch 18, total_step 281520, total loss is 11.95 , inference loss is 5.52, weight deacy loss is 6.43, training accuracy is 0.656250, time 215.171 samples/sec
+epoch 18, total_step 281540, total loss is 11.15 , inference loss is 4.72, weight deacy loss is 6.43, training accuracy is 0.531250, time 222.250 samples/sec
+epoch 18, total_step 281560, total loss is 16.00 , inference loss is 9.56, weight deacy loss is 6.43, training accuracy is 0.531250, time 228.558 samples/sec
+epoch 18, total_step 281580, total loss is 10.53 , inference loss is 4.10, weight deacy loss is 6.43, training accuracy is 0.687500, time 224.833 samples/sec
+epoch 18, total_step 281600, total loss is 12.13 , inference loss is 5.70, weight deacy loss is 6.43, training accuracy is 0.468750, time 297.181 samples/sec
+epoch 18, total_step 281620, total loss is 9.28 , inference loss is 2.85, weight deacy loss is 6.43, training accuracy is 0.718750, time 211.643 samples/sec
+epoch 18, total_step 281640, total loss is 12.64 , inference loss is 6.20, weight deacy loss is 6.43, training accuracy is 0.500000, time 210.449 samples/sec
+epoch 18, total_step 281660, total loss is 12.98 , inference loss is 6.55, weight deacy loss is 6.43, training accuracy is 0.593750, time 214.074 samples/sec
+epoch 18, total_step 281680, total loss is 17.98 , inference loss is 11.55, weight deacy loss is 6.43, training accuracy is 0.437500, time 211.280 samples/sec
+epoch 18, total_step 281700, total loss is 9.28 , inference loss is 2.85, weight deacy loss is 6.43, training accuracy is 0.625000, time 13.397 samples/sec
+epoch 18, total_step 281720, total loss is 10.35 , inference loss is 3.92, weight deacy loss is 6.43, training accuracy is 0.625000, time 298.308 samples/sec
+epoch 18, total_step 281740, total loss is 11.95 , inference loss is 5.52, weight deacy loss is 6.43, training accuracy is 0.531250, time 298.487 samples/sec
+epoch 18, total_step 281760, total loss is 12.82 , inference loss is 6.39, weight deacy loss is 6.43, training accuracy is 0.500000, time 185.816 samples/sec
+epoch 18, total_step 281780, total loss is 13.38 , inference loss is 6.95, weight deacy loss is 6.43, training accuracy is 0.437500, time 222.175 samples/sec
+epoch 18, total_step 281800, total loss is 9.65 , inference loss is 3.21, weight deacy loss is 6.43, training accuracy is 0.593750, time 210.785 samples/sec
+epoch 18, total_step 281820, total loss is 15.26 , inference loss is 8.83, weight deacy loss is 6.43, training accuracy is 0.468750, time 178.936 samples/sec
+epoch 18, total_step 281840, total loss is 12.69 , inference loss is 6.25, weight deacy loss is 6.43, training accuracy is 0.468750, time 219.543 samples/sec
+epoch 18, total_step 281860, total loss is 13.15 , inference loss is 6.72, weight deacy loss is 6.43, training accuracy is 0.437500, time 199.934 samples/sec
+epoch 18, total_step 281880, total loss is 12.25 , inference loss is 5.82, weight deacy loss is 6.43, training accuracy is 0.625000, time 200.051 samples/sec
+epoch 18, total_step 281900, total loss is 9.21 , inference loss is 2.78, weight deacy loss is 6.43, training accuracy is 0.625000, time 196.889 samples/sec
+epoch 18, total_step 281920, total loss is 10.61 , inference loss is 4.18, weight deacy loss is 6.43, training accuracy is 0.468750, time 224.988 samples/sec
+epoch 18, total_step 281940, total loss is 14.07 , inference loss is 7.64, weight deacy loss is 6.43, training accuracy is 0.531250, time 210.284 samples/sec
+epoch 18, total_step 281960, total loss is 12.38 , inference loss is 5.95, weight deacy loss is 6.43, training accuracy is 0.625000, time 188.689 samples/sec
+epoch 18, total_step 281980, total loss is 15.42 , inference loss is 8.99, weight deacy loss is 6.43, training accuracy is 0.468750, time 193.606 samples/sec
+testing verification..
+(12000, 512)
+infer time 25.439507999999968
+best_threshold_index 117 0.9885185185185185
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 117 0.9881481481481481
+best_threshold_index 117 0.9883333333333333
+best_threshold_index 117 0.9885185185185185
+best_threshold_index 117 0.9872222222222222
+best_threshold_index 117 0.9887037037037038
+best_threshold_index 117 0.9875925925925926
+best_threshold_index 121 0.9868518518518519
+best_threshold_index 117 0.987037037037037
+[lfw][282000]XNorm: 19.891002
+[lfw][282000]Accuracy-Flip: 0.98750+-0.00554
+test accuracy is: 0.9874999999999998
+epoch 18, total_step 282000, total loss is 11.86 , inference loss is 5.43, weight deacy loss is 6.43, training accuracy is 0.375000, time 17.232 samples/sec
+epoch 18, total_step 282020, total loss is 13.80 , inference loss is 7.37, weight deacy loss is 6.43, training accuracy is 0.625000, time 206.684 samples/sec
+epoch 18, total_step 282040, total loss is 15.44 , inference loss is 9.01, weight deacy loss is 6.43, training accuracy is 0.437500, time 213.341 samples/sec
+epoch 18, total_step 282060, total loss is 13.30 , inference loss is 6.87, weight deacy loss is 6.43, training accuracy is 0.406250, time 228.335 samples/sec
+epoch 18, total_step 282080, total loss is 18.38 , inference loss is 11.95, weight deacy loss is 6.43, training accuracy is 0.375000, time 300.329 samples/sec
+epoch 18, total_step 282100, total loss is 11.33 , inference loss is 4.90, weight deacy loss is 6.43, training accuracy is 0.500000, time 196.762 samples/sec
+epoch 18, total_step 282120, total loss is 13.45 , inference loss is 7.02, weight deacy loss is 6.43, training accuracy is 0.500000, time 214.145 samples/sec
+epoch 18, total_step 282140, total loss is 9.31 , inference loss is 2.88, weight deacy loss is 6.43, training accuracy is 0.687500, time 224.078 samples/sec
+epoch 18, total_step 282160, total loss is 12.90 , inference loss is 6.47, weight deacy loss is 6.43, training accuracy is 0.593750, time 232.875 samples/sec
+epoch 18, total_step 282180, total loss is 15.65 , inference loss is 9.22, weight deacy loss is 6.43, training accuracy is 0.500000, time 219.073 samples/sec
+epoch 18, total_step 282200, total loss is 13.94 , inference loss is 7.51, weight deacy loss is 6.43, training accuracy is 0.562500, time 224.264 samples/sec
+epoch 18, total_step 282220, total loss is 13.66 , inference loss is 7.23, weight deacy loss is 6.43, training accuracy is 0.468750, time 213.618 samples/sec
+epoch 18, total_step 282240, total loss is 13.22 , inference loss is 6.79, weight deacy loss is 6.43, training accuracy is 0.500000, time 210.360 samples/sec
+epoch 18, total_step 282260, total loss is 13.11 , inference loss is 6.68, weight deacy loss is 6.43, training accuracy is 0.375000, time 221.237 samples/sec
+epoch 18, total_step 282280, total loss is 17.69 , inference loss is 11.26, weight deacy loss is 6.43, training accuracy is 0.468750, time 213.532 samples/sec
+epoch 18, total_step 282300, total loss is 11.75 , inference loss is 5.32, weight deacy loss is 6.43, training accuracy is 0.562500, time 12.635 samples/sec
+epoch 18, total_step 282320, total loss is 12.72 , inference loss is 6.29, weight deacy loss is 6.43, training accuracy is 0.437500, time 299.971 samples/sec
+epoch 18, total_step 282340, total loss is 14.67 , inference loss is 8.24, weight deacy loss is 6.43, training accuracy is 0.562500, time 216.219 samples/sec
+epoch 18, total_step 282360, total loss is 10.71 , inference loss is 4.27, weight deacy loss is 6.43, training accuracy is 0.625000, time 203.647 samples/sec
+epoch 18, total_step 282380, total loss is 10.83 , inference loss is 4.40, weight deacy loss is 6.43, training accuracy is 0.593750, time 222.877 samples/sec
+epoch 18, total_step 282400, total loss is 13.66 , inference loss is 7.23, weight deacy loss is 6.43, training accuracy is 0.656250, time 189.972 samples/sec
+epoch 18, total_step 282420, total loss is 15.86 , inference loss is 9.43, weight deacy loss is 6.43, training accuracy is 0.531250, time 208.222 samples/sec
+epoch 18, total_step 282440, total loss is 14.15 , inference loss is 7.72, weight deacy loss is 6.43, training accuracy is 0.437500, time 211.466 samples/sec
+epoch 18, total_step 282460, total loss is 11.42 , inference loss is 4.99, weight deacy loss is 6.43, training accuracy is 0.687500, time 214.154 samples/sec
+epoch 18, total_step 282480, total loss is 12.32 , inference loss is 5.89, weight deacy loss is 6.43, training accuracy is 0.468750, time 207.428 samples/sec
+epoch 18, total_step 282500, total loss is 12.64 , inference loss is 6.21, weight deacy loss is 6.43, training accuracy is 0.531250, time 212.498 samples/sec
+epoch 18, total_step 282520, total loss is 16.15 , inference loss is 9.72, weight deacy loss is 6.43, training accuracy is 0.343750, time 243.321 samples/sec
+epoch 18, total_step 282540, total loss is 11.26 , inference loss is 4.83, weight deacy loss is 6.43, training accuracy is 0.656250, time 225.272 samples/sec
+epoch 18, total_step 282560, total loss is 12.84 , inference loss is 6.41, weight deacy loss is 6.43, training accuracy is 0.625000, time 221.332 samples/sec
+epoch 18, total_step 282580, total loss is 15.16 , inference loss is 8.72, weight deacy loss is 6.43, training accuracy is 0.562500, time 216.820 samples/sec
+epoch 18, total_step 282600, total loss is 10.08 , inference loss is 3.65, weight deacy loss is 6.43, training accuracy is 0.718750, time 14.469 samples/sec
+epoch 18, total_step 282620, total loss is 9.86 , inference loss is 3.43, weight deacy loss is 6.43, training accuracy is 0.781250, time 197.321 samples/sec
+epoch 18, total_step 282640, total loss is 10.49 , inference loss is 4.06, weight deacy loss is 6.43, training accuracy is 0.531250, time 209.186 samples/sec
+epoch 18, total_step 282660, total loss is 9.83 , inference loss is 3.40, weight deacy loss is 6.43, training accuracy is 0.656250, time 212.088 samples/sec
+epoch 18, total_step 282680, total loss is 12.08 , inference loss is 5.65, weight deacy loss is 6.43, training accuracy is 0.437500, time 205.331 samples/sec
+epoch 18, total_step 282700, total loss is 11.73 , inference loss is 5.30, weight deacy loss is 6.43, training accuracy is 0.656250, time 216.247 samples/sec
+epoch 18, total_step 282720, total loss is 11.08 , inference loss is 4.65, weight deacy loss is 6.43, training accuracy is 0.812500, time 225.802 samples/sec
+epoch 18, total_step 282740, total loss is 11.48 , inference loss is 5.05, weight deacy loss is 6.43, training accuracy is 0.562500, time 219.865 samples/sec
+epoch 18, total_step 282760, total loss is 9.39 , inference loss is 2.96, weight deacy loss is 6.43, training accuracy is 0.625000, time 208.236 samples/sec
+epoch 18, total_step 282780, total loss is 11.40 , inference loss is 4.97, weight deacy loss is 6.43, training accuracy is 0.562500, time 205.604 samples/sec
+epoch 18, total_step 282800, total loss is 15.07 , inference loss is 8.64, weight deacy loss is 6.43, training accuracy is 0.437500, time 219.961 samples/sec
+epoch 18, total_step 282820, total loss is 10.85 , inference loss is 4.42, weight deacy loss is 6.43, training accuracy is 0.562500, time 204.340 samples/sec
+epoch 18, total_step 282840, total loss is 15.28 , inference loss is 8.85, weight deacy loss is 6.43, training accuracy is 0.468750, time 213.428 samples/sec
+epoch 18, total_step 282860, total loss is 17.65 , inference loss is 11.22, weight deacy loss is 6.43, training accuracy is 0.406250, time 220.171 samples/sec
+epoch 18, total_step 282880, total loss is 12.93 , inference loss is 6.50, weight deacy loss is 6.43, training accuracy is 0.468750, time 195.611 samples/sec
+epoch 18, total_step 282900, total loss is 13.58 , inference loss is 7.14, weight deacy loss is 6.43, training accuracy is 0.531250, time 13.014 samples/sec
+epoch 18, total_step 282920, total loss is 15.59 , inference loss is 9.16, weight deacy loss is 6.43, training accuracy is 0.500000, time 212.224 samples/sec
+epoch 18, total_step 282940, total loss is 9.99 , inference loss is 3.56, weight deacy loss is 6.43, training accuracy is 0.625000, time 197.769 samples/sec
+epoch 18, total_step 282960, total loss is 12.19 , inference loss is 5.76, weight deacy loss is 6.43, training accuracy is 0.593750, time 185.243 samples/sec
+epoch 18, total_step 282980, total loss is 10.60 , inference loss is 4.17, weight deacy loss is 6.43, training accuracy is 0.531250, time 221.466 samples/sec
+epoch 18, total_step 283000, total loss is 9.63 , inference loss is 3.20, weight deacy loss is 6.43, training accuracy is 0.781250, time 213.360 samples/sec
+epoch 18, total_step 283020, total loss is 12.06 , inference loss is 5.63, weight deacy loss is 6.43, training accuracy is 0.437500, time 213.976 samples/sec
+epoch 18, total_step 283040, total loss is 10.98 , inference loss is 4.55, weight deacy loss is 6.43, training accuracy is 0.468750, time 220.960 samples/sec
+epoch 18, total_step 283060, total loss is 12.81 , inference loss is 6.37, weight deacy loss is 6.43, training accuracy is 0.437500, time 205.476 samples/sec
+epoch 18, total_step 283080, total loss is 10.13 , inference loss is 3.70, weight deacy loss is 6.43, training accuracy is 0.687500, time 211.144 samples/sec
+epoch 18, total_step 283100, total loss is 16.72 , inference loss is 10.29, weight deacy loss is 6.43, training accuracy is 0.468750, time 213.851 samples/sec
+epoch 18, total_step 283120, total loss is 12.87 , inference loss is 6.44, weight deacy loss is 6.43, training accuracy is 0.593750, time 209.446 samples/sec
+epoch 18, total_step 283140, total loss is 10.16 , inference loss is 3.73, weight deacy loss is 6.43, training accuracy is 0.656250, time 299.510 samples/sec
+epoch 18, total_step 283160, total loss is 11.83 , inference loss is 5.40, weight deacy loss is 6.43, training accuracy is 0.500000, time 198.054 samples/sec
+epoch 18, total_step 283180, total loss is 13.50 , inference loss is 7.06, weight deacy loss is 6.43, training accuracy is 0.406250, time 220.829 samples/sec
+epoch 18, total_step 283200, total loss is 9.83 , inference loss is 3.40, weight deacy loss is 6.43, training accuracy is 0.625000, time 12.363 samples/sec
+epoch 18, total_step 283220, total loss is 11.43 , inference loss is 5.00, weight deacy loss is 6.43, training accuracy is 0.562500, time 195.172 samples/sec
+epoch 18, total_step 283240, total loss is 17.64 , inference loss is 11.21, weight deacy loss is 6.43, training accuracy is 0.500000, time 208.737 samples/sec
+epoch 18, total_step 283260, total loss is 12.15 , inference loss is 5.72, weight deacy loss is 6.43, training accuracy is 0.468750, time 223.491 samples/sec
+epoch 18, total_step 283280, total loss is 11.45 , inference loss is 5.02, weight deacy loss is 6.43, training accuracy is 0.531250, time 244.555 samples/sec
+epoch 18, total_step 283300, total loss is 11.99 , inference loss is 5.55, weight deacy loss is 6.43, training accuracy is 0.562500, time 297.786 samples/sec
+epoch 18, total_step 283320, total loss is 13.65 , inference loss is 7.21, weight deacy loss is 6.43, training accuracy is 0.531250, time 208.145 samples/sec
+epoch 18, total_step 283340, total loss is 10.94 , inference loss is 4.51, weight deacy loss is 6.43, training accuracy is 0.625000, time 213.373 samples/sec
+epoch 18, total_step 283360, total loss is 15.60 , inference loss is 9.16, weight deacy loss is 6.43, training accuracy is 0.468750, time 227.963 samples/sec
+epoch 18, total_step 283380, total loss is 8.46 , inference loss is 2.03, weight deacy loss is 6.43, training accuracy is 0.750000, time 215.399 samples/sec
+epoch 18, total_step 283400, total loss is 12.54 , inference loss is 6.10, weight deacy loss is 6.43, training accuracy is 0.656250, time 213.432 samples/sec
+epoch 18, total_step 283420, total loss is 10.49 , inference loss is 4.05, weight deacy loss is 6.43, training accuracy is 0.562500, time 209.897 samples/sec
+epoch 18, total_step 283440, total loss is 14.86 , inference loss is 8.43, weight deacy loss is 6.43, training accuracy is 0.593750, time 295.469 samples/sec
+epoch 18, total_step 283460, total loss is 15.39 , inference loss is 8.95, weight deacy loss is 6.43, training accuracy is 0.375000, time 298.470 samples/sec
+epoch 18, total_step 283480, total loss is 9.73 , inference loss is 3.30, weight deacy loss is 6.43, training accuracy is 0.625000, time 196.967 samples/sec
+epoch 18, total_step 283500, total loss is 11.91 , inference loss is 5.48, weight deacy loss is 6.43, training accuracy is 0.562500, time 12.646 samples/sec
+epoch 18, total_step 283520, total loss is 12.59 , inference loss is 6.16, weight deacy loss is 6.43, training accuracy is 0.531250, time 300.370 samples/sec
+epoch 18, total_step 283540, total loss is 13.75 , inference loss is 7.32, weight deacy loss is 6.43, training accuracy is 0.562500, time 216.194 samples/sec
+epoch 18, total_step 283560, total loss is 10.72 , inference loss is 4.29, weight deacy loss is 6.43, training accuracy is 0.687500, time 208.486 samples/sec
+epoch 18, total_step 283580, total loss is 12.44 , inference loss is 6.00, weight deacy loss is 6.43, training accuracy is 0.656250, time 194.671 samples/sec
+epoch 18, total_step 283600, total loss is 11.34 , inference loss is 4.91, weight deacy loss is 6.43, training accuracy is 0.593750, time 220.026 samples/sec
+epoch 18, total_step 283620, total loss is 13.36 , inference loss is 6.92, weight deacy loss is 6.43, training accuracy is 0.625000, time 215.172 samples/sec
+epoch 18, total_step 283640, total loss is 13.29 , inference loss is 6.86, weight deacy loss is 6.43, training accuracy is 0.593750, time 200.931 samples/sec
+epoch 18, total_step 283660, total loss is 9.99 , inference loss is 3.56, weight deacy loss is 6.43, training accuracy is 0.625000, time 207.432 samples/sec
+epoch 18, total_step 283680, total loss is 10.85 , inference loss is 4.42, weight deacy loss is 6.43, training accuracy is 0.500000, time 210.187 samples/sec
+epoch 18, total_step 283700, total loss is 12.13 , inference loss is 5.70, weight deacy loss is 6.43, training accuracy is 0.687500, time 212.240 samples/sec
+epoch 18, total_step 283720, total loss is 12.28 , inference loss is 5.85, weight deacy loss is 6.43, training accuracy is 0.593750, time 208.336 samples/sec
+epoch 18, total_step 283740, total loss is 13.48 , inference loss is 7.05, weight deacy loss is 6.43, training accuracy is 0.468750, time 204.325 samples/sec
+epoch 18, total_step 283760, total loss is 11.10 , inference loss is 4.67, weight deacy loss is 6.43, training accuracy is 0.562500, time 210.657 samples/sec
+epoch 18, total_step 283780, total loss is 11.12 , inference loss is 4.69, weight deacy loss is 6.43, training accuracy is 0.562500, time 207.996 samples/sec
+epoch 18, total_step 283800, total loss is 8.79 , inference loss is 2.36, weight deacy loss is 6.43, training accuracy is 0.718750, time 13.702 samples/sec
+epoch 18, total_step 283820, total loss is 10.93 , inference loss is 4.50, weight deacy loss is 6.43, training accuracy is 0.593750, time 221.950 samples/sec
+epoch 18, total_step 283840, total loss is 14.77 , inference loss is 8.33, weight deacy loss is 6.43, training accuracy is 0.500000, time 223.900 samples/sec
+epoch 18, total_step 283860, total loss is 13.13 , inference loss is 6.70, weight deacy loss is 6.43, training accuracy is 0.437500, time 238.786 samples/sec
+epoch 18, total_step 283880, total loss is 14.62 , inference loss is 8.19, weight deacy loss is 6.43, training accuracy is 0.468750, time 212.653 samples/sec
+epoch 18, total_step 283900, total loss is 14.94 , inference loss is 8.51, weight deacy loss is 6.43, training accuracy is 0.500000, time 214.535 samples/sec
+epoch 18, total_step 283920, total loss is 13.15 , inference loss is 6.72, weight deacy loss is 6.43, training accuracy is 0.531250, time 217.717 samples/sec
+epoch 18, total_step 283940, total loss is 12.00 , inference loss is 5.57, weight deacy loss is 6.43, training accuracy is 0.562500, time 196.666 samples/sec
+epoch 18, total_step 283960, total loss is 10.09 , inference loss is 3.66, weight deacy loss is 6.43, training accuracy is 0.468750, time 222.803 samples/sec
+epoch 18, total_step 283980, total loss is 12.13 , inference loss is 5.70, weight deacy loss is 6.43, training accuracy is 0.656250, time 216.837 samples/sec
+testing verification..
+(12000, 512)
+infer time 24.22726800000003
+best_threshold_index 122 0.9892592592592593
+best_threshold_index 122 0.9888888888888889
+best_threshold_index 116 0.9883333333333333
+best_threshold_index 114 0.9894444444444445
+best_threshold_index 122 0.9892592592592593
+best_threshold_index 122 0.9881481481481481
+best_threshold_index 122 0.9896296296296296
+best_threshold_index 122 0.9885185185185185
+best_threshold_index 122 0.9875925925925926
+best_threshold_index 122 0.9879629629629629
+[lfw][284000]XNorm: 20.702124
+[lfw][284000]Accuracy-Flip: 0.98767+-0.00624
+test accuracy is: 0.9876666666666665
+Traceback (most recent call last):
+ File "/usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/client/session.py", line 1365, in _do_call
+ return fn(*args)
+ File "/usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/client/session.py", line 1350, in _run_fn
+ target_list, run_metadata)
+ File "/usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/client/session.py", line 1443, in _call_tf_sessionrun
+ run_metadata)
+tensorflow.python.framework.errors_impl.InternalError: GeOp11_0GEOP::::DoRunAsync Failed
+Error Message is :
+EL0004: Failed to allocate memory.
+ Possible Cause: Available memory is insufficient.
+ Solution: Close applications not in use.
+ TraceBack (most recent call last):
+ rtMalloc execute failed, reason=[driver error:out of memory][FUNC:FuncErrorReason][FILE:error_message_manage.cc][LINE:49]
+ Call rtMalloc fail, purpose:feature map,used for op input and output., size:6861537792, device_id:6[FUNC:MallocMemory][FILE:graph_mem_allocator.cc][LINE:56]
+ MallocFeatureMapMem fail, data_size:6861537792, model_id:2159, check invalid[FUNC:InitFeatureMapAndP2PMem][FILE:davinci_model.cc][LINE:373]
+
+ [[{{node GeOp11_0}}]]
+
+During handling of the above exception, another exception occurred:
+
+Traceback (most recent call last):
+ File "train_nets.py", line 179, in
+ options=config_pb2.RunOptions(report_tensor_allocations_upon_oom=True))
+ File "/usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/client/session.py", line 956, in run
+ run_metadata_ptr)
+ File "/usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/client/session.py", line 1180, in _run
+ feed_dict_tensor, options, run_metadata)
+ File "/usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/client/session.py", line 1359, in _do_run
+ run_metadata)
+ File "/usr/local/python3.7.5/lib/python3.7/site-packages/tensorflow_core/python/client/session.py", line 1384, in _do_call
+ raise type(e)(node_def, op, message)
+tensorflow.python.framework.errors_impl.InternalError: GeOp11_0GEOP::::DoRunAsync Failed
+Error Message is :
+EL0004: Failed to allocate memory.
+ Possible Cause: Available memory is insufficient.
+ Solution: Close applications not in use.
+ TraceBack (most recent call last):
+ rtMalloc execute failed, reason=[driver error:out of memory][FUNC:FuncErrorReason][FILE:error_message_manage.cc][LINE:49]
+ Call rtMalloc fail, purpose:feature map,used for op input and output., size:6861537792, device_id:6[FUNC:MallocMemory][FILE:graph_mem_allocator.cc][LINE:56]
+ MallocFeatureMapMem fail, data_size:6861537792, model_id:2159, check invalid[FUNC:InitFeatureMapAndP2PMem][FILE:davinci_model.cc][LINE:373]
+
+ [[{{node GeOp11_0}}]]
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/figures/if_v100_perfermance.log b/TensorFlow/contrib/cv/InsightFace_TF/figures/if_v100_perfermance.log
new file mode 100644
index 0000000000000000000000000000000000000000..83ac97bbbd9d90dab1660696fa095b3c324d62e8
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/figures/if_v100_perfermance.log
@@ -0,0 +1,1552 @@
+nohup: ignoring input
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:516: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint8 = np.dtype([("qint8", np.int8, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:517: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_quint8 = np.dtype([("quint8", np.uint8, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:518: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint16 = np.dtype([("qint16", np.int16, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:519: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_quint16 = np.dtype([("quint16", np.uint16, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:520: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint32 = np.dtype([("qint32", np.int32, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorflow/python/framework/dtypes.py:525: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ np_resource = np.dtype([("resource", np.ubyte, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:541: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint8 = np.dtype([("qint8", np.int8, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:542: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_quint8 = np.dtype([("quint8", np.uint8, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:543: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint16 = np.dtype([("qint16", np.int16, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:544: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_quint16 = np.dtype([("quint16", np.uint16, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:545: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ _np_qint32 = np.dtype([("qint32", np.int32, 1)])
+/root/anaconda3/envs/if/lib/python3.7/site-packages/tensorboard/compat/tensorflow_stub/dtypes.py:550: FutureWarning: Passing (type, 1) or '1type' as a synonym of type is deprecated; in a future version of numpy, it will be understood as (type, (1,)) / '(1,)type'.
+ np_resource = np.dtype([("resource", np.ubyte, 1)])
+WARNING:tensorflow:From /root/anaconda3/envs/if/lib/python3.7/site-packages/tensorlayer/layers.py:28: The name tf.GraphKeys is deprecated. Please use tf.compat.v1.GraphKeys instead.
+
+WARNING:tensorflow:From /root/anaconda3/envs/if/lib/python3.7/site-packages/tensorlayer/layers.py:4030: The name tf.nn.max_pool is deprecated. Please use tf.nn.max_pool2d instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:50: The name tf.assign_add is deprecated. Please use tf.compat.v1.assign_add instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:51: The name tf.placeholder is deprecated. Please use tf.compat.v1.placeholder instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/data/mx2tfrecords.py:63: The name tf.FixedLenFeature is deprecated. Please use tf.io.FixedLenFeature instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/data/mx2tfrecords.py:65: The name tf.parse_single_example is deprecated. Please use tf.io.parse_single_example instead.
+
+2022-12-02 19:10:44.549375: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcuda.so.1
+2022-12-02 19:10:45.107187: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:10:45.108611: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1640] Found device 0 with properties:
+name: Tesla V100-PCIE-32GB major: 7 minor: 0 memoryClockRate(GHz): 1.38
+pciBusID: 0000:00:0d.0
+2022-12-02 19:10:45.108849: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudart.so.10.0
+2022-12-02 19:10:45.109850: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcublas.so.10.0
+2022-12-02 19:10:45.110803: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcufft.so.10.0
+2022-12-02 19:10:45.111076: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcurand.so.10.0
+2022-12-02 19:10:45.112244: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusolver.so.10.0
+2022-12-02 19:10:45.113217: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusparse.so.10.0
+2022-12-02 19:10:45.116300: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
+2022-12-02 19:10:45.116417: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:10:45.117858: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:10:45.119188: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1763] Adding visible gpu devices: 0
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:64: DatasetV1.make_initializable_iterator (from tensorflow.python.data.ops.dataset_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Use `for ... in dataset:` to iterate over a dataset. If using `tf.estimator`, return the `Dataset` object directly from your input function. As a last resort, you can use `tf.compat.v1.data.make_initializable_iterator(dataset)`.
+begin db lfw convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+(12000, 112, 112, 3)
+begin db cfp_ff convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+loading bin 13000
+loading bin 14000
+(14000, 112, 112, 3)
+begin db cfp_fp convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+loading bin 13000
+loading bin 14000
+(14000, 112, 112, 3)
+begin db agedb_30 convert.
+loading bin 1000
+loading bin 2000
+loading bin 3000
+loading bin 4000
+loading bin 5000
+loading bin 6000
+loading bin 7000
+loading bin 8000
+loading bin 9000
+loading bin 10000
+loading bin 11000
+loading bin 12000
+(12000, 112, 112, 3)
+ [TL] InputLayer resnet_v1_100/input_layer: (?, 112, 112, 3)
+ [TL] Conv2dLayer resnet_v1_100/conv1: shape:[3, 3, 3, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/bn0: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/bn0/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/bn0/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/prelu0: channel_shared:False
+ [TL] PoolLayer resnet_v1_100/block1/unit_1/bottleneck_v1/shortcut: ksize:[1, 1, 1, 1] strides:[1, 2, 2, 1] padding:SAME pool:max_pool
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block1/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_1/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_2/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_3/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 64, 128] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(64,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(64,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block2/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_1/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_2/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_3/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_4/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_5/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_6/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_7/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_8/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_9/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_10/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_11/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_12/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_13/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 128, 256] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(128,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(128,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1: shape:[3, 3, 128, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block3/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_1/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_2/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_3/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_4/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_5/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_6/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_7/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_8/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_9/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_10/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_11/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_12/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_13/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_14/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_15/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_16/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_17/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_18/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_19/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_20/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_21/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_22/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_23/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_24/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_25/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_26/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_27/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_28/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_29/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_30/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 256, 512] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/Identity:0", shape=(256,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(256,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1: shape:[3, 3, 256, 512] strides:[1, 1, 1, 1] pad:SAME act:identityWARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/nets/L_Resnet_E_IR_fix_issue9.py:314: calling dropout (from tensorflow.python.ops.nn_ops) with keep_prob is deprecated and will be removed in a future version.
+Instructions for updating:
+Please use `rate` instead of `keep_prob`. Rate should be set to `rate = 1 - keep_prob`.
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/losses/face_losses.py:20: calling norm (from tensorflow.python.ops.linalg_ops) with keep_dims is deprecated and will be removed in a future version.
+Instructions for updating:
+keep_dims is deprecated, use keepdims instead
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/losses/face_losses.py:21: div (from tensorflow.python.ops.math_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Deprecated in favor of operator or tf.math.divide.
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/losses/face_losses.py:40: add_dispatch_support..wrapper (from tensorflow.python.ops.array_ops) is deprecated and will be removed in a future version.
+Instructions for updating:
+Use tf.where in 2.0, which has the same broadcast rule as np.where
+
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block4/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_1/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_2/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] PReluLayer resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_3/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/E_BN1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/E_BN1/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/E_BN1/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] ReshapeLayer resnet_v1_100/E_Reshapelayer: (?, 25088)
+ [TL] DenseLayer resnet_v1_100/E_DenseLayer: 512 identity
+ [TL] BatchNormLayer resnet_v1_100/E_BN2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+Tensor("resnet_v1_100/E_BN2/Identity:0", shape=(512,), dtype=float32)
+Tensor("resnet_v1_100/E_BN2/Identity_1:0", shape=(512,), dtype=float32)
+ [TL] InputLayer resnet_v1_100/input_layer: (?, 112, 112, 3)
+ [TL] Conv2dLayer resnet_v1_100/conv1: shape:[3, 3, 3, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/bn0: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/prelu0: channel_shared:False
+ [TL] PoolLayer resnet_v1_100/block1/unit_1/bottleneck_v1/shortcut: ksize:[1, 1, 1, 1] strides:[1, 2, 2, 1] padding:SAME pool:max_pool
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block1/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_1/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_2/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2: shape:[3, 3, 64, 64] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block1/unit_3/bottleneck_v1/combine_layer: size:(?, 56, 56, 64) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 64, 128] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1: shape:[3, 3, 64, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block2/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_1/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_2/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_3/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_4/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_5/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_6/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_7/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_8/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_9/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_10/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_11/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_12/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2: shape:[3, 3, 128, 128] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block2/unit_13/bottleneck_v1/combine_layer: size:(?, 28, 28, 128) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 128, 256] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1: shape:[3, 3, 128, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block3/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_1/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_2/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_3/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_4/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_5/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_6/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_7/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_8/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_9/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_10/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_11/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_12/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_13/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_14/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_15/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_16/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_17/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_18/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_19/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_20/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_21/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_22/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_23/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_24/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_25/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_26/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_27/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_28/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_29/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2: shape:[3, 3, 256, 256] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block3/unit_30/bottleneck_v1/combine_layer: size:(?, 14, 14, 256) fn:add
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv: shape:[1, 1, 256, 512] strides:[1, 2, 2, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1: shape:[3, 3, 256, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] PadLayer resnet_v1_100/block4/unit_1/bottleneck_v1/padding_conv2: paddings:[[0, 0], [1, 1], [1, 1], [0, 0]] mode:CONSTANT
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 2, 2, 1] pad:VALID act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_1/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_2/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] PReluLayer resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer: channel_shared:False
+ [TL] Conv2dLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2: shape:[3, 3, 512, 512] strides:[1, 1, 1, 1] pad:SAME act:identity
+ [TL] BatchNormLayer resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ElementwiseLayer resnet_v1_100/block4/unit_3/bottleneck_v1/combine_layer: size:(?, 7, 7, 512) fn:add
+ [TL] BatchNormLayer resnet_v1_100/E_BN1: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [TL] ReshapeLayer resnet_v1_100/E_Reshapelayer: (?, 25088)
+ [TL] DenseLayer resnet_v1_100/E_DenseLayer: 512 identity
+ [TL] BatchNormLayer resnet_v1_100/E_BN2: decay:0.900000 epsilon:0.000020 act:identity is_train:True
+ [*] geting variables with W_conv2d
+ got 0: resnet_v1_100/conv1/W_conv2d:0 (3, 3, 3, 64)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 2: resnet_v1_100/block1/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 3: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 4: resnet_v1_100/block1/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 5: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 64)
+ got 6: resnet_v1_100/block1/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 64, 64)
+ got 7: resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 64, 128)
+ got 8: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 64, 128)
+ got 9: resnet_v1_100/block2/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 10: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 11: resnet_v1_100/block2/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 12: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 13: resnet_v1_100/block2/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 14: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 15: resnet_v1_100/block2/unit_4/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 16: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 17: resnet_v1_100/block2/unit_5/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 18: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 19: resnet_v1_100/block2/unit_6/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 20: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 21: resnet_v1_100/block2/unit_7/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 22: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 23: resnet_v1_100/block2/unit_8/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 24: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 25: resnet_v1_100/block2/unit_9/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 26: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 27: resnet_v1_100/block2/unit_10/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 28: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 29: resnet_v1_100/block2/unit_11/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 30: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 31: resnet_v1_100/block2/unit_12/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 32: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 128)
+ got 33: resnet_v1_100/block2/unit_13/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 128, 128)
+ got 34: resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 128, 256)
+ got 35: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 128, 256)
+ got 36: resnet_v1_100/block3/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 37: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 38: resnet_v1_100/block3/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 39: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 40: resnet_v1_100/block3/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 41: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 42: resnet_v1_100/block3/unit_4/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 43: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 44: resnet_v1_100/block3/unit_5/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 45: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 46: resnet_v1_100/block3/unit_6/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 47: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 48: resnet_v1_100/block3/unit_7/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 49: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 50: resnet_v1_100/block3/unit_8/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 51: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 52: resnet_v1_100/block3/unit_9/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 53: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 54: resnet_v1_100/block3/unit_10/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 55: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 56: resnet_v1_100/block3/unit_11/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 57: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 58: resnet_v1_100/block3/unit_12/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 59: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 60: resnet_v1_100/block3/unit_13/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 61: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 62: resnet_v1_100/block3/unit_14/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 63: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 64: resnet_v1_100/block3/unit_15/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 65: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 66: resnet_v1_100/block3/unit_16/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 67: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 68: resnet_v1_100/block3/unit_17/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 69: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 70: resnet_v1_100/block3/unit_18/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 71: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 72: resnet_v1_100/block3/unit_19/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 73: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 74: resnet_v1_100/block3/unit_20/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 75: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 76: resnet_v1_100/block3/unit_21/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 77: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 78: resnet_v1_100/block3/unit_22/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 79: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 80: resnet_v1_100/block3/unit_23/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 81: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 82: resnet_v1_100/block3/unit_24/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 83: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 84: resnet_v1_100/block3/unit_25/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 85: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 86: resnet_v1_100/block3/unit_26/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 87: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 88: resnet_v1_100/block3/unit_27/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 89: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 90: resnet_v1_100/block3/unit_28/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 91: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 92: resnet_v1_100/block3/unit_29/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 93: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 256)
+ got 94: resnet_v1_100/block3/unit_30/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 256, 256)
+ got 95: resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_conv/W_conv2d:0 (1, 1, 256, 512)
+ got 96: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 256, 512)
+ got 97: resnet_v1_100/block4/unit_1/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ got 98: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 512, 512)
+ got 99: resnet_v1_100/block4/unit_2/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ got 100: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1/W_conv2d:0 (3, 3, 512, 512)
+ got 101: resnet_v1_100/block4/unit_3/bottleneck_v1/conv2/W_conv2d:0 (3, 3, 512, 512)
+ [*] geting variables with resnet_v1_50/E_DenseLayer/W
+ [*] geting variables with embedding_weights
+ got 0: arcface_loss/embedding_weights:0 (512, 85164)
+ [*] geting variables with gamma
+ got 0: resnet_v1_100/bn0/gamma:0 (64,)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 2: resnet_v1_100/block1/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 3: resnet_v1_100/block1/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 4: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 5: resnet_v1_100/block1/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 6: resnet_v1_100/block1/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 7: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 8: resnet_v1_100/block1/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (64,)
+ got 9: resnet_v1_100/block1/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (64,)
+ got 10: resnet_v1_100/block2/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (128,)
+ got 11: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (64,)
+ got 12: resnet_v1_100/block2/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 13: resnet_v1_100/block2/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 14: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 15: resnet_v1_100/block2/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 16: resnet_v1_100/block2/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 17: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 18: resnet_v1_100/block2/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 19: resnet_v1_100/block2/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 20: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 21: resnet_v1_100/block2/unit_4/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 22: resnet_v1_100/block2/unit_4/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 23: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 24: resnet_v1_100/block2/unit_5/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 25: resnet_v1_100/block2/unit_5/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 26: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 27: resnet_v1_100/block2/unit_6/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 28: resnet_v1_100/block2/unit_6/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 29: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 30: resnet_v1_100/block2/unit_7/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 31: resnet_v1_100/block2/unit_7/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 32: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 33: resnet_v1_100/block2/unit_8/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 34: resnet_v1_100/block2/unit_8/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 35: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 36: resnet_v1_100/block2/unit_9/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 37: resnet_v1_100/block2/unit_9/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 38: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 39: resnet_v1_100/block2/unit_10/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 40: resnet_v1_100/block2/unit_10/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 41: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 42: resnet_v1_100/block2/unit_11/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 43: resnet_v1_100/block2/unit_11/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 44: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 45: resnet_v1_100/block2/unit_12/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 46: resnet_v1_100/block2/unit_12/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 47: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 48: resnet_v1_100/block2/unit_13/bottleneck_v1/conv1_bn2/gamma:0 (128,)
+ got 49: resnet_v1_100/block2/unit_13/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (128,)
+ got 50: resnet_v1_100/block3/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (256,)
+ got 51: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (128,)
+ got 52: resnet_v1_100/block3/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 53: resnet_v1_100/block3/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 54: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 55: resnet_v1_100/block3/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 56: resnet_v1_100/block3/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 57: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 58: resnet_v1_100/block3/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 59: resnet_v1_100/block3/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 60: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 61: resnet_v1_100/block3/unit_4/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 62: resnet_v1_100/block3/unit_4/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 63: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 64: resnet_v1_100/block3/unit_5/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 65: resnet_v1_100/block3/unit_5/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 66: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 67: resnet_v1_100/block3/unit_6/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 68: resnet_v1_100/block3/unit_6/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 69: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 70: resnet_v1_100/block3/unit_7/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 71: resnet_v1_100/block3/unit_7/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 72: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 73: resnet_v1_100/block3/unit_8/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 74: resnet_v1_100/block3/unit_8/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 75: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 76: resnet_v1_100/block3/unit_9/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 77: resnet_v1_100/block3/unit_9/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 78: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 79: resnet_v1_100/block3/unit_10/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 80: resnet_v1_100/block3/unit_10/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 81: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 82: resnet_v1_100/block3/unit_11/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 83: resnet_v1_100/block3/unit_11/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 84: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 85: resnet_v1_100/block3/unit_12/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 86: resnet_v1_100/block3/unit_12/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 87: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 88: resnet_v1_100/block3/unit_13/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 89: resnet_v1_100/block3/unit_13/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 90: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 91: resnet_v1_100/block3/unit_14/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 92: resnet_v1_100/block3/unit_14/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 93: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 94: resnet_v1_100/block3/unit_15/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 95: resnet_v1_100/block3/unit_15/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 96: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 97: resnet_v1_100/block3/unit_16/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 98: resnet_v1_100/block3/unit_16/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 99: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 100: resnet_v1_100/block3/unit_17/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 101: resnet_v1_100/block3/unit_17/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 102: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 103: resnet_v1_100/block3/unit_18/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 104: resnet_v1_100/block3/unit_18/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 105: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 106: resnet_v1_100/block3/unit_19/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 107: resnet_v1_100/block3/unit_19/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 108: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 109: resnet_v1_100/block3/unit_20/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 110: resnet_v1_100/block3/unit_20/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 111: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 112: resnet_v1_100/block3/unit_21/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 113: resnet_v1_100/block3/unit_21/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 114: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 115: resnet_v1_100/block3/unit_22/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 116: resnet_v1_100/block3/unit_22/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 117: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 118: resnet_v1_100/block3/unit_23/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 119: resnet_v1_100/block3/unit_23/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 120: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 121: resnet_v1_100/block3/unit_24/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 122: resnet_v1_100/block3/unit_24/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 123: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 124: resnet_v1_100/block3/unit_25/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 125: resnet_v1_100/block3/unit_25/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 126: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 127: resnet_v1_100/block3/unit_26/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 128: resnet_v1_100/block3/unit_26/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 129: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 130: resnet_v1_100/block3/unit_27/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 131: resnet_v1_100/block3/unit_27/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 132: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 133: resnet_v1_100/block3/unit_28/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 134: resnet_v1_100/block3/unit_28/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 135: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 136: resnet_v1_100/block3/unit_29/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 137: resnet_v1_100/block3/unit_29/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 138: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 139: resnet_v1_100/block3/unit_30/bottleneck_v1/conv1_bn2/gamma:0 (256,)
+ got 140: resnet_v1_100/block3/unit_30/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (256,)
+ got 141: resnet_v1_100/block4/unit_1/bottleneck_v1/shortcut_bn/BatchNorm/gamma:0 (512,)
+ got 142: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn1/gamma:0 (256,)
+ got 143: resnet_v1_100/block4/unit_1/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 144: resnet_v1_100/block4/unit_1/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 145: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn1/gamma:0 (512,)
+ got 146: resnet_v1_100/block4/unit_2/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 147: resnet_v1_100/block4/unit_2/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 148: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn1/gamma:0 (512,)
+ got 149: resnet_v1_100/block4/unit_3/bottleneck_v1/conv1_bn2/gamma:0 (512,)
+ got 150: resnet_v1_100/block4/unit_3/bottleneck_v1/conv2_bn/BatchNorm/gamma:0 (512,)
+ got 151: resnet_v1_100/E_BN1/gamma:0 (512,)
+ [*] geting variables with alphas
+ got 0: resnet_v1_100/prelu0/alphas:0 (64,)
+ got 1: resnet_v1_100/block1/unit_1/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 2: resnet_v1_100/block1/unit_2/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 3: resnet_v1_100/block1/unit_3/bottleneck_v1/prelu_layer/alphas:0 (64,)
+ got 4: resnet_v1_100/block2/unit_1/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 5: resnet_v1_100/block2/unit_2/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 6: resnet_v1_100/block2/unit_3/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 7: resnet_v1_100/block2/unit_4/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 8: resnet_v1_100/block2/unit_5/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 9: resnet_v1_100/block2/unit_6/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 10: resnet_v1_100/block2/unit_7/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 11: resnet_v1_100/block2/unit_8/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 12: resnet_v1_100/block2/unit_9/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 13: resnet_v1_100/block2/unit_10/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 14: resnet_v1_100/block2/unit_11/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 15: resnet_v1_100/block2/unit_12/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 16: resnet_v1_100/block2/unit_13/bottleneck_v1/prelu_layer/alphas:0 (128,)
+ got 17: resnet_v1_100/block3/unit_1/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 18: resnet_v1_100/block3/unit_2/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 19: resnet_v1_100/block3/unit_3/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 20: resnet_v1_100/block3/unit_4/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 21: resnet_v1_100/block3/unit_5/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 22: resnet_v1_100/block3/unit_6/bottleneck_v1/prelu_layer/alphas:0 (256,)WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:112: The name tf.train.piecewise_constant is deprecated. Please use tf.compat.v1.train.piecewise_constant instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:114: The name tf.train.MomentumOptimizer is deprecated. Please use tf.compat.v1.train.MomentumOptimizer instead.
+
+2022-12-02 19:11:50.971551: I tensorflow/core/platform/cpu_feature_guard.cc:142] Your CPU supports instructions that this TensorFlow binary was not compiled to use: AVX2 AVX512F FMA
+2022-12-02 19:11:50.978062: I tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency: 2600000000 Hz
+2022-12-02 19:11:50.978391: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0xf905610 executing computations on platform Host. Devices:
+2022-12-02 19:11:50.978417: I tensorflow/compiler/xla/service/service.cc:175] StreamExecutor device (0): ,
+2022-12-02 19:11:51.079465: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:11:51.080958: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0xf92d0d0 executing computations on platform CUDA. Devices:
+2022-12-02 19:11:51.080988: I tensorflow/compiler/xla/service/service.cc:175] StreamExecutor device (0): Tesla V100-PCIE-32GB, Compute Capability 7.0
+2022-12-02 19:11:51.081191: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:11:51.082576: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1640] Found device 0 with properties:
+name: Tesla V100-PCIE-32GB major: 7 minor: 0 memoryClockRate(GHz): 1.38
+pciBusID: 0000:00:0d.0
+2022-12-02 19:11:51.082664: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudart.so.10.0
+2022-12-02 19:11:51.082690: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcublas.so.10.0
+2022-12-02 19:11:51.082714: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcufft.so.10.0
+2022-12-02 19:11:51.082735: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcurand.so.10.0
+2022-12-02 19:11:51.082765: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusolver.so.10.0
+2022-12-02 19:11:51.082786: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcusparse.so.10.0
+2022-12-02 19:11:51.082808: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
+2022-12-02 19:11:51.082901: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:11:51.084306: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:11:51.085660: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1763] Adding visible gpu devices: 0
+2022-12-02 19:11:51.085739: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudart.so.10.0
+2022-12-02 19:11:51.088069: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1181] Device interconnect StreamExecutor with strength 1 edge matrix:
+2022-12-02 19:11:51.088095: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1187] 0
+2022-12-02 19:11:51.088104: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1200] 0: N
+2022-12-02 19:11:51.088216: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:11:51.089645: I tensorflow/stream_executor/cuda/cuda_gpu_executor.cc:1005] successful NUMA node read from SysFS had negative value (-1), but there must be at least one NUMA node, so returning NUMA node zero
+2022-12-02 19:11:51.091029: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1326] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 30555 MB memory) -> physical GPU (device: 0, name: Tesla V100-PCIE-32GB, pci bus id: 0000:00:0d.0, compute capability: 7.0)
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:130: The name tf.summary.FileWriter is deprecated. Please use tf.compat.v1.summary.FileWriter instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:135: The name tf.summary.histogram is deprecated. Please use tf.compat.v1.summary.histogram instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:140: The name tf.summary.scalar is deprecated. Please use tf.compat.v1.summary.scalar instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:145: The name tf.summary.merge is deprecated. Please use tf.compat.v1.summary.merge instead.
+
+WARNING:tensorflow:From /root/dave/InsightFace_TF/InsightFace_TF_GPU_2/train_nets.py:147: The name tf.train.Saver is deprecated. Please use tf.compat.v1.train.Saver instead.
+
+2022-12-02 19:12:00.632160: W tensorflow/compiler/jit/mark_for_compilation_pass.cc:1412] (One-time warning): Not using XLA:CPU for cluster because envvar TF_XLA_FLAGS=--tf_xla_cpu_global_jit was not set. If you want XLA:CPU, either set that envvar, or use experimental_jit_scope to enable XLA:CPU. To confirm that XLA is active, pass --vmodule=xla_compilation_cache=1 (as a proper command-line flag, not via TF_XLA_FLAGS) or set the envvar XLA_FLAGS=--xla_hlo_profile.
+2022-12-02 19:12:37.229855: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcublas.so.10.0
+2022-12-02 19:12:37.500333: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
+
+ got 23: resnet_v1_100/block3/unit_7/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 24: resnet_v1_100/block3/unit_8/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 25: resnet_v1_100/block3/unit_9/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 26: resnet_v1_100/block3/unit_10/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 27: resnet_v1_100/block3/unit_11/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 28: resnet_v1_100/block3/unit_12/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 29: resnet_v1_100/block3/unit_13/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 30: resnet_v1_100/block3/unit_14/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 31: resnet_v1_100/block3/unit_15/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 32: resnet_v1_100/block3/unit_16/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 33: resnet_v1_100/block3/unit_17/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 34: resnet_v1_100/block3/unit_18/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 35: resnet_v1_100/block3/unit_19/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 36: resnet_v1_100/block3/unit_20/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 37: resnet_v1_100/block3/unit_21/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 38: resnet_v1_100/block3/unit_22/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 39: resnet_v1_100/block3/unit_23/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 40: resnet_v1_100/block3/unit_24/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 41: resnet_v1_100/block3/unit_25/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 42: resnet_v1_100/block3/unit_26/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 43: resnet_v1_100/block3/unit_27/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 44: resnet_v1_100/block3/unit_28/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 45: resnet_v1_100/block3/unit_29/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 46: resnet_v1_100/block3/unit_30/bottleneck_v1/prelu_layer/alphas:0 (256,)
+ got 47: resnet_v1_100/block4/unit_1/bottleneck_v1/prelu_layer/alphas:0 (512,)
+ got 48: resnet_v1_100/block4/unit_2/bottleneck_v1/prelu_layer/alphas:0 (512,)
+ got 49: resnet_v1_100/block4/unit_3/bottleneck_v1/prelu_layer/alphas:0 (512,)
+[640000, 960000, 1280000]
+epoch 0, total_step 20, total loss is 58.93 , inference loss is 44.46, weight deacy loss is 14.47, training accuracy is 0.000000, time 80.302 samples/sec
+epoch 0, total_step 40, total loss is 57.10 , inference loss is 42.62, weight deacy loss is 14.47, training accuracy is 0.000000, time 81.657 samples/sec
+epoch 0, total_step 60, total loss is 55.01 , inference loss is 40.53, weight deacy loss is 14.47, training accuracy is 0.000000, time 78.763 samples/sec
+epoch 0, total_step 80, total loss is 51.63 , inference loss is 37.16, weight deacy loss is 14.47, training accuracy is 0.000000, time 80.031 samples/sec
+epoch 0, total_step 100, total loss is 52.69 , inference loss is 38.22, weight deacy loss is 14.47, training accuracy is 0.000000, time 80.972 samples/sec
+epoch 0, total_step 120, total loss is 50.58 , inference loss is 36.10, weight deacy loss is 14.47, training accuracy is 0.000000, time 80.001 samples/sec
+epoch 0, total_step 140, total loss is 52.84 , inference loss is 38.37, weight deacy loss is 14.47, training accuracy is 0.000000, time 81.194 samples/sec
+epoch 0, total_step 160, total loss is 52.77 , inference loss is 38.31, weight deacy loss is 14.47, training accuracy is 0.000000, time 80.557 samples/sec
+epoch 0, total_step 180, total loss is 52.79 , inference loss is 38.32, weight deacy loss is 14.47, training accuracy is 0.000000, time 79.738 samples/sec
+epoch 0, total_step 200, total loss is 52.04 , inference loss is 37.57, weight deacy loss is 14.47, training accuracy is 0.000000, time 81.152 samples/sec
+epoch 0, total_step 220, total loss is 53.48 , inference loss is 39.01, weight deacy loss is 14.46, training accuracy is 0.000000, time 79.425 samples/sec
+epoch 0, total_step 240, total loss is 51.62 , inference loss is 37.16, weight deacy loss is 14.46, training accuracy is 0.000000, time 81.547 samples/sec
+epoch 0, total_step 260, total loss is 46.71 , inference loss is 32.25, weight deacy loss is 14.46, training accuracy is 0.000000, time 78.698 samples/sec
+epoch 0, total_step 280, total loss is 49.36 , inference loss is 34.90, weight deacy loss is 14.46, training accuracy is 0.000000, time 79.106 samples/sec
+epoch 0, total_step 300, total loss is 52.60 , inference loss is 38.14, weight deacy loss is 14.46, training accuracy is 0.000000, time 78.632 samples/sec
+epoch 0, total_step 320, total loss is 49.78 , inference loss is 35.32, weight deacy loss is 14.46, training accuracy is 0.000000, time 80.554 samples/sec
+epoch 0, total_step 340, total loss is 51.65 , inference loss is 37.20, weight deacy loss is 14.46, training accuracy is 0.031250, time 81.581 samples/sec
+epoch 0, total_step 360, total loss is 51.36 , inference loss is 36.91, weight deacy loss is 14.45, training accuracy is 0.000000, time 79.464 samples/sec
+epoch 0, total_step 380, total loss is 49.51 , inference loss is 35.06, weight deacy loss is 14.45, training accuracy is 0.000000, time 81.285 samples/sec
+epoch 0, total_step 400, total loss is 50.16 , inference loss is 35.71, weight deacy loss is 14.45, training accuracy is 0.000000, time 80.461 samples/sec
+epoch 0, total_step 420, total loss is 46.60 , inference loss is 32.15, weight deacy loss is 14.45, training accuracy is 0.000000, time 79.361 samples/sec
+epoch 0, total_step 440, total loss is 47.09 , inference loss is 32.64, weight deacy loss is 14.45, training accuracy is 0.000000, time 81.191 samples/sec
+epoch 0, total_step 460, total loss is 44.51 , inference loss is 30.07, weight deacy loss is 14.45, training accuracy is 0.031250, time 81.208 samples/sec
+epoch 0, total_step 480, total loss is 47.41 , inference loss is 32.97, weight deacy loss is 14.44, training accuracy is 0.000000, time 81.569 samples/sec
+epoch 0, total_step 500, total loss is 46.68 , inference loss is 32.24, weight deacy loss is 14.44, training accuracy is 0.000000, time 80.453 samples/sec
+epoch 0, total_step 520, total loss is 47.01 , inference loss is 32.57, weight deacy loss is 14.44, training accuracy is 0.000000, time 79.177 samples/sec
+epoch 0, total_step 540, total loss is 48.83 , inference loss is 34.40, weight deacy loss is 14.44, training accuracy is 0.000000, time 81.401 samples/sec
+epoch 0, total_step 560, total loss is 47.72 , inference loss is 33.28, weight deacy loss is 14.44, training accuracy is 0.000000, time 79.275 samples/sec
+epoch 0, total_step 580, total loss is 51.74 , inference loss is 37.30, weight deacy loss is 14.44, training accuracy is 0.000000, time 80.160 samples/sec
+epoch 0, total_step 600, total loss is 47.55 , inference loss is 33.11, weight deacy loss is 14.43, training accuracy is 0.000000, time 81.104 samples/sec
+epoch 0, total_step 620, total loss is 48.41 , inference loss is 33.98, weight deacy loss is 14.43, training accuracy is 0.000000, time 78.750 samples/sec
+epoch 0, total_step 640, total loss is 48.72 , inference loss is 34.29, weight deacy loss is 14.43, training accuracy is 0.031250, time 78.792 samples/sec
+epoch 0, total_step 660, total loss is 48.29 , inference loss is 33.86, weight deacy loss is 14.43, training accuracy is 0.031250, time 80.242 samples/sec
+epoch 0, total_step 680, total loss is 43.06 , inference loss is 28.63, weight deacy loss is 14.43, training accuracy is 0.031250, time 80.291 samples/sec
+epoch 0, total_step 700, total loss is 47.08 , inference loss is 32.65, weight deacy loss is 14.43, training accuracy is 0.031250, time 80.610 samples/sec
+epoch 0, total_step 720, total loss is 47.33 , inference loss is 32.90, weight deacy loss is 14.42, training accuracy is 0.000000, time 79.080 samples/sec
+epoch 0, total_step 740, total loss is 48.12 , inference loss is 33.70, weight deacy loss is 14.42, training accuracy is 0.000000, time 80.171 samples/sec
+epoch 0, total_step 760, total loss is 48.31 , inference loss is 33.89, weight deacy loss is 14.42, training accuracy is 0.000000, time 79.259 samples/sec
+epoch 0, total_step 780, total loss is 46.71 , inference loss is 32.30, weight deacy loss is 14.42, training accuracy is 0.000000, time 81.302 samples/sec
\ No newline at end of file
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_250k.png b/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_250k.png
new file mode 100644
index 0000000000000000000000000000000000000000..1ce134aa29ccc63736c4b5f16f03cca5a91a072e
Binary files /dev/null and b/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_250k.png differ
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_310k.png b/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_310k.png
new file mode 100644
index 0000000000000000000000000000000000000000..358732117cdb741ca8248b5bd7a641099a7362e8
Binary files /dev/null and b/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_310k.png differ
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_730k.png b/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_730k.png
new file mode 100644
index 0000000000000000000000000000000000000000..a7205690deb105b1a719be75c86bef94aead40d6
Binary files /dev/null and b/TensorFlow/contrib/cv/InsightFace_TF/figures/lfw_730k.png differ
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/figures/model_a_170k.png b/TensorFlow/contrib/cv/InsightFace_TF/figures/model_a_170k.png
new file mode 100644
index 0000000000000000000000000000000000000000..403f6eefcb84e3ef80f2262e4ae91bac8003a29a
Binary files /dev/null and b/TensorFlow/contrib/cv/InsightFace_TF/figures/model_a_170k.png differ
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/losses/__init__.py b/TensorFlow/contrib/cv/InsightFace_TF/losses/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/losses/face_losses.py b/TensorFlow/contrib/cv/InsightFace_TF/losses/face_losses.py
new file mode 100644
index 0000000000000000000000000000000000000000..be97b115bcab122fc571c1e6ebb71a50090da98f
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/losses/face_losses.py
@@ -0,0 +1,122 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import math
+
+
+def arcface_loss(embedding, labels, out_num, w_init=None, s=64., m=0.5):
+ '''
+ :param embedding: the input embedding vectors
+ :param labels: the input labels, the shape should be eg: (batch_size, 1)
+ :param s: scalar value default is 64
+ :param out_num: output class num
+ :param m: the margin value, default is 0.5
+ :return: the final cacualted output, this output is send into the tf.nn.softmax directly
+ '''
+ cos_m = math.cos(m)
+ sin_m = math.sin(m)
+ mm = sin_m * m # issue 1
+ threshold = math.cos(math.pi - m)
+ with tf.variable_scope('arcface_loss'):
+ # inputs and weights norm
+ embedding_norm = tf.norm(embedding, axis=1, keep_dims=True)
+ embedding = tf.div(embedding, embedding_norm, name='norm_embedding')
+ weights = tf.get_variable(name='embedding_weights', shape=(embedding.get_shape().as_list()[-1], out_num),
+ initializer=w_init, dtype=tf.float32)
+ weights_norm = tf.norm(weights, axis=0, keep_dims=True)
+ weights = tf.div(weights, weights_norm, name='norm_weights')
+ # cos(theta+m)
+ cos_t = tf.matmul(embedding, weights, name='cos_t')
+ cos_t2 = tf.square(cos_t, name='cos_2')
+ sin_t2 = tf.subtract(1., cos_t2, name='sin_2')
+ sin_t = tf.sqrt(sin_t2, name='sin_t')
+ cos_mt = s * tf.subtract(tf.multiply(cos_t, cos_m), tf.multiply(sin_t, sin_m), name='cos_mt')
+
+ # this condition controls the theta+m should in range [0, pi]
+ # 0<=theta+m<=pi
+ # -m<=theta<=pi-m
+ cond_v = cos_t - threshold
+ cond = tf.cast(tf.nn.relu(cond_v, name='if_else'), dtype=tf.bool)
+
+ keep_val = s*(cos_t - mm)
+ cos_mt_temp = tf.where(cond, cos_mt, keep_val)
+
+ mask = tf.one_hot(labels, depth=out_num, name='one_hot_mask')
+ # mask = tf.squeeze(mask, 1)
+ inv_mask = tf.subtract(1., mask, name='inverse_mask')
+
+ s_cos_t = tf.multiply(s, cos_t, name='scalar_cos_t')
+
+ output = tf.add(tf.multiply(s_cos_t, inv_mask), tf.multiply(cos_mt_temp, mask), name='arcface_loss_output')
+ return output
+
+
+def cosineface_losses(embedding, labels, out_num, w_init=None, s=30., m=0.4):
+ '''
+ :param embedding: the input embedding vectors
+ :param labels: the input labels, the shape should be eg: (batch_size, 1)
+ :param s: scalar value, default is 30
+ :param out_num: output class num
+ :param m: the margin value, default is 0.4
+ :return: the final cacualted output, this output is send into the tf.nn.softmax directly
+ '''
+ with tf.variable_scope('cosineface_loss'):
+ # inputs and weights norm
+ embedding_norm = tf.norm(embedding, axis=1, keep_dims=True)
+ embedding = tf.div(embedding, embedding_norm, name='norm_embedding')
+ weights = tf.get_variable(name='embedding_weights', shape=(embedding.get_shape().as_list()[-1], out_num),
+ initializer=w_init, dtype=tf.float32)
+ weights_norm = tf.norm(weights, axis=0, keep_dims=True)
+ weights = tf.div(weights, weights_norm, name='norm_weights')
+ # cos_theta - m
+ cos_t = tf.matmul(embedding, weights, name='cos_t')
+ cos_t_m = tf.subtract(cos_t, m, name='cos_t_m')
+
+ mask = tf.one_hot(labels, depth=out_num, name='one_hot_mask')
+ inv_mask = tf.subtract(1., mask, name='inverse_mask')
+
+ output = tf.add(s * tf.multiply(cos_t, inv_mask), s * tf.multiply(cos_t_m, mask), name='cosineface_loss_output')
+ return output
+
+
+def combine_loss_val(embedding, labels, w_init, out_num, margin_a, margin_m, margin_b, s):
+ '''
+ This code is contributed by RogerLo. Thanks for you contribution.
+
+ :param embedding: the input embedding vectors
+ :param labels: the input labels, the shape should be eg: (batch_size, 1)
+ :param s: scalar value default is 64
+ :param out_num: output class num
+ :param m: the margin value, default is 0.5
+ :return: the final cacualted output, this output is send into the tf.nn.softmax directly
+ '''
+ weights = tf.get_variable(name='embedding_weights', shape=(embedding.get_shape().as_list()[-1], out_num),
+ initializer=w_init, dtype=tf.float32)
+ weights_unit = tf.nn.l2_normalize(weights, axis=0)
+ embedding_unit = tf.nn.l2_normalize(embedding, axis=1)
+ cos_t = tf.matmul(embedding_unit, weights_unit)
+ ordinal = tf.constant(list(range(0, embedding.get_shape().as_list()[0])), tf.int64)
+ ordinal_y = tf.stack([ordinal, labels], axis=1)
+ zy = cos_t * s
+ sel_cos_t = tf.gather_nd(zy, ordinal_y)
+ if margin_a != 1.0 or margin_m != 0.0 or margin_b != 0.0:
+ if margin_a == 1.0 and margin_m == 0.0:
+ s_m = s * margin_b
+ new_zy = sel_cos_t - s_m
+ else:
+ cos_value = sel_cos_t / s
+ t = tf.acos(cos_value)
+ if margin_a != 1.0:
+ t = t * margin_a
+ if margin_m > 0.0:
+ t = t + margin_m
+ body = tf.cos(t)
+ if margin_b > 0.0:
+ body = body - margin_b
+ new_zy = body * s
+ updated_logits = tf.add(zy, tf.scatter_nd(ordinal_y, tf.subtract(new_zy, sel_cos_t), zy.get_shape()))
+ loss = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(labels=labels, logits=updated_logits))
+ predict_cls = tf.argmax(updated_logits, 1)
+ accuracy = tf.reduce_mean(tf.cast(tf.equal(tf.cast(predict_cls, tf.int64), tf.cast(labels, tf.int64)), 'float'))
+ predict_cls_s = tf.argmax(zy, 1)
+ accuracy_s = tf.reduce_mean(tf.cast(tf.equal(tf.cast(predict_cls_s, tf.int64), tf.cast(labels, tf.int64)), 'float'))
+ return zy, loss, accuracy, accuracy_s, predict_cls_s
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR.py
new file mode 100644
index 0000000000000000000000000000000000000000..e8e8f26d98add1887f1505996cd8b53ee950dace
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR.py
@@ -0,0 +1,418 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+from tensorflow.contrib.layers.python.layers import utils
+import collections
+from tensorlayer.layers import Layer, list_remove_repeat
+
+
+class ElementwiseLayer(Layer):
+ """
+ The :class:`ElementwiseLayer` class combines multiple :class:`Layer` which have the same output shapes by a given elemwise-wise operation.
+
+ Parameters
+ ----------
+ layer : a list of :class:`Layer` instances
+ The `Layer` class feeding into this layer.
+ combine_fn : a TensorFlow elemwise-merge function
+ e.g. AND is ``tf.minimum`` ; OR is ``tf.maximum`` ; ADD is ``tf.add`` ; MUL is ``tf.multiply`` and so on.
+ See `TensorFlow Math API `_ .
+ name : a string or None
+ An optional name to attach to this layer.
+ """
+ def __init__(
+ self,
+ layer = [],
+ combine_fn = tf.minimum,
+ name ='elementwise_layer',
+ act = None,
+ ):
+ Layer.__init__(self, name=name)
+
+ if act:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s, act:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__, act.__name__))
+ else:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__))
+
+ self.outputs = layer[0].outputs
+ # print(self.outputs._shape, type(self.outputs._shape))
+ for l in layer[1:]:
+ # assert str(self.outputs.get_shape()) == str(l.outputs.get_shape()), "Hint: the input shapes should be the same. %s != %s" % (self.outputs.get_shape() , str(l.outputs.get_shape()))
+ self.outputs = combine_fn(self.outputs, l.outputs, name=name)
+ if act:
+ self.outputs = act(self.outputs)
+ self.all_layers = list(layer[0].all_layers)
+ self.all_params = list(layer[0].all_params)
+ self.all_drop = dict(layer[0].all_drop)
+
+ for i in range(1, len(layer)):
+ self.all_layers.extend(list(layer[i].all_layers))
+ self.all_params.extend(list(layer[i].all_params))
+ self.all_drop.update(dict(layer[i].all_drop))
+
+ self.all_layers = list_remove_repeat(self.all_layers)
+ self.all_params = list_remove_repeat(self.all_params)
+
+
+class BatchNormLayer(Layer):
+ """
+ The :class:`BatchNormLayer` class is a normalization layer, see ``tf.nn.batch_normalization`` and ``tf.nn.moments``.
+
+ Batch normalization on fully-connected or convolutional maps.
+
+ ```
+ https://www.tensorflow.org/api_docs/python/tf/cond
+ If x < y, the tf.add operation will be executed and tf.square operation will not be executed.
+ Since z is needed for at least one branch of the cond, the tf.multiply operation is always executed, unconditionally.
+ ```
+
+ Parameters
+ -----------
+ layer : a :class:`Layer` instance
+ The `Layer` class feeding into this layer.
+ decay : float, default is 0.9.
+ A decay factor for ExponentialMovingAverage, use larger value for large dataset.
+ epsilon : float
+ A small float number to avoid dividing by 0.
+ act : activation function.
+ is_train : boolean
+ Whether train or inference.
+ beta_init : beta initializer
+ The initializer for initializing beta
+ gamma_init : gamma initializer
+ The initializer for initializing gamma
+ dtype : tf.float32 (default) or tf.float16
+ name : a string or None
+ An optional name to attach to this layer.
+
+ References
+ ----------
+ - `Source `_
+ - `stackoverflow `_
+
+ """
+
+ def __init__(
+ self,
+ layer=None,
+ decay=0.9,
+ epsilon=2e-5,
+ act=tf.identity,
+ is_train=False,
+ fix_gamma=True,
+ beta_init=tf.zeros_initializer,
+ gamma_init=tf.random_normal_initializer(mean=1.0, stddev=0.002), # tf.ones_initializer,
+ # dtype = tf.float32,
+ trainable=None,
+ name='batchnorm_layer',
+ ):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ print(" [TL] BatchNormLayer %s: decay:%f epsilon:%f act:%s is_train:%s" % (self.name, decay, epsilon, act.__name__, is_train))
+ x_shape = self.inputs.get_shape()
+ params_shape = x_shape[-1:]
+
+ from tensorflow.python.training import moving_averages
+ from tensorflow.python.ops import control_flow_ops
+
+ with tf.variable_scope(name) as vs:
+ axis = list(range(len(x_shape) - 1))
+
+ ## 1. beta, gamma
+ if tf.__version__ > '0.12.1' and beta_init == tf.zeros_initializer:
+ beta_init = beta_init()
+ beta = tf.get_variable('beta', shape=params_shape, initializer=beta_init, dtype=tf.float32, trainable=is_train) #, restore=restore)
+
+ gamma = tf.get_variable(
+ 'gamma',
+ shape=params_shape,
+ initializer=gamma_init,
+ dtype=tf.float32,
+ trainable=fix_gamma,
+ ) #restore=restore)
+
+ ## 2.
+ if tf.__version__ > '0.12.1':
+ moving_mean_init = tf.zeros_initializer()
+ else:
+ moving_mean_init = tf.zeros_initializer
+ moving_mean = tf.get_variable('moving_mean', params_shape, initializer=moving_mean_init, dtype=tf.float32, trainable=False) # restore=restore)
+ moving_variance = tf.get_variable(
+ 'moving_variance',
+ params_shape,
+ initializer=tf.constant_initializer(1.),
+ dtype=tf.float32,
+ trainable=False,
+ ) # restore=restore)
+
+ ## 3.
+ # These ops will only be preformed when training.
+ mean, variance = tf.nn.moments(self.inputs, axis)
+ try: # TF12
+ update_moving_mean = moving_averages.assign_moving_average(moving_mean, mean, decay, zero_debias=False) # if zero_debias=True, has bias
+ update_moving_variance = moving_averages.assign_moving_average(
+ moving_variance, variance, decay, zero_debias=False) # if zero_debias=True, has bias
+ # print("TF12 moving")
+ except Exception as e: # TF11
+ update_moving_mean = moving_averages.assign_moving_average(moving_mean, mean, decay)
+ update_moving_variance = moving_averages.assign_moving_average(moving_variance, variance, decay)
+ # print("TF11 moving")
+
+ def mean_var_with_update():
+ with tf.control_dependencies([update_moving_mean, update_moving_variance]):
+ return tf.identity(mean), tf.identity(variance)
+ if trainable:
+ mean, var = mean_var_with_update()
+ print(mean)
+ print(var)
+ self.outputs = act(tf.nn.batch_normalization(self.inputs, mean, var, beta, gamma, epsilon))
+ else:
+ self.outputs = act(tf.nn.batch_normalization(self.inputs, moving_mean, moving_variance, beta, gamma, epsilon))
+ variables = [beta, gamma, moving_mean, moving_variance]
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+ self.all_layers.extend([self.outputs])
+ self.all_params.extend(variables)
+
+
+def subsample(inputs, factor, scope=None):
+ if factor == 1:
+ return inputs
+ else:
+ return tl.layers.MaxPool2d(inputs, [1, 1], strides=(factor, factor), name=scope)
+
+
+def conv2d_same(inputs, num_outputs, kernel_size, strides, rate=1, w_init=None, scope=None, trainable=None):
+ '''
+ Reference slim resnet
+ :param inputs:
+ :param num_outputs:
+ :param kernel_size:
+ :param strides:
+ :param rate:
+ :param scope:
+ :return:
+ '''
+ if strides == 1:
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='SAME', name=scope,
+ use_cudnn_on_gpu=True)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size),
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ return nets
+ else:
+ kernel_size_effective = kernel_size + (kernel_size - 1) * (rate - 1)
+ pad_total = kernel_size_effective - 1
+ pad_beg = pad_total // 2
+ pad_end = pad_total - pad_beg
+ inputs = tl.layers.PadLayer(inputs, [[0, 0], [pad_beg, pad_end], [pad_beg, pad_end], [0, 0]], name='padding_%s' % scope)
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='VALID', name=scope,
+ use_cudnn_on_gpu=True)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ return nets
+
+
+def bottleneck_IR(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut = BatchNormLayer(shortcut, act=tf.identity, is_train=True, trainable=trainable, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = BatchNormLayer(inputs, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn1')
+ residual = tl.layers.Conv2d(residual, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual = BatchNormLayer(residual, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn2')
+ # bottleneck prelu
+ residual = tl.layers.PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ output = ElementwiseLayer(layer=[shortcut, residual],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=None)
+ return output
+
+
+def bottleneck_IR_SE(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut = BatchNormLayer(shortcut, act=tf.identity, is_train=True, trainable=trainable, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = BatchNormLayer(inputs, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn1')
+ residual = tl.layers.Conv2d(residual, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual = BatchNormLayer(residual, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn2')
+ # bottleneck prelu
+ residual = tl.layers.PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ # squeeze
+ squeeze = tl.layers.InputLayer(tf.reduce_mean(residual.outputs, axis=[1, 2]), name='squeeze_layer')
+ # excitation
+ excitation1 = tl.layers.DenseLayer(squeeze, n_units=int(depth/16.0), act=tf.nn.relu,
+ W_init=w_init, name='excitation_1')
+ # excitation1 = tl.layers.PReluLayer(excitation1, name='excitation_prelu')
+ excitation2 = tl.layers.DenseLayer(excitation1, n_units=depth, act=tf.nn.sigmoid,
+ W_init=w_init, name='excitation_2')
+ # scale
+ scale = tl.layers.ReshapeLayer(excitation2, shape=[tf.shape(excitation2.outputs)[0], 1, 1, depth], name='excitation_reshape')
+
+ residual_se = ElementwiseLayer(layer=[residual, scale],
+ combine_fn=tf.multiply,
+ name='scale_layer',
+ act=None)
+
+ output = ElementwiseLayer(layer=[shortcut, residual_se],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=tf.nn.relu)
+ return output
+
+
+def resnet(inputs, bottle_neck, blocks, w_init=None, trainable=None, reuse=False, keep_rate=None, scope=None):
+ with tf.variable_scope(scope, reuse=reuse):
+ # inputs = tf.subtract(inputs, 127.5)
+ # inputs = tf.multiply(inputs, 0.0078125)
+ net_inputs = tl.layers.InputLayer(inputs, name='input_layer')
+ if bottle_neck:
+ net = tl.layers.Conv2d(net_inputs, n_filter=64, filter_size=(3, 3), strides=(1, 1),
+ act=None, W_init=w_init, b_init=None, name='conv1', use_cudnn_on_gpu=True)
+ net = BatchNormLayer(net, act=tf.identity, name='bn0', is_train=True, trainable=trainable)
+ net = tl.layers.PReluLayer(net, name='prelu0')
+ else:
+ raise ValueError('The standard resnet must support the bottleneck layer')
+ for block in blocks:
+ with tf.variable_scope(block.scope):
+ for i, var in enumerate(block.args):
+ with tf.variable_scope('unit_%d' % (i+1)):
+ net = block.unit_fn(net, depth=var['depth'], depth_bottleneck=var['depth_bottleneck'],
+ w_init=w_init, stride=var['stride'], rate=var['rate'], scope=None,
+ trainable=trainable)
+ net = BatchNormLayer(net, act=tf.identity, is_train=True, name='E_BN1', trainable=trainable)
+ # net = tl.layers.DropoutLayer(net, keep=0.4, name='E_Dropout')
+ net.outputs = npu_ops.dropout(net.outputs, keep_prob=keep_rate, name='E_Dropout')
+ net_shape = net.outputs.get_shape()
+ net = tl.layers.ReshapeLayer(net, shape=[-1, net_shape[1]*net_shape[2]*net_shape[3]], name='E_Reshapelayer')
+ net = tl.layers.DenseLayer(net, n_units=512, W_init=w_init, name='E_DenseLayer')
+ net = BatchNormLayer(net, act=tf.identity, is_train=True, fix_gamma=False, trainable=trainable, name='E_BN2')
+ return net
+
+
+class Block(collections.namedtuple('Block', ['scope', 'unit_fn', 'args'])):
+ """A named tuple describing a ResNet block.
+
+ Its parts are:
+ scope: The scope of the `Block`.
+ unit_fn: The ResNet unit function which takes as input a `Tensor` and
+ returns another `Tensor` with the output of the ResNet unit.
+ args: A list of length equal to the number of units in the `Block`. The list
+ contains one (depth, depth_bottleneck, stride) tuple for each unit in the
+ block to serve as argument to unit_fn.
+ """
+
+
+def resnetse_v1_block(scope, base_depth, num_units, stride, rate=1, unit_fn=None):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, unit_fn, [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': stride,
+ 'rate': rate
+ }] + [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': 1,
+ 'rate': rate
+ }] * (num_units - 1))
+
+
+def get_resnet(inputs, num_layers, type=None, w_init=None, trainable=None, sess=None, reuse=False, keep_rate=None):
+ if type == 'ir':
+ unit_fn = bottleneck_IR
+ elif type == 'se_ir':
+ unit_fn = bottleneck_IR_SE
+ else:
+ raise ValueError('the input fn is unknown')
+
+ if num_layers == 50:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=14, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 101:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=13, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=30, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 152:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=8, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=36, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ else:
+ raise ValueError('Resnet layer %d is not supported now.' % num_layers)
+ net = resnet(inputs=inputs,
+ bottle_neck=True,
+ blocks=blocks,
+ w_init=w_init,
+ trainable=trainable,
+ reuse=reuse,
+ keep_rate = keep_rate,
+ scope='resnet_v1_%d' % num_layers)
+ return net
+
+
+if __name__ == '__main__':
+ x = tf.placeholder(dtype=tf.float32, shape=[None, 112, 112, 3], name='input_place')
+ sess = tf.Session(config=npu_config_proto())
+ # w_init = tf.truncated_normal_initializer(mean=10, stddev=5e-2)
+ w_init = tf.contrib.layers.xavier_initializer(uniform=False)
+ # test resnetse
+ nets = get_resnet(x, 50, type='ir', w_init=w_init, sess=sess)
+ tl.layers.initialize_global_variables(sess)
+
+ for p in tl.layers.get_variables_with_name('W_conv2d', True, True):
+ print(p.op.name)
+ print('##############'*30)
+ with sess:
+ nets.print_params()
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_GBN.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_GBN.py
new file mode 100644
index 0000000000000000000000000000000000000000..49a76e607ec61570d409dbd1c2a2c906ef1a89b0
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_GBN.py
@@ -0,0 +1,354 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+from tensorflow.contrib.layers.python.layers import utils
+import collections
+from tensorlayer.layers import Layer, list_remove_repeat
+from tl_layers_modify import GroupNormLayer
+
+
+class ElementwiseLayer(Layer):
+ """
+ The :class:`ElementwiseLayer` class combines multiple :class:`Layer` which have the same output shapes by a given elemwise-wise operation.
+
+ Parameters
+ ----------
+ layer : a list of :class:`Layer` instances
+ The `Layer` class feeding into this layer.
+ combine_fn : a TensorFlow elemwise-merge function
+ e.g. AND is ``tf.minimum`` ; OR is ``tf.maximum`` ; ADD is ``tf.add`` ; MUL is ``tf.multiply`` and so on.
+ See `TensorFlow Math API `_ .
+ name : a string or None
+ An optional name to attach to this layer.
+ """
+ def __init__(
+ self,
+ layer = [],
+ combine_fn = tf.minimum,
+ name ='elementwise_layer',
+ act = None,
+ ):
+ Layer.__init__(self, name=name)
+
+ if act:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s, act:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__, act.__name__))
+ else:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__))
+
+ self.outputs = layer[0].outputs
+ # print(self.outputs._shape, type(self.outputs._shape))
+ for l in layer[1:]:
+ # assert str(self.outputs.get_shape()) == str(l.outputs.get_shape()), "Hint: the input shapes should be the same. %s != %s" % (self.outputs.get_shape() , str(l.outputs.get_shape()))
+ self.outputs = combine_fn(self.outputs, l.outputs, name=name)
+ if act:
+ self.outputs = act(self.outputs)
+ self.all_layers = list(layer[0].all_layers)
+ self.all_params = list(layer[0].all_params)
+ self.all_drop = dict(layer[0].all_drop)
+
+ for i in range(1, len(layer)):
+ self.all_layers.extend(list(layer[i].all_layers))
+ self.all_params.extend(list(layer[i].all_params))
+ self.all_drop.update(dict(layer[i].all_drop))
+
+ self.all_layers = list_remove_repeat(self.all_layers)
+ self.all_params = list_remove_repeat(self.all_params)
+
+
+def subsample(inputs, factor, scope=None):
+ if factor == 1:
+ return inputs
+ else:
+ return tl.layers.MaxPool2d(inputs, [1, 1], strides=(factor, factor), name=scope)
+
+
+def conv2d_same(inputs, num_outputs, kernel_size, strides, rate=1, w_init=None, scope=None, trainable=None):
+ '''
+ Reference slim resnet
+ :param inputs:
+ :param num_outputs:
+ :param kernel_size:
+ :param strides:
+ :param rate:
+ :param scope:
+ :return:
+ '''
+ if strides == 1:
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='SAME', name=scope,
+ use_cudnn_on_gpu=True)
+ nets = GroupNormLayer(layer=nets, act=tf.identity, name=scope+'_bn/GroupNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size),
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets = GroupNormLayer(layer=nets, act=tf.identity, name=scope+'_bn/GroupNorm')
+ return nets
+ else:
+ kernel_size_effective = kernel_size + (kernel_size - 1) * (rate - 1)
+ pad_total = kernel_size_effective - 1
+ pad_beg = pad_total // 2
+ pad_end = pad_total - pad_beg
+ inputs = tl.layers.PadLayer(inputs, [[0, 0], [pad_beg, pad_end], [pad_beg, pad_end], [0, 0]], name='padding_%s' % scope)
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='VALID', name=scope,
+ use_cudnn_on_gpu=True)
+ nets = GroupNormLayer(layer=nets, act=tf.identity, name=scope+'_bn/GroupNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets = GroupNormLayer(layer=nets, act=tf.identity, name=scope+'_bn/GroupNorm')
+ return nets
+
+
+def bottleneck(inputs, depth, depth_bottleneck, stride, rate=1, scope=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ b_init=None, name='shortcut_conv')
+ shortcut = GroupNormLayer(layer=shortcut, act=tf.identity, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = tl.layers.Conv2d(inputs, depth_bottleneck, filter_size=(1, 1), strides=(1, 1), act=None, b_init=None,
+ name='conv1')
+ residual = GroupNormLayer(layer=residual, act=tf.nn.relu, name='conv1_bn/BatchNorm')
+
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth_bottleneck, kernel_size=3, strides= stride, rate=rate, scope='conv2')
+
+ # bottleneck layer 3
+ residual = tl.layers.Conv2d(residual, depth, filter_size=(1, 1), strides=(1, 1), act=None, b_init=None,
+ name='conv3')
+ residual = GroupNormLayer(layer=residual, act=tf.identity, name='conv3_bn/BatchNorm',
+ scale_init=tf.constant_initializer(0.0))
+ output = ElementwiseLayer(layer=[shortcut, residual],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=tf.nn.relu)
+ return output
+
+
+def bottleneck_IR(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut = GroupNormLayer(layer=shortcut, act=tf.identity, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = GroupNormLayer(layer=inputs, act=tf.identity, name='conv1_bn1')
+ residual = tl.layers.Conv2d(residual, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual = GroupNormLayer(layer=residual, act=tf.identity, name='conv1_bn2')
+ # bottleneck prelu
+ residual = tl.layers.PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ output = ElementwiseLayer(layer=[shortcut, residual],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=None)
+ return output
+
+
+def bottleneck_IR_SE(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut = GroupNormLayer(layer=shortcut, act=tf.identity, name='shortcut_bn/BatchNorm')
+ residual = GroupNormLayer(layer=inputs, act=tf.identity, name='conv1_bn1')
+ residual = tl.layers.Conv2d(residual, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual = GroupNormLayer(layer=residual, act=tf.identity, name='conv1_bn2')
+ # bottleneck prelu
+ residual = tl.layers.PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ # squeeze
+ squeeze = tl.layers.InputLayer(tf.reduce_mean(residual.outputs, axis=[1, 2]), name='squeeze_layer')
+ # excitation
+ excitation1 = tl.layers.DenseLayer(squeeze, n_units=int(depth/16.0), act=tf.nn.relu,
+ W_init=w_init, name='excitation_1')
+ # excitation1 = tl.layers.PReluLayer(excitation1, name='excitation_prelu')
+ excitation2 = tl.layers.DenseLayer(excitation1, n_units=depth, act=tf.nn.sigmoid,
+ W_init=w_init, name='excitation_2')
+ # scale
+ scale = tl.layers.ReshapeLayer(excitation2, shape=[tf.shape(excitation2.outputs)[0], 1, 1, depth], name='excitation_reshape')
+
+ residual_se = ElementwiseLayer(layer=[residual, scale],
+ combine_fn=tf.multiply,
+ name='scale_layer',
+ act=None)
+
+ output = ElementwiseLayer(layer=[shortcut, residual_se],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=tf.nn.relu)
+ return output
+
+
+def resnet(inputs, bottle_neck, blocks, w_init=None, trainable=None, scope=None):
+ with tf.variable_scope(scope):
+ net_inputs = tl.layers.InputLayer(inputs, name='input_layer')
+ if bottle_neck:
+ net = tl.layers.Conv2d(net_inputs, n_filter=64, filter_size=(3, 3), strides=(1, 1),
+ act=None, W_init=w_init, b_init=None, name='conv1', use_cudnn_on_gpu=True)
+ net = GroupNormLayer(layer=net, act=tf.identity, name='group_norm_0')
+ net = tl.layers.PReluLayer(net, name='prelu0')
+ else:
+ raise ValueError('The standard resnet must support the bottleneck layer')
+ for block in blocks:
+ with tf.variable_scope(block.scope):
+ for i, var in enumerate(block.args):
+ with tf.variable_scope('unit_%d' % (i+1)):
+ net = block.unit_fn(net, depth=var['depth'], depth_bottleneck=var['depth_bottleneck'],
+ w_init=w_init, stride=var['stride'], rate=var['rate'], scope=None,
+ trainable=trainable)
+ net = GroupNormLayer(layer=net, act=tf.identity, name='E_GN_0')
+ net = tl.layers.DropoutLayer(net, keep=0.4, name='E_Dropout')
+ net_shape = net.outputs.get_shape()
+ net = tl.layers.ReshapeLayer(net, shape=[-1, net_shape[1]*net_shape[2]*net_shape[3]], name='E_Reshapelayer')
+ net = tl.layers.DenseLayer(net, n_units=512, W_init=w_init, name='E_DenseLayer')
+ # net = GroupNormLayer(layer=net, act=tf.identity, name='E_GN_1')
+ return net
+
+
+class Block(collections.namedtuple('Block', ['scope', 'unit_fn', 'args'])):
+ """A named tuple describing a ResNet block.
+
+ Its parts are:
+ scope: The scope of the `Block`.
+ unit_fn: The ResNet unit function which takes as input a `Tensor` and
+ returns another `Tensor` with the output of the ResNet unit.
+ args: A list of length equal to the number of units in the `Block`. The list
+ contains one (depth, depth_bottleneck, stride) tuple for each unit in the
+ block to serve as argument to unit_fn.
+ """
+
+
+def resnetse_v1_block(scope, base_depth, num_units, stride, rate=1, unit_fn=None):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, unit_fn, [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': stride,
+ 'rate': rate
+ }] + [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': 1,
+ 'rate': rate
+ }] * (num_units - 1))
+
+
+def resnetse_v1_block_2(scope, base_depth, num_units, stride, rate=1, unit_fn=None):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, unit_fn, [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': 1,
+ 'rate': rate
+ }] * (num_units - 1) + [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': stride,
+ 'rate': rate
+ }])
+
+
+def get_resnet(inputs, num_layers, type=None, w_init=None, trainable=None, sess=None):
+ if type == 'ir':
+ unit_fn = bottleneck_IR
+ elif type == 'se_ir':
+ unit_fn = bottleneck_IR_SE
+ # elif type == 'resnet':
+ # unit_fn = bottleneck
+ # blocks = [
+ # resnetse_v1_block_2('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ # resnetse_v1_block_2('block2', base_depth=128, num_units=4, stride=2, rate=1, unit_fn=unit_fn),
+ # resnetse_v1_block_2('block3', base_depth=256, num_units=6, stride=2, rate=1, unit_fn=unit_fn),
+ # resnetse_v1_block_2('block4', base_depth=512, num_units=3, stride=1, rate=1, unit_fn=unit_fn)
+ # ]
+ else:
+ raise ValueError('the input fn is unknown')
+
+ if num_layers == 50:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=14, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 101:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=13, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=30, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 152:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=8, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=36, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ else:
+ raise ValueError('Resnet layer %d is not supported now.' % num_layers)
+ net = resnet(inputs=inputs,
+ bottle_neck=True,
+ blocks=blocks,
+ w_init=w_init,
+ trainable=trainable,
+ scope='resnet_v1_%d' % num_layers)
+ return net
+
+
+if __name__ == '__main__':
+ x = tf.placeholder(dtype=tf.float32, shape=[None, 112, 112, 3], name='input_place')
+ sess = tf.Session(config=npu_config_proto())
+ # w_init = tf.truncated_normal_initializer(mean=10, stddev=5e-2)
+ w_init = tf.contrib.layers.xavier_initializer(uniform=False)
+ # test resnetse
+ nets = get_resnet(x, 50, type='ir', w_init=w_init, sess=sess)
+ tl.layers.initialize_global_variables(sess)
+
+ for p in tl.layers.get_variables_with_name('W_conv2d', True, True):
+ print(p.op.name)
+ print('##############'*30)
+ with sess:
+ nets.print_params()
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_MGPU.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_MGPU.py
new file mode 100644
index 0000000000000000000000000000000000000000..3f96317c6c36425f4c1007881a014110b70273b1
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_MGPU.py
@@ -0,0 +1,242 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+from tensorflow.contrib.layers.python.layers import utils
+import collections
+from tl_layers_modify import ElementwiseLayer, BatchNormLayer, Conv2d, PReluLayer, DenseLayer
+
+
+def subsample(inputs, factor, scope=None):
+ if factor == 1:
+ return inputs
+ else:
+ return tl.layers.MaxPool2d(inputs, [1, 1], strides=(factor, factor), name=scope)
+
+
+def conv2d_same(inputs, num_outputs, kernel_size, strides, rate=1, w_init=None, scope=None, trainable=None):
+ '''
+ Reference slim resnet
+ :param inputs:
+ :param num_outputs:
+ :param kernel_size:
+ :param strides:
+ :param rate:
+ :param scope:
+ :return:
+ '''
+ if strides == 1:
+ if rate == 1:
+ nets = Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='SAME', name=scope,
+ use_cudnn_on_gpu=True)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size),
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ return nets
+ else:
+ kernel_size_effective = kernel_size + (kernel_size - 1) * (rate - 1)
+ pad_total = kernel_size_effective - 1
+ pad_beg = pad_total // 2
+ pad_end = pad_total - pad_beg
+ inputs = tl.layers.PadLayer(inputs, [[0, 0], [pad_beg, pad_end], [pad_beg, pad_end], [0, 0]], name='padding_%s' % scope)
+ if rate == 1:
+ nets = Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='VALID', name=scope,
+ use_cudnn_on_gpu=True)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ return nets
+
+
+def bottleneck_IR(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut = BatchNormLayer(shortcut, act=tf.identity, is_train=True, trainable=trainable, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = BatchNormLayer(inputs, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn1')
+ residual = Conv2d(residual, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual = BatchNormLayer(residual, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn2')
+ # bottleneck prelu
+ residual = PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ output = ElementwiseLayer(layer=[shortcut, residual],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=None)
+ return output
+
+
+def bottleneck_IR_SE(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut = BatchNormLayer(shortcut, act=tf.identity, is_train=True, trainable=trainable, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = BatchNormLayer(inputs, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn1')
+ residual = Conv2d(residual, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual = BatchNormLayer(residual, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn2')
+ # bottleneck prelu
+ residual = PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ # squeeze
+ squeeze = tl.layers.InputLayer(tf.reduce_mean(residual.outputs, axis=[1, 2]), name='squeeze_layer')
+ # excitation
+ excitation1 = DenseLayer(squeeze, n_units=int(depth/16.0), act=tf.nn.relu,
+ W_init=w_init, name='excitation_1')
+ # excitation1 = tl.layers.PReluLayer(excitation1, name='excitation_prelu')
+ excitation2 = DenseLayer(excitation1, n_units=depth, act=tf.nn.sigmoid,
+ W_init=w_init, name='excitation_2')
+ # scale
+ scale = tl.layers.ReshapeLayer(excitation2, shape=[tf.shape(excitation2.outputs)[0], 1, 1, depth], name='excitation_reshape')
+
+ residual_se = ElementwiseLayer(layer=[residual, scale],
+ combine_fn=tf.multiply,
+ name='scale_layer',
+ act=None)
+
+ output = ElementwiseLayer(layer=[shortcut, residual_se],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=tf.nn.relu)
+ return output
+
+
+def resnet(inputs, bottle_neck, blocks, w_init=None, trainable=None, keep_rate=None, scope=None):
+ with tf.variable_scope(scope):
+ net_inputs = tl.layers.InputLayer(inputs, name='input_layer')
+ if bottle_neck:
+ net = Conv2d(net_inputs, n_filter=64, filter_size=(3, 3), strides=(1, 1),
+ act=None, W_init=w_init, b_init=None, name='conv1', use_cudnn_on_gpu=True)
+ net = BatchNormLayer(net, act=tf.identity, name='bn0', is_train=True, trainable=trainable)
+ net = PReluLayer(net, name='prelu0')
+ else:
+ raise ValueError('The standard resnet must support the bottleneck layer')
+ for block in blocks:
+ with tf.variable_scope(block.scope):
+ for i, var in enumerate(block.args):
+ with tf.variable_scope('unit_%d' % (i+1)):
+ net = block.unit_fn(net, depth=var['depth'], depth_bottleneck=var['depth_bottleneck'],
+ w_init=w_init, stride=var['stride'], rate=var['rate'], scope=None,
+ trainable=trainable)
+ net = BatchNormLayer(net, act=tf.identity, is_train=True, name='E_BN1', trainable=trainable)
+ net = tl.layers.DropoutLayer(net, keep=keep_rate, name='E_Dropout')
+ net_shape = net.outputs.get_shape()
+ net = tl.layers.ReshapeLayer(net, shape=[-1, net_shape[1]*net_shape[2]*net_shape[3]], name='E_Reshapelayer')
+ net = DenseLayer(net, n_units=512, W_init=w_init, name='E_DenseLayer')
+ net = BatchNormLayer(net, act=tf.identity, is_train=True, fix_gamma=False, trainable=trainable, name='E_BN2')
+ return net
+
+
+class Block(collections.namedtuple('Block', ['scope', 'unit_fn', 'args'])):
+ """A named tuple describing a ResNet block.
+
+ Its parts are:
+ scope: The scope of the `Block`.
+ unit_fn: The ResNet unit function which takes as input a `Tensor` and
+ returns another `Tensor` with the output of the ResNet unit.
+ args: A list of length equal to the number of units in the `Block`. The list
+ contains one (depth, depth_bottleneck, stride) tuple for each unit in the
+ block to serve as argument to unit_fn.
+ """
+
+
+def resnetse_v1_block(scope, base_depth, num_units, stride, rate=1, unit_fn=None):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, unit_fn, [{
+ 'depth': base_depth,
+ 'depth_bottleneck': base_depth,
+ 'stride': stride,
+ 'rate': rate
+ }] + [{
+ 'depth': base_depth,
+ 'depth_bottleneck': base_depth,
+ 'stride': 1,
+ 'rate': rate
+ }] * (num_units - 1))
+
+
+def get_resnet(inputs, num_layers, type=None, w_init=None, trainable=None, keep_rate=None, sess=None):
+ if type == 'ir':
+ unit_fn = bottleneck_IR
+ elif type == 'se_ir':
+ unit_fn = bottleneck_IR_SE
+ else:
+ raise ValueError('the input fn is unknown')
+
+ if num_layers == 50:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=14, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 100:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=13, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=30, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 152:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=8, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=36, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ else:
+ raise ValueError('Resnet layer %d is not supported now.' % num_layers)
+ net = resnet(inputs=inputs,
+ bottle_neck=True,
+ blocks=blocks,
+ w_init=w_init,
+ trainable=trainable,
+ keep_rate=keep_rate,
+ scope='resnet_v1_%d' % num_layers)
+ return net
+
+
+if __name__ == '__main__':
+ x = tf.placeholder(dtype=tf.float32, shape=[None, 112, 112, 3], name='input_place')
+ sess = tf.Session(config=npu_config_proto())
+ # w_init = tf.truncated_normal_initializer(mean=10, stddev=5e-2)
+ w_init = tf.contrib.layers.xavier_initializer(uniform=False)
+ # test resnetse
+ nets = get_resnet(x, 50, type='ir', w_init=w_init, sess=sess)
+ tl.layers.initialize_global_variables(sess)
+
+ for p in tl.layers.get_variables_with_name('W_conv2d', True, True):
+ print(p.op.name)
+ print('##############'*30)
+ with sess:
+ nets.print_params()
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_RBN.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_RBN.py
new file mode 100644
index 0000000000000000000000000000000000000000..d6a54a4600e30fc286ef2219314ba4651e375807
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_RBN.py
@@ -0,0 +1,322 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+from tensorflow.contrib.layers.python.layers import utils
+import collections
+from tensorlayer.layers import Layer, list_remove_repeat
+
+
+class ElementwiseLayer(Layer):
+ """
+ The :class:`ElementwiseLayer` class combines multiple :class:`Layer` which have the same output shapes by a given elemwise-wise operation.
+
+ Parameters
+ ----------
+ layer : a list of :class:`Layer` instances
+ The `Layer` class feeding into this layer.
+ combine_fn : a TensorFlow elemwise-merge function
+ e.g. AND is ``tf.minimum`` ; OR is ``tf.maximum`` ; ADD is ``tf.add`` ; MUL is ``tf.multiply`` and so on.
+ See `TensorFlow Math API `_ .
+ name : a string or None
+ An optional name to attach to this layer.
+ """
+ def __init__(
+ self,
+ layer = [],
+ combine_fn = tf.minimum,
+ name ='elementwise_layer',
+ act = None,
+ ):
+ Layer.__init__(self, name=name)
+
+ if act:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s, act:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__, act.__name__))
+ else:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__))
+
+ self.outputs = layer[0].outputs
+ # print(self.outputs._shape, type(self.outputs._shape))
+ for l in layer[1:]:
+ # assert str(self.outputs.get_shape()) == str(l.outputs.get_shape()), "Hint: the input shapes should be the same. %s != %s" % (self.outputs.get_shape() , str(l.outputs.get_shape()))
+ self.outputs = combine_fn(self.outputs, l.outputs, name=name)
+ if act:
+ self.outputs = act(self.outputs)
+ self.all_layers = list(layer[0].all_layers)
+ self.all_params = list(layer[0].all_params)
+ self.all_drop = dict(layer[0].all_drop)
+
+ for i in range(1, len(layer)):
+ self.all_layers.extend(list(layer[i].all_layers))
+ self.all_params.extend(list(layer[i].all_params))
+ self.all_drop.update(dict(layer[i].all_drop))
+
+ self.all_layers = list_remove_repeat(self.all_layers)
+ self.all_params = list_remove_repeat(self.all_params)
+
+
+def subsample(inputs, factor, scope=None):
+ if factor == 1:
+ return inputs
+ else:
+ return tl.layers.MaxPool2d(inputs, [1, 1], strides=(factor, factor), name=scope)
+
+
+def conv2d_same(inputs, num_outputs, kernel_size, strides, rate=1, w_init=None, scope=None, trainable=None):
+ '''
+ Reference slim resnet
+ :param inputs:
+ :param num_outputs:
+ :param kernel_size:
+ :param strides:
+ :param rate:
+ :param scope:
+ :return:
+ '''
+ if strides == 1:
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='SAME', name=scope,
+ use_cudnn_on_gpu=True)
+ nets.outputs = tf.layers.batch_normalization(inputs=nets.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax':3, 'rmin':0.3333,
+ 'dmax':5},
+ renorm_momentum=0.9,
+ name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size),
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets.outputs = tf.layers.batch_normalization(inputs=nets.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax':3, 'rmin':0.3333,
+ 'dmax':5},
+ renorm_momentum=0.9,
+ name=scope+'_bn/BatchNorm')
+ return nets
+ else:
+ kernel_size_effective = kernel_size + (kernel_size - 1) * (rate - 1)
+ pad_total = kernel_size_effective - 1
+ pad_beg = pad_total // 2
+ pad_end = pad_total - pad_beg
+ inputs = tl.layers.PadLayer(inputs, [[0, 0], [pad_beg, pad_end], [pad_beg, pad_end], [0, 0]], name='padding_%s' % scope)
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='VALID', name=scope,
+ use_cudnn_on_gpu=True)
+ nets.outputs = tf.layers.batch_normalization(inputs=nets.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax':3, 'rmin':0.3333,
+ 'dmax':5},
+ renorm_momentum=0.9,
+ name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets.outputs = tf.layers.batch_normalization(inputs=nets.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax':3, 'rmin':0.3333,
+ 'dmax':5},
+ renorm_momentum=0.9,
+ name=scope+'_bn/BatchNorm')
+ return nets
+
+
+def bottleneck_IR(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut.outputs = tf.layers.batch_normalization(inputs=shortcut.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax': 3, 'rmin': 0.3333,
+ 'dmax': 5},
+ renorm_momentum=0.9,
+ name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ inputs.outputs = tf.layers.batch_normalization(inputs=inputs.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax': 3, 'rmin': 0.3333,
+ 'dmax': 5},
+ renorm_momentum=0.9,
+ name='conv1_bn1')
+ residual = tl.layers.Conv2d(inputs, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual.outputs = tf.layers.batch_normalization(inputs=residual.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax': 3, 'rmin': 0.3333,
+ 'dmax': 5},
+ renorm_momentum=0.9,
+ name='conv1_bn2')
+ # bottleneck prelu
+ residual = tl.layers.PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ output = ElementwiseLayer(layer=[shortcut, residual],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=None)
+ return output
+
+
+def resnet(inputs, bottle_neck, blocks, w_init=None, trainable=None, scope=None):
+ with tf.variable_scope(scope):
+ # inputs = tf.subtract(inputs, 127.5)
+ # inputs = tf.multiply(inputs, 0.0078125)
+ net_inputs = tl.layers.InputLayer(inputs, name='input_layer')
+ if bottle_neck:
+ net = tl.layers.Conv2d(net_inputs, n_filter=64, filter_size=(3, 3), strides=(1, 1),
+ act=None, W_init=w_init, b_init=None, name='conv1', use_cudnn_on_gpu=True)
+ net.outputs = tf.layers.batch_normalization(inputs=net.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax': 3, 'rmin': 0.3333,
+ 'dmax': 5},
+ renorm_momentum=0.9,
+ name='bn0')
+ net = tl.layers.PReluLayer(net, name='prelu0')
+ else:
+ raise ValueError('The standard resnet must support the bottleneck layer')
+ for block in blocks:
+ with tf.variable_scope(block.scope):
+ for i, var in enumerate(block.args):
+ with tf.variable_scope('unit_%d' % (i+1)):
+ net = block.unit_fn(net, depth=var['depth'], depth_bottleneck=var['depth_bottleneck'],
+ w_init=w_init, stride=var['stride'], rate=var['rate'], scope=None,
+ trainable=trainable)
+ net.outputs = tf.layers.batch_normalization(inputs=net.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax': 3, 'rmin': 0.3333,
+ 'dmax': 5},
+ renorm_momentum=0.9,
+ name='E_BN1')
+ net = tl.layers.DropoutLayer(net, keep=0.4, name='E_Dropout')
+ net_shape = net.outputs.get_shape()
+ net = tl.layers.ReshapeLayer(net, shape=[-1, net_shape[1]*net_shape[2]*net_shape[3]], name='E_Reshapelayer')
+ net = tl.layers.DenseLayer(net, n_units=512, W_init=w_init, name='E_DenseLayer')
+ net.outputs = tf.layers.batch_normalization(inputs=net.outputs,
+ momentum=0.9,
+ training=trainable,
+ renorm=True,
+ renorm_clipping={'rmax': 3, 'rmin': 0.3333,
+ 'dmax': 5},
+ renorm_momentum=0.9,
+ name='E_BN2')
+ return net
+
+
+class Block(collections.namedtuple('Block', ['scope', 'unit_fn', 'args'])):
+ """A named tuple describing a ResNet block.
+
+ Its parts are:
+ scope: The scope of the `Block`.
+ unit_fn: The ResNet unit function which takes as input a `Tensor` and
+ returns another `Tensor` with the output of the ResNet unit.
+ args: A list of length equal to the number of units in the `Block`. The list
+ contains one (depth, depth_bottleneck, stride) tuple for each unit in the
+ block to serve as argument to unit_fn.
+ """
+
+
+def resnetse_v1_block(scope, base_depth, num_units, stride, rate=1, unit_fn=None):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, unit_fn, [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': stride,
+ 'rate': rate
+ }] + [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': 1,
+ 'rate': rate
+ }] * (num_units - 1))
+
+
+def get_resnet(inputs, num_layers, type=None, w_init=None, trainable=None, sess=None):
+ if type == 'ir':
+ unit_fn = bottleneck_IR
+ # elif type == 'se_ir':
+ # unit_fn = bottleneck_IR_SE
+ else:
+ raise ValueError('the input fn is unknown')
+
+ if num_layers == 50:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=14, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 101:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=13, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=30, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 152:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=8, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=36, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ else:
+ raise ValueError('Resnet layer %d is not supported now.' % num_layers)
+ net = resnet(inputs=inputs,
+ bottle_neck=True,
+ blocks=blocks,
+ w_init=w_init,
+ trainable=trainable,
+ scope='resnet_v1_%d' % num_layers)
+ return net
+
+
+if __name__ == '__main__':
+ x = tf.placeholder(dtype=tf.float32, shape=[None, 112, 112, 3], name='input_place')
+ sess = tf.Session(config=npu_config_proto())
+ # w_init = tf.truncated_normal_initializer(mean=10, stddev=5e-2)
+ w_init = tf.contrib.layers.xavier_initializer(uniform=False)
+ # test resnetse
+ nets = get_resnet(x, 50, type='ir', w_init=w_init, sess=sess)
+ tl.layers.initialize_global_variables(sess)
+
+ for p in tl.layers.get_variables_with_name('W_conv2d', True, True):
+ print(p.op.name)
+ print('##############'*30)
+ with sess:
+ nets.print_params()
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_fix_issue9.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_fix_issue9.py
new file mode 100644
index 0000000000000000000000000000000000000000..d7b67ca444157cf6e041651f2ebe796b45cf7052
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/L_Resnet_E_IR_fix_issue9.py
@@ -0,0 +1,418 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+from tensorflow.contrib.layers.python.layers import utils
+import collections
+from tensorlayer.layers import Layer, list_remove_repeat
+
+
+class ElementwiseLayer(Layer):
+ """
+ The :class:`ElementwiseLayer` class combines multiple :class:`Layer` which have the same output shapes by a given elemwise-wise operation.
+
+ Parameters
+ ----------
+ layer : a list of :class:`Layer` instances
+ The `Layer` class feeding into this layer.
+ combine_fn : a TensorFlow elemwise-merge function
+ e.g. AND is ``tf.minimum`` ; OR is ``tf.maximum`` ; ADD is ``tf.add`` ; MUL is ``tf.multiply`` and so on.
+ See `TensorFlow Math API `_ .
+ name : a string or None
+ An optional name to attach to this layer.
+ """
+ def __init__(
+ self,
+ layer = [],
+ combine_fn = tf.minimum,
+ name ='elementwise_layer',
+ act = None,
+ ):
+ Layer.__init__(self, name=name)
+
+ if act:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s, act:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__, act.__name__))
+ else:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__))
+
+ self.outputs = layer[0].outputs
+ # print(self.outputs._shape, type(self.outputs._shape))
+ for l in layer[1:]:
+ # assert str(self.outputs.get_shape()) == str(l.outputs.get_shape()), "Hint: the input shapes should be the same. %s != %s" % (self.outputs.get_shape() , str(l.outputs.get_shape()))
+ self.outputs = combine_fn(self.outputs, l.outputs, name=name)
+ if act:
+ self.outputs = act(self.outputs)
+ self.all_layers = list(layer[0].all_layers)
+ self.all_params = list(layer[0].all_params)
+ self.all_drop = dict(layer[0].all_drop)
+
+ for i in range(1, len(layer)):
+ self.all_layers.extend(list(layer[i].all_layers))
+ self.all_params.extend(list(layer[i].all_params))
+ self.all_drop.update(dict(layer[i].all_drop))
+
+ self.all_layers = list_remove_repeat(self.all_layers)
+ self.all_params = list_remove_repeat(self.all_params)
+
+
+class BatchNormLayer(Layer):
+ """
+ The :class:`BatchNormLayer` class is a normalization layer, see ``tf.nn.batch_normalization`` and ``tf.nn.moments``.
+
+ Batch normalization on fully-connected or convolutional maps.
+
+ ```
+ https://www.tensorflow.org/api_docs/python/tf/cond
+ If x < y, the tf.add operation will be executed and tf.square operation will not be executed.
+ Since z is needed for at least one branch of the cond, the tf.multiply operation is always executed, unconditionally.
+ ```
+
+ Parameters
+ -----------
+ layer : a :class:`Layer` instance
+ The `Layer` class feeding into this layer.
+ decay : float, default is 0.9.
+ A decay factor for ExponentialMovingAverage, use larger value for large dataset.
+ epsilon : float
+ A small float number to avoid dividing by 0.
+ act : activation function.
+ is_train : boolean
+ Whether train or inference.
+ beta_init : beta initializer
+ The initializer for initializing beta
+ gamma_init : gamma initializer
+ The initializer for initializing gamma
+ dtype : tf.float32 (default) or tf.float16
+ name : a string or None
+ An optional name to attach to this layer.
+
+ References
+ ----------
+ - `Source `_
+ - `stackoverflow `_
+
+ """
+
+ def __init__(
+ self,
+ layer=None,
+ decay=0.9,
+ epsilon=2e-5,
+ act=tf.identity,
+ is_train=False,
+ fix_gamma=True,
+ beta_init=tf.zeros_initializer,
+ gamma_init=tf.random_normal_initializer(mean=1.0, stddev=0.002), # tf.ones_initializer,
+ # dtype = tf.float32,
+ trainable=None,
+ name='batchnorm_layer',
+ ):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ print(" [TL] BatchNormLayer %s: decay:%f epsilon:%f act:%s is_train:%s" % (self.name, decay, epsilon, act.__name__, is_train))
+ x_shape = self.inputs.get_shape()
+ params_shape = x_shape[-1:]
+
+ from tensorflow.python.training import moving_averages
+ from tensorflow.python.ops import control_flow_ops
+
+ with tf.variable_scope(name) as vs:
+ axis = list(range(len(x_shape) - 1))
+
+ ## 1. beta, gamma
+ if tf.__version__ > '0.12.1' and beta_init == tf.zeros_initializer:
+ beta_init = beta_init()
+ beta = tf.get_variable('beta', shape=params_shape, initializer=beta_init, dtype=tf.float32, trainable=is_train) #, restore=restore)
+
+ gamma = tf.get_variable(
+ 'gamma',
+ shape=params_shape,
+ initializer=gamma_init,
+ dtype=tf.float32,
+ trainable=fix_gamma,
+ ) #restore=restore)
+
+ ## 2.
+ if tf.__version__ > '0.12.1':
+ moving_mean_init = tf.zeros_initializer()
+ else:
+ moving_mean_init = tf.zeros_initializer
+ moving_mean = tf.get_variable('moving_mean', params_shape, initializer=moving_mean_init, dtype=tf.float32, trainable=False) # restore=restore)
+ moving_variance = tf.get_variable(
+ 'moving_variance',
+ params_shape,
+ initializer=tf.constant_initializer(1.),
+ dtype=tf.float32,
+ trainable=False,
+ ) # restore=restore)
+
+ ## 3.
+ # These ops will only be preformed when training.
+ mean, variance = tf.nn.moments(self.inputs, axis)
+ try: # TF12
+ update_moving_mean = moving_averages.assign_moving_average(moving_mean, mean, decay, zero_debias=False) # if zero_debias=True, has bias
+ update_moving_variance = moving_averages.assign_moving_average(
+ moving_variance, variance, decay, zero_debias=False) # if zero_debias=True, has bias
+ # print("TF12 moving")
+ except Exception as e: # TF11
+ update_moving_mean = moving_averages.assign_moving_average(moving_mean, mean, decay)
+ update_moving_variance = moving_averages.assign_moving_average(moving_variance, variance, decay)
+ # print("TF11 moving")
+
+ def mean_var_with_update():
+ with tf.control_dependencies([update_moving_mean, update_moving_variance]):
+ return tf.identity(mean), tf.identity(variance)
+ if trainable:
+ mean, var = mean_var_with_update()
+ print(mean)
+ print(var)
+ self.outputs = act(tf.nn.batch_normalization(self.inputs, mean, var, beta, gamma, epsilon))
+ else:
+ self.outputs = act(tf.nn.batch_normalization(self.inputs, moving_mean, moving_variance, beta, gamma, epsilon))
+ variables = [beta, gamma, moving_mean, moving_variance]
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+ self.all_layers.extend([self.outputs])
+ self.all_params.extend(variables)
+
+
+def subsample(inputs, factor, scope=None):
+ if factor == 1:
+ return inputs
+ else:
+ return tl.layers.MaxPool2d(inputs, [1, 1], strides=(factor, factor), name=scope)
+
+
+def conv2d_same(inputs, num_outputs, kernel_size, strides, rate=1, w_init=None, scope=None, trainable=None):
+ '''
+ Reference slim resnet
+ :param inputs:
+ :param num_outputs:
+ :param kernel_size:
+ :param strides:
+ :param rate:
+ :param scope:
+ :return:
+ '''
+ if strides == 1:
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='SAME', name=scope,
+ use_cudnn_on_gpu=True)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size),
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ return nets
+ else:
+ kernel_size_effective = kernel_size + (kernel_size - 1) * (rate - 1)
+ pad_total = kernel_size_effective - 1
+ pad_beg = pad_total // 2
+ pad_end = pad_total - pad_beg
+ inputs = tl.layers.PadLayer(inputs, [[0, 0], [pad_beg, pad_end], [pad_beg, pad_end], [0, 0]], name='padding_%s' % scope)
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), W_init=w_init, act=None, padding='VALID', name=scope,
+ use_cudnn_on_gpu=True)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ rate=rate, act=None, W_init=w_init, padding='SAME', name=scope)
+ nets = BatchNormLayer(nets, act=tf.identity, is_train=True, trainable=trainable, name=scope+'_bn/BatchNorm')
+ return nets
+
+
+def bottleneck_IR(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut = BatchNormLayer(shortcut, act=tf.identity, is_train=True, trainable=trainable, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = BatchNormLayer(inputs, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn1')
+ residual = tl.layers.Conv2d(residual, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual = BatchNormLayer(residual, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn2')
+ # bottleneck prelu
+ residual = tl.layers.PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ output = ElementwiseLayer(layer=[shortcut, residual],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=None)
+ return output
+
+
+def bottleneck_IR_SE(inputs, depth, depth_bottleneck, stride, rate=1, w_init=None, scope=None, trainable=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ W_init=w_init, b_init=None, name='shortcut_conv', use_cudnn_on_gpu=True)
+ shortcut = BatchNormLayer(shortcut, act=tf.identity, is_train=True, trainable=trainable, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = BatchNormLayer(inputs, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn1')
+ residual = tl.layers.Conv2d(residual, depth_bottleneck, filter_size=(3, 3), strides=(1, 1), act=None, b_init=None,
+ W_init=w_init, name='conv1', use_cudnn_on_gpu=True)
+ residual = BatchNormLayer(residual, act=tf.identity, is_train=True, trainable=trainable, name='conv1_bn2')
+ # bottleneck prelu
+ residual = tl.layers.PReluLayer(residual)
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth, kernel_size=3, strides=stride, rate=rate, w_init=w_init, scope='conv2', trainable=trainable)
+ # squeeze
+ squeeze = tl.layers.InputLayer(tf.reduce_mean(residual.outputs, axis=[1, 2]), name='squeeze_layer')
+ # excitation
+ excitation1 = tl.layers.DenseLayer(squeeze, n_units=int(depth/16.0), act=tf.nn.relu,
+ W_init=w_init, name='excitation_1')
+ # excitation1 = tl.layers.PReluLayer(excitation1, name='excitation_prelu')
+ excitation2 = tl.layers.DenseLayer(excitation1, n_units=depth, act=tf.nn.sigmoid,
+ W_init=w_init, name='excitation_2')
+ # scale
+ scale = tl.layers.ReshapeLayer(excitation2, shape=[tf.shape(excitation2.outputs)[0], 1, 1, depth], name='excitation_reshape')
+
+ residual_se = ElementwiseLayer(layer=[residual, scale],
+ combine_fn=tf.multiply,
+ name='scale_layer',
+ act=None)
+
+ output = ElementwiseLayer(layer=[shortcut, residual_se],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=tf.nn.relu)
+ return output
+
+
+def resnet(inputs, bottle_neck, blocks, w_init=None, trainable=None, reuse=False, keep_rate=None, scope=None):
+ with tf.variable_scope(scope, reuse=reuse):
+ # inputs = tf.subtract(inputs, 127.5)
+ # inputs = tf.multiply(inputs, 0.0078125)
+ net_inputs = tl.layers.InputLayer(inputs, name='input_layer')
+ if bottle_neck:
+ net = tl.layers.Conv2d(net_inputs, n_filter=64, filter_size=(3, 3), strides=(1, 1),
+ act=None, W_init=w_init, b_init=None, name='conv1', use_cudnn_on_gpu=True)
+ net = BatchNormLayer(net, act=tf.identity, name='bn0', is_train=True, trainable=trainable)
+ net = tl.layers.PReluLayer(net, name='prelu0')
+ else:
+ raise ValueError('The standard resnet must support the bottleneck layer')
+ for block in blocks:
+ with tf.variable_scope(block.scope):
+ for i, var in enumerate(block.args):
+ with tf.variable_scope('unit_%d' % (i+1)):
+ net = block.unit_fn(net, depth=var['depth'], depth_bottleneck=var['depth_bottleneck'],
+ w_init=w_init, stride=var['stride'], rate=var['rate'], scope=None,
+ trainable=trainable)
+ net = BatchNormLayer(net, act=tf.identity, is_train=True, name='E_BN1', trainable=trainable)
+ # net = tl.layers.DropoutLayer(net, keep=0.4, name='E_Dropout')
+ net.outputs = npu_ops.dropout(net.outputs, keep_prob=keep_rate, name='E_Dropout')
+ net_shape = net.outputs.get_shape()
+ net = tl.layers.ReshapeLayer(net, shape=[-1, net_shape[1]*net_shape[2]*net_shape[3]], name='E_Reshapelayer')
+ net = tl.layers.DenseLayer(net, n_units=512, W_init=w_init, name='E_DenseLayer')
+ net = BatchNormLayer(net, act=tf.identity, is_train=True, fix_gamma=False, trainable=trainable, name='E_BN2')
+ return net
+
+
+class Block(collections.namedtuple('Block', ['scope', 'unit_fn', 'args'])):
+ """A named tuple describing a ResNet block.
+
+ Its parts are:
+ scope: The scope of the `Block`.
+ unit_fn: The ResNet unit function which takes as input a `Tensor` and
+ returns another `Tensor` with the output of the ResNet unit.
+ args: A list of length equal to the number of units in the `Block`. The list
+ contains one (depth, depth_bottleneck, stride) tuple for each unit in the
+ block to serve as argument to unit_fn.
+ """
+
+
+def resnetse_v1_block(scope, base_depth, num_units, stride, rate=1, unit_fn=None):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, unit_fn, [{
+ 'depth': base_depth,
+ 'depth_bottleneck': base_depth,
+ 'stride': stride,
+ 'rate': rate
+ }] + [{
+ 'depth': base_depth,
+ 'depth_bottleneck': base_depth,
+ 'stride': 1,
+ 'rate': rate
+ }] * (num_units - 1))
+
+
+def get_resnet(inputs, num_layers, type=None, w_init=None, trainable=None, sess=None, reuse=False, keep_rate=None):
+ if type == 'ir':
+ unit_fn = bottleneck_IR
+ elif type == 'se_ir':
+ unit_fn = bottleneck_IR_SE
+ else:
+ raise ValueError('the input fn is unknown')
+
+ if num_layers == 50:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=14, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 100:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=13, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=30, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ elif num_layers == 152:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block2', base_depth=128, num_units=8, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block3', base_depth=256, num_units=36, stride=2, rate=1, unit_fn=unit_fn),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=2, rate=1, unit_fn=unit_fn)
+ ]
+ else:
+ raise ValueError('Resnet layer %d is not supported now.' % num_layers)
+ net = resnet(inputs=inputs,
+ bottle_neck=True,
+ blocks=blocks,
+ w_init=w_init,
+ trainable=trainable,
+ reuse=reuse,
+ keep_rate=keep_rate,
+ scope='resnet_v1_%d' % num_layers)
+ return net
+
+
+if __name__ == '__main__':
+ x = tf.placeholder(dtype=tf.float32, shape=[None, 112, 112, 3], name='input_place')
+ sess = tf.Session(config=npu_config_proto())
+ # w_init = tf.truncated_normal_initializer(mean=10, stddev=5e-2)
+ w_init = tf.contrib.layers.xavier_initializer(uniform=False)
+ # test resnetse
+ nets = get_resnet(x, 50, type='ir', w_init=w_init, sess=sess)
+ tl.layers.initialize_global_variables(sess)
+
+ for p in tl.layers.get_variables_with_name('W_conv2d', True, True):
+ print(p.op.name)
+ print('##############'*30)
+ with sess:
+ nets.print_params()
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/__init__.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/imagenet_classes.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/imagenet_classes.py
new file mode 100644
index 0000000000000000000000000000000000000000..d7d7ab7ec1b2f691191f20477a12386cf9602212
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/imagenet_classes.py
@@ -0,0 +1,1000 @@
+class_names = '''tench, Tinca tinca
+goldfish, Carassius auratus
+great white shark, white shark, man-eater, man-eating shark, Carcharodon carcharias
+tiger shark, Galeocerdo cuvieri
+hammerhead, hammerhead shark
+electric ray, crampfish, numbfish, torpedo
+stingray
+cock
+hen
+ostrich, Struthio camelus
+brambling, Fringilla montifringilla
+goldfinch, Carduelis carduelis
+house finch, linnet, Carpodacus mexicanus
+junco, snowbird
+indigo bunting, indigo finch, indigo bird, Passerina cyanea
+robin, American robin, Turdus migratorius
+bulbul
+jay
+magpie
+chickadee
+water ouzel, dipper
+kite
+bald eagle, American eagle, Haliaeetus leucocephalus
+vulture
+great grey owl, great gray owl, Strix nebulosa
+European fire salamander, Salamandra salamandra
+common newt, Triturus vulgaris
+eft
+spotted salamander, Ambystoma maculatum
+axolotl, mud puppy, Ambystoma mexicanum
+bullfrog, Rana catesbeiana
+tree frog, tree-frog
+tailed frog, bell toad, ribbed toad, tailed toad, Ascaphus trui
+loggerhead, loggerhead turtle, Caretta caretta
+leatherback turtle, leatherback, leathery turtle, Dermochelys coriacea
+mud turtle
+terrapin
+box turtle, box tortoise
+banded gecko
+common iguana, iguana, Iguana iguana
+American chameleon, anole, Anolis carolinensis
+whiptail, whiptail lizard
+agama
+frilled lizard, Chlamydosaurus kingi
+alligator lizard
+Gila monster, Heloderma suspectum
+green lizard, Lacerta viridis
+African chameleon, Chamaeleo chamaeleon
+Komodo dragon, Komodo lizard, dragon lizard, giant lizard, Varanus komodoensis
+African crocodile, Nile crocodile, Crocodylus niloticus
+American alligator, Alligator mississipiensis
+triceratops
+thunder snake, worm snake, Carphophis amoenus
+ringneck snake, ring-necked snake, ring snake
+hognose snake, puff adder, sand viper
+green snake, grass snake
+king snake, kingsnake
+garter snake, grass snake
+water snake
+vine snake
+night snake, Hypsiglena torquata
+boa constrictor, Constrictor constrictor
+rock python, rock snake, Python sebae
+Indian cobra, Naja naja
+green mamba
+sea snake
+horned viper, cerastes, sand viper, horned asp, Cerastes cornutus
+diamondback, diamondback rattlesnake, Crotalus adamanteus
+sidewinder, horned rattlesnake, Crotalus cerastes
+trilobite
+harvestman, daddy longlegs, Phalangium opilio
+scorpion
+black and gold garden spider, Argiope aurantia
+barn spider, Araneus cavaticus
+garden spider, Aranea diademata
+black widow, Latrodectus mactans
+tarantula
+wolf spider, hunting spider
+tick
+centipede
+black grouse
+ptarmigan
+ruffed grouse, partridge, Bonasa umbellus
+prairie chicken, prairie grouse, prairie fowl
+peacock
+quail
+partridge
+African grey, African gray, Psittacus erithacus
+macaw
+sulphur-crested cockatoo, Kakatoe galerita, Cacatua galerita
+lorikeet
+coucal
+bee eater
+hornbill
+hummingbird
+jacamar
+toucan
+drake
+red-breasted merganser, Mergus serrator
+goose
+black swan, Cygnus atratus
+tusker
+echidna, spiny anteater, anteater
+platypus, duckbill, duckbilled platypus, duck-billed platypus, Ornithorhynchus anatinus
+wallaby, brush kangaroo
+koala, koala bear, kangaroo bear, native bear, Phascolarctos cinereus
+wombat
+jellyfish
+sea anemone, anemone
+brain coral
+flatworm, platyhelminth
+nematode, nematode worm, roundworm
+conch
+snail
+slug
+sea slug, nudibranch
+chiton, coat-of-mail shell, sea cradle, polyplacophore
+chambered nautilus, pearly nautilus, nautilus
+Dungeness crab, Cancer magister
+rock crab, Cancer irroratus
+fiddler crab
+king crab, Alaska crab, Alaskan king crab, Alaska king crab, Paralithodes camtschatica
+American lobster, Northern lobster, Maine lobster, Homarus americanus
+spiny lobster, langouste, rock lobster, crawfish, crayfish, sea crawfish
+crayfish, crawfish, crawdad, crawdaddy
+hermit crab
+isopod
+white stork, Ciconia ciconia
+black stork, Ciconia nigra
+spoonbill
+flamingo
+little blue heron, Egretta caerulea
+American egret, great white heron, Egretta albus
+bittern
+crane
+limpkin, Aramus pictus
+European gallinule, Porphyrio porphyrio
+American coot, marsh hen, mud hen, water hen, Fulica americana
+bustard
+ruddy turnstone, Arenaria interpres
+red-backed sandpiper, dunlin, Erolia alpina
+redshank, Tringa totanus
+dowitcher
+oystercatcher, oyster catcher
+pelican
+king penguin, Aptenodytes patagonica
+albatross, mollymawk
+grey whale, gray whale, devilfish, Eschrichtius gibbosus, Eschrichtius robustus
+killer whale, killer, orca, grampus, sea wolf, Orcinus orca
+dugong, Dugong dugon
+sea lion
+Chihuahua
+Japanese spaniel
+Maltese dog, Maltese terrier, Maltese
+Pekinese, Pekingese, Peke
+Shih-Tzu
+Blenheim spaniel
+papillon
+toy terrier
+Rhodesian ridgeback
+Afghan hound, Afghan
+basset, basset hound
+beagle
+bloodhound, sleuthhound
+bluetick
+black-and-tan coonhound
+Walker hound, Walker foxhound
+English foxhound
+redbone
+borzoi, Russian wolfhound
+Irish wolfhound
+Italian greyhound
+whippet
+Ibizan hound, Ibizan Podenco
+Norwegian elkhound, elkhound
+otterhound, otter hound
+Saluki, gazelle hound
+Scottish deerhound, deerhound
+Weimaraner
+Staffordshire bullterrier, Staffordshire bull terrier
+American Staffordshire terrier, Staffordshire terrier, American pit bull terrier, pit bull terrier
+Bedlington terrier
+Border terrier
+Kerry blue terrier
+Irish terrier
+Norfolk terrier
+Norwich terrier
+Yorkshire terrier
+wire-haired fox terrier
+Lakeland terrier
+Sealyham terrier, Sealyham
+Airedale, Airedale terrier
+cairn, cairn terrier
+Australian terrier
+Dandie Dinmont, Dandie Dinmont terrier
+Boston bull, Boston terrier
+miniature schnauzer
+giant schnauzer
+standard schnauzer
+Scotch terrier, Scottish terrier, Scottie
+Tibetan terrier, chrysanthemum dog
+silky terrier, Sydney silky
+soft-coated wheaten terrier
+West Highland white terrier
+Lhasa, Lhasa apso
+flat-coated retriever
+curly-coated retriever
+golden retriever
+Labrador retriever
+Chesapeake Bay retriever
+German short-haired pointer
+vizsla, Hungarian pointer
+English setter
+Irish setter, red setter
+Gordon setter
+Brittany spaniel
+clumber, clumber spaniel
+English springer, English springer spaniel
+Welsh springer spaniel
+cocker spaniel, English cocker spaniel, cocker
+Sussex spaniel
+Irish water spaniel
+kuvasz
+schipperke
+groenendael
+malinois
+briard
+kelpie
+komondor
+Old English sheepdog, bobtail
+Shetland sheepdog, Shetland sheep dog, Shetland
+collie
+Border collie
+Bouvier des Flandres, Bouviers des Flandres
+Rottweiler
+German shepherd, German shepherd dog, German police dog, alsatian
+Doberman, Doberman pinscher
+miniature pinscher
+Greater Swiss Mountain dog
+Bernese mountain dog
+Appenzeller
+EntleBucher
+boxer
+bull mastiff
+Tibetan mastiff
+French bulldog
+Great Dane
+Saint Bernard, St Bernard
+Eskimo dog, husky
+malamute, malemute, Alaskan malamute
+Siberian husky
+dalmatian, coach dog, carriage dog
+affenpinscher, monkey pinscher, monkey dog
+basenji
+pug, pug-dog
+Leonberg
+Newfoundland, Newfoundland dog
+Great Pyrenees
+Samoyed, Samoyede
+Pomeranian
+chow, chow chow
+keeshond
+Brabancon griffon
+Pembroke, Pembroke Welsh corgi
+Cardigan, Cardigan Welsh corgi
+toy poodle
+miniature poodle
+standard poodle
+Mexican hairless
+timber wolf, grey wolf, gray wolf, Canis lupus
+white wolf, Arctic wolf, Canis lupus tundrarum
+red wolf, maned wolf, Canis rufus, Canis niger
+coyote, prairie wolf, brush wolf, Canis latrans
+dingo, warrigal, warragal, Canis dingo
+dhole, Cuon alpinus
+African hunting dog, hyena dog, Cape hunting dog, Lycaon pictus
+hyena, hyaena
+red fox, Vulpes vulpes
+kit fox, Vulpes macrotis
+Arctic fox, white fox, Alopex lagopus
+grey fox, gray fox, Urocyon cinereoargenteus
+tabby, tabby cat
+tiger cat
+Persian cat
+Siamese cat, Siamese
+Egyptian cat
+cougar, puma, catamount, mountain lion, painter, panther, Felis concolor
+lynx, catamount
+leopard, Panthera pardus
+snow leopard, ounce, Panthera uncia
+jaguar, panther, Panthera onca, Felis onca
+lion, king of beasts, Panthera leo
+tiger, Panthera tigris
+cheetah, chetah, Acinonyx jubatus
+brown bear, bruin, Ursus arctos
+American black bear, black bear, Ursus americanus, Euarctos americanus
+ice bear, polar bear, Ursus Maritimus, Thalarctos maritimus
+sloth bear, Melursus ursinus, Ursus ursinus
+mongoose
+meerkat, mierkat
+tiger beetle
+ladybug, ladybeetle, lady beetle, ladybird, ladybird beetle
+ground beetle, carabid beetle
+long-horned beetle, longicorn, longicorn beetle
+leaf beetle, chrysomelid
+dung beetle
+rhinoceros beetle
+weevil
+fly
+bee
+ant, emmet, pismire
+grasshopper, hopper
+cricket
+walking stick, walkingstick, stick insect
+cockroach, roach
+mantis, mantid
+cicada, cicala
+leafhopper
+lacewing, lacewing fly
+dragonfly, darning needle, devil's darning needle, sewing needle, snake feeder, snake doctor, mosquito hawk, skeeter hawk
+damselfly
+admiral
+ringlet, ringlet butterfly
+monarch, monarch butterfly, milkweed butterfly, Danaus plexippus
+cabbage butterfly
+sulphur butterfly, sulfur butterfly
+lycaenid, lycaenid butterfly
+starfish, sea star
+sea urchin
+sea cucumber, holothurian
+wood rabbit, cottontail, cottontail rabbit
+hare
+Angora, Angora rabbit
+hamster
+porcupine, hedgehog
+fox squirrel, eastern fox squirrel, Sciurus niger
+marmot
+beaver
+guinea pig, Cavia cobaya
+sorrel
+zebra
+hog, pig, grunter, squealer, Sus scrofa
+wild boar, boar, Sus scrofa
+warthog
+hippopotamus, hippo, river horse, Hippopotamus amphibius
+ox
+water buffalo, water ox, Asiatic buffalo, Bubalus bubalis
+bison
+ram, tup
+bighorn, bighorn sheep, cimarron, Rocky Mountain bighorn, Rocky Mountain sheep, Ovis canadensis
+ibex, Capra ibex
+hartebeest
+impala, Aepyceros melampus
+gazelle
+Arabian camel, dromedary, Camelus dromedarius
+llama
+weasel
+mink
+polecat, fitch, foulmart, foumart, Mustela putorius
+black-footed ferret, ferret, Mustela nigripes
+otter
+skunk, polecat, wood pussy
+badger
+armadillo
+three-toed sloth, ai, Bradypus tridactylus
+orangutan, orang, orangutang, Pongo pygmaeus
+gorilla, Gorilla gorilla
+chimpanzee, chimp, Pan troglodytes
+gibbon, Hylobates lar
+siamang, Hylobates syndactylus, Symphalangus syndactylus
+guenon, guenon monkey
+patas, hussar monkey, Erythrocebus patas
+baboon
+macaque
+langur
+colobus, colobus monkey
+proboscis monkey, Nasalis larvatus
+marmoset
+capuchin, ringtail, Cebus capucinus
+howler monkey, howler
+titi, titi monkey
+spider monkey, Ateles geoffroyi
+squirrel monkey, Saimiri sciureus
+Madagascar cat, ring-tailed lemur, Lemur catta
+indri, indris, Indri indri, Indri brevicaudatus
+Indian elephant, Elephas maximus
+African elephant, Loxodonta africana
+lesser panda, red panda, panda, bear cat, cat bear, Ailurus fulgens
+giant panda, panda, panda bear, coon bear, Ailuropoda melanoleuca
+barracouta, snoek
+eel
+coho, cohoe, coho salmon, blue jack, silver salmon, Oncorhynchus kisutch
+rock beauty, Holocanthus tricolor
+anemone fish
+sturgeon
+gar, garfish, garpike, billfish, Lepisosteus osseus
+lionfish
+puffer, pufferfish, blowfish, globefish
+abacus
+abaya
+academic gown, academic robe, judge's robe
+accordion, piano accordion, squeeze box
+acoustic guitar
+aircraft carrier, carrier, flattop, attack aircraft carrier
+airliner
+airship, dirigible
+altar
+ambulance
+amphibian, amphibious vehicle
+analog clock
+apiary, bee house
+apron
+ashcan, trash can, garbage can, wastebin, ash bin, ash-bin, ashbin, dustbin, trash barrel, trash bin
+assault rifle, assault gun
+backpack, back pack, knapsack, packsack, rucksack, haversack
+bakery, bakeshop, bakehouse
+balance beam, beam
+balloon
+ballpoint, ballpoint pen, ballpen, Biro
+Band Aid
+banjo
+bannister, banister, balustrade, balusters, handrail
+barbell
+barber chair
+barbershop
+barn
+barometer
+barrel, cask
+barrow, garden cart, lawn cart, wheelbarrow
+baseball
+basketball
+bassinet
+bassoon
+bathing cap, swimming cap
+bath towel
+bathtub, bathing tub, bath, tub
+beach wagon, station wagon, wagon, estate car, beach waggon, station waggon, waggon
+beacon, lighthouse, beacon light, pharos
+beaker
+bearskin, busby, shako
+beer bottle
+beer glass
+bell cote, bell cot
+bib
+bicycle-built-for-two, tandem bicycle, tandem
+bikini, two-piece
+binder, ring-binder
+binoculars, field glasses, opera glasses
+birdhouse
+boathouse
+bobsled, bobsleigh, bob
+bolo tie, bolo, bola tie, bola
+bonnet, poke bonnet
+bookcase
+bookshop, bookstore, bookstall
+bottlecap
+bow
+bow tie, bow-tie, bowtie
+brass, memorial tablet, plaque
+brassiere, bra, bandeau
+breakwater, groin, groyne, mole, bulwark, seawall, jetty
+breastplate, aegis, egis
+broom
+bucket, pail
+buckle
+bulletproof vest
+bullet train, bullet
+butcher shop, meat market
+cab, hack, taxi, taxicab
+caldron, cauldron
+candle, taper, wax light
+cannon
+canoe
+can opener, tin opener
+cardigan
+car mirror
+carousel, carrousel, merry-go-round, roundabout, whirligig
+carpenter's kit, tool kit
+carton
+car wheel
+cash machine, cash dispenser, automated teller machine, automatic teller machine, automated teller, automatic teller, ATM
+cassette
+cassette player
+castle
+catamaran
+CD player
+cello, violoncello
+cellular telephone, cellular phone, cellphone, cell, mobile phone
+chain
+chainlink fence
+chain mail, ring mail, mail, chain armor, chain armour, ring armor, ring armour
+chain saw, chainsaw
+chest
+chiffonier, commode
+chime, bell, gong
+china cabinet, china closet
+Christmas stocking
+church, church building
+cinema, movie theater, movie theatre, movie house, picture palace
+cleaver, meat cleaver, chopper
+cliff dwelling
+cloak
+clog, geta, patten, sabot
+cocktail shaker
+coffee mug
+coffeepot
+coil, spiral, volute, whorl, helix
+combination lock
+computer keyboard, keypad
+confectionery, confectionary, candy store
+container ship, containership, container vessel
+convertible
+corkscrew, bottle screw
+cornet, horn, trumpet, trump
+cowboy boot
+cowboy hat, ten-gallon hat
+cradle
+crane
+crash helmet
+crate
+crib, cot
+Crock Pot
+croquet ball
+crutch
+cuirass
+dam, dike, dyke
+desk
+desktop computer
+dial telephone, dial phone
+diaper, nappy, napkin
+digital clock
+digital watch
+dining table, board
+dishrag, dishcloth
+dishwasher, dish washer, dishwashing machine
+disk brake, disc brake
+dock, dockage, docking facility
+dogsled, dog sled, dog sleigh
+dome
+doormat, welcome mat
+drilling platform, offshore rig
+drum, membranophone, tympan
+drumstick
+dumbbell
+Dutch oven
+electric fan, blower
+electric guitar
+electric locomotive
+entertainment center
+envelope
+espresso maker
+face powder
+feather boa, boa
+file, file cabinet, filing cabinet
+fireboat
+fire engine, fire truck
+fire screen, fireguard
+flagpole, flagstaff
+flute, transverse flute
+folding chair
+football helmet
+forklift
+fountain
+fountain pen
+four-poster
+freight car
+French horn, horn
+frying pan, frypan, skillet
+fur coat
+garbage truck, dustcart
+gasmask, respirator, gas helmet
+gas pump, gasoline pump, petrol pump, island dispenser
+goblet
+go-kart
+golf ball
+golfcart, golf cart
+gondola
+gong, tam-tam
+gown
+grand piano, grand
+greenhouse, nursery, glasshouse
+grille, radiator grille
+grocery store, grocery, food market, market
+guillotine
+hair slide
+hair spray
+half track
+hammer
+hamper
+hand blower, blow dryer, blow drier, hair dryer, hair drier
+hand-held computer, hand-held microcomputer
+handkerchief, hankie, hanky, hankey
+hard disc, hard disk, fixed disk
+harmonica, mouth organ, harp, mouth harp
+harp
+harvester, reaper
+hatchet
+holster
+home theater, home theatre
+honeycomb
+hook, claw
+hoopskirt, crinoline
+horizontal bar, high bar
+horse cart, horse-cart
+hourglass
+iPod
+iron, smoothing iron
+jack-o'-lantern
+jean, blue jean, denim
+jeep, landrover
+jersey, T-shirt, tee shirt
+jigsaw puzzle
+jinrikisha, ricksha, rickshaw
+joystick
+kimono
+knee pad
+knot
+lab coat, laboratory coat
+ladle
+lampshade, lamp shade
+laptop, laptop computer
+lawn mower, mower
+lens cap, lens cover
+letter opener, paper knife, paperknife
+library
+lifeboat
+lighter, light, igniter, ignitor
+limousine, limo
+liner, ocean liner
+lipstick, lip rouge
+Loafer
+lotion
+loudspeaker, speaker, speaker unit, loudspeaker system, speaker system
+loupe, jeweler's loupe
+lumbermill, sawmill
+magnetic compass
+mailbag, postbag
+mailbox, letter box
+maillot
+maillot, tank suit
+manhole cover
+maraca
+marimba, xylophone
+mask
+matchstick
+maypole
+maze, labyrinth
+measuring cup
+medicine chest, medicine cabinet
+megalith, megalithic structure
+microphone, mike
+microwave, microwave oven
+military uniform
+milk can
+minibus
+miniskirt, mini
+minivan
+missile
+mitten
+mixing bowl
+mobile home, manufactured home
+Model T
+modem
+monastery
+monitor
+moped
+mortar
+mortarboard
+mosque
+mosquito net
+motor scooter, scooter
+mountain bike, all-terrain bike, off-roader
+mountain tent
+mouse, computer mouse
+mousetrap
+moving van
+muzzle
+nail
+neck brace
+necklace
+nipple
+notebook, notebook computer
+obelisk
+oboe, hautboy, hautbois
+ocarina, sweet potato
+odometer, hodometer, mileometer, milometer
+oil filter
+organ, pipe organ
+oscilloscope, scope, cathode-ray oscilloscope, CRO
+overskirt
+oxcart
+oxygen mask
+packet
+paddle, boat paddle
+paddlewheel, paddle wheel
+padlock
+paintbrush
+pajama, pyjama, pj's, jammies
+palace
+panpipe, pandean pipe, syrinx
+paper towel
+parachute, chute
+parallel bars, bars
+park bench
+parking meter
+passenger car, coach, carriage
+patio, terrace
+pay-phone, pay-station
+pedestal, plinth, footstall
+pencil box, pencil case
+pencil sharpener
+perfume, essence
+Petri dish
+photocopier
+pick, plectrum, plectron
+pickelhaube
+picket fence, paling
+pickup, pickup truck
+pier
+piggy bank, penny bank
+pill bottle
+pillow
+ping-pong ball
+pinwheel
+pirate, pirate ship
+pitcher, ewer
+plane, carpenter's plane, woodworking plane
+planetarium
+plastic bag
+plate rack
+plow, plough
+plunger, plumber's helper
+Polaroid camera, Polaroid Land camera
+pole
+police van, police wagon, paddy wagon, patrol wagon, wagon, black Maria
+poncho
+pool table, billiard table, snooker table
+pop bottle, soda bottle
+pot, flowerpot
+potter's wheel
+power drill
+prayer rug, prayer mat
+printer
+prison, prison house
+projectile, missile
+projector
+puck, hockey puck
+punching bag, punch bag, punching ball, punchball
+purse
+quill, quill pen
+quilt, comforter, comfort, puff
+racer, race car, racing car
+racket, racquet
+radiator
+radio, wireless
+radio telescope, radio reflector
+rain barrel
+recreational vehicle, RV, R.V.
+reel
+reflex camera
+refrigerator, icebox
+remote control, remote
+restaurant, eating house, eating place, eatery
+revolver, six-gun, six-shooter
+rifle
+rocking chair, rocker
+rotisserie
+rubber eraser, rubber, pencil eraser
+rugby ball
+rule, ruler
+running shoe
+safe
+safety pin
+saltshaker, salt shaker
+sandal
+sarong
+sax, saxophone
+scabbard
+scale, weighing machine
+school bus
+schooner
+scoreboard
+screen, CRT screen
+screw
+screwdriver
+seat belt, seatbelt
+sewing machine
+shield, buckler
+shoe shop, shoe-shop, shoe store
+shoji
+shopping basket
+shopping cart
+shovel
+shower cap
+shower curtain
+ski
+ski mask
+sleeping bag
+slide rule, slipstick
+sliding door
+slot, one-armed bandit
+snorkel
+snowmobile
+snowplow, snowplough
+soap dispenser
+soccer ball
+sock
+solar dish, solar collector, solar furnace
+sombrero
+soup bowl
+space bar
+space heater
+space shuttle
+spatula
+speedboat
+spider web, spider's web
+spindle
+sports car, sport car
+spotlight, spot
+stage
+steam locomotive
+steel arch bridge
+steel drum
+stethoscope
+stole
+stone wall
+stopwatch, stop watch
+stove
+strainer
+streetcar, tram, tramcar, trolley, trolley car
+stretcher
+studio couch, day bed
+stupa, tope
+submarine, pigboat, sub, U-boat
+suit, suit of clothes
+sundial
+sunglass
+sunglasses, dark glasses, shades
+sunscreen, sunblock, sun blocker
+suspension bridge
+swab, swob, mop
+sweatshirt
+swimming trunks, bathing trunks
+swing
+switch, electric switch, electrical switch
+syringe
+table lamp
+tank, army tank, armored combat vehicle, armoured combat vehicle
+tape player
+teapot
+teddy, teddy bear
+television, television system
+tennis ball
+thatch, thatched roof
+theater curtain, theatre curtain
+thimble
+thresher, thrasher, threshing machine
+throne
+tile roof
+toaster
+tobacco shop, tobacconist shop, tobacconist
+toilet seat
+torch
+totem pole
+tow truck, tow car, wrecker
+toyshop
+tractor
+trailer truck, tractor trailer, trucking rig, rig, articulated lorry, semi
+tray
+trench coat
+tricycle, trike, velocipede
+trimaran
+tripod
+triumphal arch
+trolleybus, trolley coach, trackless trolley
+trombone
+tub, vat
+turnstile
+typewriter keyboard
+umbrella
+unicycle, monocycle
+upright, upright piano
+vacuum, vacuum cleaner
+vase
+vault
+velvet
+vending machine
+vestment
+viaduct
+violin, fiddle
+volleyball
+waffle iron
+wall clock
+wallet, billfold, notecase, pocketbook
+wardrobe, closet, press
+warplane, military plane
+washbasin, handbasin, washbowl, lavabo, wash-hand basin
+washer, automatic washer, washing machine
+water bottle
+water jug
+water tower
+whiskey jug
+whistle
+wig
+window screen
+window shade
+Windsor tie
+wine bottle
+wing
+wok
+wooden spoon
+wool, woolen, woollen
+worm fence, snake fence, snake-rail fence, Virginia fence
+wreck
+yawl
+yurt
+web site, website, internet site, site
+comic book
+crossword puzzle, crossword
+street sign
+traffic light, traffic signal, stoplight
+book jacket, dust cover, dust jacket, dust wrapper
+menu
+plate
+guacamole
+consomme
+hot pot, hotpot
+trifle
+ice cream, icecream
+ice lolly, lolly, lollipop, popsicle
+French loaf
+bagel, beigel
+pretzel
+cheeseburger
+hotdog, hot dog, red hot
+mashed potato
+head cabbage
+broccoli
+cauliflower
+zucchini, courgette
+spaghetti squash
+acorn squash
+butternut squash
+cucumber, cuke
+artichoke, globe artichoke
+bell pepper
+cardoon
+mushroom
+Granny Smith
+strawberry
+orange
+lemon
+fig
+pineapple, ananas
+banana
+jackfruit, jak, jack
+custard apple
+pomegranate
+hay
+carbonara
+chocolate sauce, chocolate syrup
+dough
+meat loaf, meatloaf
+pizza, pizza pie
+potpie
+burrito
+red wine
+espresso
+cup
+eggnog
+alp
+bubble
+cliff, drop, drop-off
+coral reef
+geyser
+lakeside, lakeshore
+promontory, headland, head, foreland
+sandbar, sand bar
+seashore, coast, seacoast, sea-coast
+valley, vale
+volcano
+ballplayer, baseball player
+groom, bridegroom
+scuba diver
+rapeseed
+daisy
+yellow lady's slipper, yellow lady-slipper, Cypripedium calceolus, Cypripedium parviflorum
+corn
+acorn
+hip, rose hip, rosehip
+buckeye, horse chestnut, conker
+coral fungus
+agaric
+gyromitra
+stinkhorn, carrion fungus
+earthstar
+hen-of-the-woods, hen of the woods, Polyporus frondosus, Grifola frondosa
+bolete
+ear, spike, capitulum
+toilet tissue, toilet paper, bathroom tissue'''.split("\n")
\ No newline at end of file
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/nets_utils.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/nets_utils.py
new file mode 100644
index 0000000000000000000000000000000000000000..16b0bf54c8edc6769a9baed666b89b375da5c11f
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/nets_utils.py
@@ -0,0 +1,31 @@
+from npu_bridge.npu_init import *
+from tensorflow.python import pywrap_tensorflow
+import collections
+import numpy as np
+
+
+var_stat = collections.namedtuple('stats', ['mean', 'median', 'std'])
+
+
+def get_variables_in_checkpoint_file(file_name):
+ try:
+ reader = pywrap_tensorflow.NewCheckpointReader(file_name)
+ var_to_shape_map = reader.get_variable_to_shape_map()
+ return var_to_shape_map
+ except Exception as e: # pylint: disable=broad-except
+ print(str(e))
+ if "corrupted compressed block contents" in str(e):
+ print("It's likely that your checkpoint file has been compressed "
+ "with SNAPPY.")
+
+
+def get_tensor_static_val(file_name, all_tensors, all_tensor_names):
+ reader = pywrap_tensorflow.NewCheckpointReader(file_name)
+ vars_dict = {}
+ if all_tensors or all_tensor_names:
+ var_to_shape_map = reader.get_variable_to_shape_map()
+ for key in sorted(var_to_shape_map):
+ if all_tensors:
+ vars_dict[key] = var_stat(np.mean(reader.get_tensor(key)), np.median(reader.get_tensor(key)),
+ np.std(reader.get_tensor(key)))
+ return vars_dict
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/networks.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/networks.py
new file mode 100644
index 0000000000000000000000000000000000000000..e345b80c190d8f7fb3a0759c2561142dd6cedf31
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/networks.py
@@ -0,0 +1,20 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+from vgg16 import get_vgg16
+from vgg19 import get_vgg19
+
+
+def get_model(inputs, sess, type, pretrained=True):
+ if type == 'vgg16':
+ return get_vgg16(inputs, sess, pretrained)
+ elif type == 'vgg19':
+ return get_vgg19(inputs, sess, pretrained)
+
+
+if __name__ == '__main__':
+ tfconfig = tf.ConfigProto(allow_soft_placement=True)
+ x = tf.placeholder(dtype=tf.float32, shape=[None, 224, 224, 3], name='inpust')
+ with tf.Session(config=npu_config_proto(config_proto=tfconfig)) as sess:
+ network = get_model(x, sess, type='vgg19', pretrained=True)
+ network.print_params()
+ network.print_layers()
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/readme.md b/TensorFlow/contrib/cv/InsightFace_TF/nets/readme.md
new file mode 100644
index 0000000000000000000000000000000000000000..614cc24f3bd22216287d40d1633330a4633671c4
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/readme.md
@@ -0,0 +1,5 @@
+1. `vgg16.py`
+ the vgg16 model, input is an image that with shape 224*224 and mean substract. The input should be first resized to 224*224, and then substract the mean. The channels of the input image is `RGB`.
+2. 'vgg19.py'
+ the vgg19 model. The input should be first normalized to [0, 1], then resized to 224*224 and then do substract to mean, the channel of the input should be `RGB`.
+3. 'resnet'
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/resnet.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/resnet.py
new file mode 100644
index 0000000000000000000000000000000000000000..71d23518587c11dd8c6c49dc9998a957e4ec40a4
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/resnet.py
@@ -0,0 +1,502 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+from tensorflow.contrib.layers.python.layers import utils
+import collections
+from tensorlayer.layers import Layer, list_remove_repeat
+from nets_utils import get_variables_in_checkpoint_file
+
+
+class ElementwiseLayer(Layer):
+ """
+ The :class:`ElementwiseLayer` class combines multiple :class:`Layer` which have the same output shapes by a given elemwise-wise operation.
+
+ Parameters
+ ----------
+ layer : a list of :class:`Layer` instances
+ The `Layer` class feeding into this layer.
+ combine_fn : a TensorFlow elemwise-merge function
+ e.g. AND is ``tf.minimum`` ; OR is ``tf.maximum`` ; ADD is ``tf.add`` ; MUL is ``tf.multiply`` and so on.
+ See `TensorFlow Math API `_ .
+ name : a string or None
+ An optional name to attach to this layer.
+ """
+ def __init__(
+ self,
+ layer = [],
+ combine_fn = tf.minimum,
+ name ='elementwise_layer',
+ act = None,
+ ):
+ Layer.__init__(self, name=name)
+
+ if act:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s, act:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__, act.__name__))
+ else:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__))
+
+ self.outputs = layer[0].outputs
+ # print(self.outputs._shape, type(self.outputs._shape))
+ for l in layer[1:]:
+ # assert str(self.outputs.get_shape()) == str(l.outputs.get_shape()), "Hint: the input shapes should be the same. %s != %s" % (self.outputs.get_shape() , str(l.outputs.get_shape()))
+ self.outputs = combine_fn(self.outputs, l.outputs, name=name)
+ if act:
+ self.outputs = act(self.outputs)
+ self.all_layers = list(layer[0].all_layers)
+ self.all_params = list(layer[0].all_params)
+ self.all_drop = dict(layer[0].all_drop)
+
+ for i in range(1, len(layer)):
+ self.all_layers.extend(list(layer[i].all_layers))
+ self.all_params.extend(list(layer[i].all_params))
+ self.all_drop.update(dict(layer[i].all_drop))
+
+ self.all_layers = list_remove_repeat(self.all_layers)
+ self.all_params = list_remove_repeat(self.all_params)
+
+
+def subsample(inputs, factor, scope=None):
+ if factor == 1:
+ return inputs
+ else:
+ return tl.layers.MaxPool2d(inputs, [1, 1], strides=(factor, factor), name=scope)
+
+
+def conv2d_same(inputs, num_outputs, kernel_size, strides, rate=1, scope=None):
+ '''
+ Reference slim resnet
+ :param inputs:
+ :param num_outputs:
+ :param kernel_size:
+ :param strides:
+ :param rate:
+ :param scope:
+ :return:
+ '''
+ if strides == 1:
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), act=None, padding='SAME', name=scope)
+ nets = tl.layers.BatchNormLayer(nets, act=tf.nn.relu, is_train=True, name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size),
+ rate=rate, act=None, padding='SAME', name=scope)
+ nets = tl.layers.BatchNormLayer(nets, act=tf.nn.relu, is_train=True, name=scope+'_bn/BatchNorm')
+ return nets
+ else:
+ kernel_size_effective = kernel_size + (kernel_size - 1) * (rate - 1)
+ pad_total = kernel_size_effective - 1
+ pad_beg = pad_total // 2
+ pad_end = pad_total - pad_beg
+ inputs = tl.layers.PadLayer(inputs, [[0, 0], [pad_beg, pad_end], [pad_beg, pad_end], [0, 0]], name='padding_%s' % scope)
+ if rate == 1:
+ nets = tl.layers.Conv2d(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ strides=(strides, strides), act=None, padding='VALID', name=scope)
+ nets = tl.layers.BatchNormLayer(nets, act=tf.nn.relu, is_train=True, name=scope+'_bn/BatchNorm')
+ else:
+ nets = tl.layers.AtrousConv2dLayer(inputs, n_filter=num_outputs, filter_size=(kernel_size, kernel_size), b_init=None,
+ rate=rate, act=None, padding='SAME', name=scope)
+ nets = tl.layers.BatchNormLayer(nets, act=tf.nn.relu, is_train=True, name=scope+'_bn/BatchNorm')
+ return nets
+
+
+def bottleneck(inputs, depth, depth_bottleneck, stride, rate=1, scope=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ b_init=None, name='shortcut_conv')
+ shortcut = tl.layers.BatchNormLayer(shortcut, act=tf.identity, is_train=True, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = tl.layers.Conv2d(inputs, depth_bottleneck, filter_size=(1, 1), strides=(1, 1), act=None, b_init=None,
+ name='conv1')
+ residual = tl.layers.BatchNormLayer(residual, act=tf.nn.relu, is_train=True, name='conv1_bn/BatchNorm')
+
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth_bottleneck, kernel_size=3, strides= stride, rate=rate, scope='conv2')
+
+ # bottleneck layer 3
+ residual = tl.layers.Conv2d(residual, depth, filter_size=(1, 1), strides=(1, 1), act=None, b_init=None,
+ name='conv3')
+ residual = tl.layers.BatchNormLayer(residual, act=tf.identity, is_train=True, name='conv3_bn/BatchNorm')
+
+ output = ElementwiseLayer(layer=[shortcut, residual],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=tf.nn.relu)
+ return output
+
+
+def bottleneck_SE(inputs, depth, depth_bottleneck, stride, rate=1, scope=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ b_init=None, name='shortcut_conv')
+ shortcut = tl.layers.BatchNormLayer(shortcut, act=tf.identity, is_train=True, name='shortcut_bn/BatchNorm')
+ # bottleneck layer 1
+ residual = tl.layers.Conv2d(inputs, depth_bottleneck, filter_size=(1, 1), strides=(1, 1), act=None, b_init=None,
+ name='conv1')
+ residual = tl.layers.BatchNormLayer(residual, act=tf.nn.relu, is_train=True, name='conv1_bn/BatchNorm')
+
+ # bottleneck layer 2
+ residual = conv2d_same(residual, depth_bottleneck, kernel_size=3, strides= stride, rate=rate, scope='conv2')
+
+ # bottleneck layer 3
+ residual = tl.layers.Conv2d(residual, depth, filter_size=(1, 1), strides=(1, 1), act=None, b_init=None,
+ name='conv3')
+ residual = tl.layers.BatchNormLayer(residual, act=tf.identity, is_train=True, name='conv3_bn/BatchNorm')
+
+ # squeeze
+ squeeze = tl.layers.InputLayer(tf.reduce_mean(residual.outputs, axis=[1, 2]), name='squeeze_layer')
+ # excitation
+ excitation1 = tl.layers.DenseLayer(squeeze, n_units=int(depth/16.0), act=tf.nn.relu, name='excitation_1')
+ excitation2 = tl.layers.DenseLayer(excitation1, n_units=depth, act=tf.nn.sigmoid, name='excitation_2')
+ # scale
+ scale = tl.layers.ReshapeLayer(excitation2, shape=[tf.shape(excitation2.outputs)[0], 1, 1, depth], name='excitation_reshape')
+
+ residual_se = ElementwiseLayer(layer=[residual, scale],
+ combine_fn=tf.multiply,
+ name='scale_layer',
+ act=None)
+
+ output = ElementwiseLayer(layer=[shortcut, residual_se],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=tf.nn.relu)
+ return output
+
+
+def bottleneck_Xt(inputs, depth, stride, cardinality, cardinality_dim, rate=1, scope=None):
+ with tf.variable_scope(scope, 'bottleneck_v1') as sc:
+ depth_in = utils.last_dimension(inputs.outputs.get_shape(), min_rank=4)
+ if depth == depth_in:
+ shortcut = subsample(inputs, stride, 'shortcut')
+ else:
+ shortcut = tl.layers.Conv2d(inputs, depth, filter_size=(1, 1), strides=(stride, stride), act=None,
+ b_init=None, name='shortcut_conv')
+ shortcut = tl.layers.BatchNormLayer(shortcut, act=tf.identity, is_train=True, name='shortcut_bn/BatchNorm')
+
+ cardinality_layers = []
+ for i in range(cardinality):
+ # bottleneck layer 1
+ residual = tl.layers.Conv2d(inputs, cardinality_dim, filter_size=(1, 1), strides=(1, 1), act=None, b_init=None,
+ name='cardinality_%d/conv1' % i)
+ residual = tl.layers.BatchNormLayer(residual, act=tf.nn.relu, is_train=True, name='cardinality_%d/conv1_bn/BatchNorm' % i)
+
+ # bottleneck layer 2
+ residual = conv2d_same(residual, cardinality_dim, kernel_size=3, strides= stride, rate=rate, scope='cardinality_%d/conv2' % i)
+
+ # bottleneck layer 3
+ residual = tl.layers.Conv2d(residual, depth, filter_size=(1, 1), strides=(1, 1), act=None, b_init=None,
+ name='cardinality_%d/conv3' % i)
+ residual = tl.layers.BatchNormLayer(residual, act=tf.identity, is_train=True, name='cardinality_%d/conv3_bn/BatchNorm' % i)
+ cardinality_layers.append(residual)
+
+ residual_total = ElementwiseLayer(layer=cardinality_layers,
+ combine_fn=tf.add,
+ name='cardinality_cmobine',
+ act=None)
+ with tf.control_dependencies([residual_total.outputs]):
+ output = ElementwiseLayer(layer=[shortcut, residual_total],
+ combine_fn=tf.add,
+ name='combine_layer',
+ act=tf.nn.relu)
+ return output
+
+
+def resnet(inputs, bottle_neck, blocks, num_classes=1000, scope=None, type=None):
+ # mean_rgb_var = tf.Variable()
+ with tf.variable_scope(scope):
+ mean_rgb_var = tf.Variable(dtype=tf.float32, name='mean_rgb', trainable=False, initial_value=[128.0, 128.0, 128.0])
+ rgb_mean_dims = tf.reshape(mean_rgb_var, shape=[1, 1, 1, 3])
+ inputs = tf.subtract(inputs, rgb_mean_dims)
+ net_inputs = tl.layers.InputLayer(inputs, name='input_layer')
+ if bottle_neck:
+ net = conv2d_same(net_inputs, 64, 7, strides=2, rate=1, scope='conv1')
+ net = tl.layers.MaxPool2d(net, (2, 2), padding='SAME', name='pool1')
+ else:
+ raise ValueError('The standard resnet must support the bottleneck layer')
+ for block in blocks:
+ with tf.variable_scope(block.scope):
+ for i, var in enumerate(block.args):
+ with tf.variable_scope('unit_%d' % (i+1)):
+ if type=='resnext':
+ net = block.unit_fn(net, depth=var['depth'], stride=var['stride'],
+ cardinality=var['cardinality'],
+ cardinality_dim=var['cardinality_dim'], rate=1, scope=None)
+ else:
+ net = block.unit_fn(net, depth=var['depth'], depth_bottleneck=var['depth_bottleneck'],
+ stride=var['stride'], rate=var['rate'], scope=None)
+ net.outputs = tf.reduce_mean(net.outputs, [1, 2], keep_dims=True)
+ net = tl.layers.Conv2d(net, num_classes, filter_size=(1, 1), strides=(1, 1), act=None, name='logits')
+ net.outputs = tf.squeeze(net.outputs, [1, 2], name='SpatialSqueeze')
+ return net
+
+
+class Block(collections.namedtuple('Block', ['scope', 'unit_fn', 'args'])):
+ """A named tuple describing a ResNet block.
+
+ Its parts are:
+ scope: The scope of the `Block`.
+ unit_fn: The ResNet unit function which takes as input a `Tensor` and
+ returns another `Tensor` with the output of the ResNet unit.
+ args: A list of length equal to the number of units in the `Block`. The list
+ contains one (depth, depth_bottleneck, stride) tuple for each unit in the
+ block to serve as argument to unit_fn.
+ """
+
+
+def resnet_v1_block(scope, base_depth, num_units, stride, rate=1):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, bottleneck, [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': 1,
+ 'rate': rate
+ }] * (num_units - 1) + [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': stride,
+ 'rate': rate
+ }])
+
+
+def resnetse_v1_block(scope, base_depth, num_units, stride, rate=1):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, bottleneck_SE, [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': 1,
+ 'rate': rate
+ }] * (num_units - 1) + [{
+ 'depth': base_depth * 4,
+ 'depth_bottleneck': base_depth,
+ 'stride': stride,
+ 'rate': rate
+ }])
+
+
+def resnext_v1_block(scope, base_depth, num_units, stride, cardinality, cardinality_dim, rate=1):
+ """Helper function for creating a resnet_v1 bottleneck block.
+
+ Args:
+ scope: The scope of the block.
+ base_depth: The depth of the bottleneck layer for each unit.
+ num_units: The number of units in the block.
+ stride: The stride of the block, implemented as a stride in the last unit.
+ All other units have stride=1.
+
+ Returns:
+ A resnet_v1 bottleneck block.
+ """
+ return Block(scope, bottleneck_Xt, [{
+ 'depth': base_depth * 4,
+ 'stride': 1,
+ 'rate': rate,
+ 'cardinality': cardinality,
+ 'cardinality_dim': cardinality_dim
+ }] * (num_units - 1) + [{
+ 'depth': base_depth * 4,
+ 'stride': stride,
+ 'rate': 1,
+ 'cardinality': cardinality,
+ 'cardinality_dim': cardinality_dim
+ }])
+
+
+def get_resnet(inputs, num_classes, num_layers, type='resnet', sess=None, pretrained=True):
+ '''
+ :param inputs: inputs is an tensorflow placeholder
+ :param num_classes:
+ :param num_layers:
+ :param type: choose weather using se xt or just oridinary resnet ['resnet', 'resnetse', 'resnext']
+ :return:
+ '''
+ if type == 'resnet':
+ if num_layers == 50:
+ blocks = [
+ resnet_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1),
+ resnet_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1),
+ resnet_v1_block('block3', base_depth=256, num_units=6, stride=2, rate=1),
+ resnet_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1)
+ ]
+ ckpt_file_path = '../model_weights/resnet_v1_50.ckpt'
+ elif num_layers == 101:
+ blocks = [
+ resnet_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1),
+ resnet_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1),
+ resnet_v1_block('block3', base_depth=256, num_units=23, stride=2, rate=1),
+ resnet_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1)
+ ]
+ ckpt_file_path = '../model_weights/resnet_v1_101.ckpt'
+ elif num_layers == 152:
+ blocks = [
+ resnet_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1),
+ resnet_v1_block('block2', base_depth=128, num_units=8, stride=2, rate=1),
+ resnet_v1_block('block3', base_depth=256, num_units=36, stride=2, rate=1),
+ resnet_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1)
+ ]
+ ckpt_file_path = '../model_weights/resnet_v1_152.ckpt'
+ else:
+ raise ValueError('Resnet layer %d is not supported now.' % num_layers)
+ net = resnet(inputs=inputs,
+ bottle_neck=True,
+ blocks=blocks,
+ num_classes=num_classes,
+ scope='resnet_v1_%d' % num_layers,
+ type=type)
+ if pretrained:
+ var_ckpt = get_variables_in_checkpoint_file(ckpt_file_path)
+ vars = tf.get_collection(tf.GraphKeys.GLOBAL_VARIABLES)
+ vars_dict = {}
+ for var in vars:
+ var_name = var.op.name
+ var_name_new = var_name
+ if '_bn' in var_name:
+ var_name_new = var_name_new.replace('_bn', '')
+ if 'W_conv2d' in var_name:
+ var_name_new = var_name_new.replace('W_conv2d', 'weights')
+ if 'b_conv2d' in var_name:
+ var_name_new = var_name_new.replace('b_conv2d', 'biases')
+ if 'shortcut_conv' in var_name:
+ var_name_new = var_name_new.replace('shortcut_conv', 'shortcut')
+ print(var_name_new)
+ if var_name_new in var_ckpt:
+ vars_dict[var_name_new] = var
+ tl.layers.initialize_global_variables(sess)
+ if len(vars_dict.keys()) > 0:
+ saver = tf.train.Saver(vars_dict)
+ saver.restore(sess, ckpt_file_path)
+ return net
+ else:
+ return net
+ elif type == 'resnetse':
+ if pretrained:
+ raise ValueError('resnetse have no pretrained model!')
+ if num_layers == 50:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1),
+ resnetse_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1),
+ resnetse_v1_block('block3', base_depth=256, num_units=6, stride=2, rate=1),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1)
+ ]
+ elif num_layers == 101:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1),
+ resnetse_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1),
+ resnetse_v1_block('block3', base_depth=256, num_units=23, stride=2, rate=1),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1)
+ ]
+ elif num_layers == 152:
+ blocks = [
+ resnetse_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1),
+ resnetse_v1_block('block2', base_depth=128, num_units=8, stride=2, rate=1),
+ resnetse_v1_block('block3', base_depth=256, num_units=36, stride=2, rate=1),
+ resnetse_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1)
+ ]
+ else:
+ raise ValueError('Resnet layer %d is not supported now.' % num_layers)
+ net = resnet(inputs=inputs,
+ bottle_neck=True,
+ blocks=blocks,
+ num_classes=num_classes,
+ scope='resnet_v1_%d' % num_layers,
+ type=type)
+ return net
+ elif type == 'resnext':
+ if pretrained:
+ raise ValueError('resnetse have no pretrained model!')
+ if num_layers == 50:
+ blocks = [
+ resnext_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block3', base_depth=256, num_units=6, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1, cardinality=32,
+ cardinality_dim=4)
+ ]
+ elif num_layers == 101:
+ blocks = [
+ resnext_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block2', base_depth=128, num_units=4, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block3', base_depth=256, num_units=23, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1, cardinality=32,
+ cardinality_dim=4)
+ ]
+ elif num_layers == 152:
+ blocks = [
+ resnext_v1_block('block1', base_depth=64, num_units=3, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block2', base_depth=128, num_units=8, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block3', base_depth=256, num_units=36, stride=2, rate=1, cardinality=32,
+ cardinality_dim=4),
+ resnext_v1_block('block4', base_depth=512, num_units=3, stride=1, rate=1, cardinality=32,
+ cardinality_dim=4)
+ ]
+ else:
+ raise ValueError('Resnet layer %d is not supported now.' % num_layers)
+ net = resnet(inputs=inputs,
+ bottle_neck=True,
+ blocks=blocks,
+ num_classes=num_classes,
+ scope='resnet_v1_%d' % num_layers,
+ type=type)
+ return net
+ else:
+ raise ValueError('resnet type %s does not support now!' % type)
+
+
+if __name__ == '__main__':
+ x = tf.placeholder(dtype=tf.float32, shape=[1, 224, 224, 3], name='input_place')
+ sess = tf.Session(config=npu_config_proto())
+ # # test resnet
+ # nets = get_resnet(x, 1000, 50, type='resnet', sess=sess, pretrained=True)
+ # with sess:
+ # nets.print_params()
+
+ # # test resnetse
+ # nets = get_resnet(x, 1000, 50, type='resnetse', sess=sess, pretrained=False)
+ # tl.layers.initialize_global_variables(sess)
+ # with sess:
+ # nets.print_params()
+
+ # test resnext
+ nets = get_resnet(x, 1000, 50, type='resnext', sess=sess, pretrained=False)
+ tl.layers.initialize_global_variables(sess)
+ with sess:
+ nets.print_params()
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/tl_layers_modify.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/tl_layers_modify.py
new file mode 100644
index 0000000000000000000000000000000000000000..f76e6e926b054e1354f85fdb6a56be1a65998eec
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/tl_layers_modify.py
@@ -0,0 +1,567 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+from tensorlayer.layers import Layer, list_remove_repeat
+
+
+D_TYPE = tf.float32
+TF_GRAPHKEYS_VARIABLES = tf.GraphKeys.GLOBAL_VARIABLES
+
+
+class ElementwiseLayer(Layer):
+ """
+ The :class:`ElementwiseLayer` class combines multiple :class:`Layer` which have the same output shapes by a given elemwise-wise operation.
+
+ Parameters
+ ----------
+ layer : a list of :class:`Layer` instances
+ The `Layer` class feeding into this layer.
+ combine_fn : a TensorFlow elemwise-merge function
+ e.g. AND is ``tf.minimum`` ; OR is ``tf.maximum`` ; ADD is ``tf.add`` ; MUL is ``tf.multiply`` and so on.
+ See `TensorFlow Math API `_ .
+ name : a string or None
+ An optional name to attach to this layer.
+ """
+ def __init__(
+ self,
+ layer = [],
+ combine_fn = tf.minimum,
+ name ='elementwise_layer',
+ act = None,
+ ):
+ Layer.__init__(self, name=name)
+
+ if act:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s, act:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__, act.__name__))
+ else:
+ print(" [TL] ElementwiseLayer %s: size:%s fn:%s" % (
+ self.name, layer[0].outputs.get_shape(), combine_fn.__name__))
+
+ self.outputs = layer[0].outputs
+ # print(self.outputs._shape, type(self.outputs._shape))
+ for l in layer[1:]:
+ # assert str(self.outputs.get_shape()) == str(l.outputs.get_shape()), "Hint: the input shapes should be the same. %s != %s" % (self.outputs.get_shape() , str(l.outputs.get_shape()))
+ self.outputs = combine_fn(self.outputs, l.outputs, name=name)
+ if act:
+ self.outputs = act(self.outputs)
+ self.all_layers = list(layer[0].all_layers)
+ self.all_params = list(layer[0].all_params)
+ self.all_drop = dict(layer[0].all_drop)
+
+ for i in range(1, len(layer)):
+ self.all_layers.extend(list(layer[i].all_layers))
+ self.all_params.extend(list(layer[i].all_params))
+ self.all_drop.update(dict(layer[i].all_drop))
+
+ self.all_layers = list_remove_repeat(self.all_layers)
+ self.all_params = list_remove_repeat(self.all_params)
+
+
+class BatchNormLayer(Layer):
+ """
+ The :class:`BatchNormLayer` class is a normalization layer, see ``tf.nn.batch_normalization`` and ``tf.nn.moments``.
+
+ Batch normalization on fully-connected or convolutional maps.
+
+ ```
+ https://www.tensorflow.org/api_docs/python/tf/cond
+ If x < y, the tf.add operation will be executed and tf.square operation will not be executed.
+ Since z is needed for at least one branch of the cond, the tf.multiply operation is always executed, unconditionally.
+ ```
+
+ Parameters
+ -----------
+ layer : a :class:`Layer` instance
+ The `Layer` class feeding into this layer.
+ decay : float, default is 0.9.
+ A decay factor for ExponentialMovingAverage, use larger value for large dataset.
+ epsilon : float
+ A small float number to avoid dividing by 0.
+ act : activation function.
+ is_train : boolean
+ Whether train or inference.
+ beta_init : beta initializer
+ The initializer for initializing beta
+ gamma_init : gamma initializer
+ The initializer for initializing gamma
+ dtype : tf.float32 (default) or tf.float16
+ name : a string or None
+ An optional name to attach to this layer.
+
+ References
+ ----------
+ - `Source `_
+ - `stackoverflow `_
+
+ """
+
+ def __init__(
+ self,
+ layer=None,
+ decay=0.9,
+ epsilon=2e-5,
+ act=tf.identity,
+ is_train=False,
+ fix_gamma=True,
+ beta_init=tf.zeros_initializer,
+ gamma_init=tf.random_normal_initializer(mean=1.0, stddev=0.002), # tf.ones_initializer,
+ # dtype = tf.float32,
+ trainable=None,
+ name='batchnorm_layer',
+ ):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ print(" [TL] BatchNormLayer %s: decay:%f epsilon:%f act:%s is_train:%s" % (self.name, decay, epsilon, act.__name__, is_train))
+ x_shape = self.inputs.get_shape()
+ params_shape = x_shape[-1:]
+
+ from tensorflow.python.training import moving_averages
+ from tensorflow.python.ops import control_flow_ops
+
+ with tf.variable_scope(name) as vs:
+ axis = list(range(len(x_shape) - 1))
+
+ ## 1. beta, gamma
+ if tf.__version__ > '0.12.1' and beta_init == tf.zeros_initializer:
+ beta_init = beta_init()
+ with tf.device('/cpu:0'):
+ beta = tf.get_variable('beta', shape=params_shape, initializer=beta_init, dtype=tf.float32, trainable=is_train) #, restore=restore)
+
+ gamma = tf.get_variable(
+ 'gamma',
+ shape=params_shape,
+ initializer=gamma_init,
+ dtype=tf.float32,
+ trainable=fix_gamma,
+ ) #restore=restore)
+
+ ## 2.
+ if tf.__version__ > '0.12.1':
+ moving_mean_init = tf.zeros_initializer()
+ else:
+ moving_mean_init = tf.zeros_initializer
+ with tf.device('/cpu:0'):
+ moving_mean = tf.get_variable('moving_mean', params_shape, initializer=moving_mean_init, dtype=tf.float32, trainable=False) # restore=restore)
+ moving_variance = tf.get_variable(
+ 'moving_variance',
+ params_shape,
+ initializer=tf.constant_initializer(1.),
+ dtype=tf.float32,
+ trainable=False,
+ ) # restore=restore)
+
+ ## 3.
+ # These ops will only be preformed when training.
+ mean, variance = tf.nn.moments(self.inputs, axis)
+ try: # TF12
+ update_moving_mean = moving_averages.assign_moving_average(moving_mean, mean, decay, zero_debias=False) # if zero_debias=True, has bias
+ update_moving_variance = moving_averages.assign_moving_average(
+ moving_variance, variance, decay, zero_debias=False) # if zero_debias=True, has bias
+ # print("TF12 moving")
+ except Exception as e: # TF11
+ update_moving_mean = moving_averages.assign_moving_average(moving_mean, mean, decay)
+ update_moving_variance = moving_averages.assign_moving_average(moving_variance, variance, decay)
+ # print("TF11 moving")
+
+ def mean_var_with_update():
+ with tf.control_dependencies([update_moving_mean, update_moving_variance]):
+ return tf.identity(mean), tf.identity(variance)
+ if trainable:
+ mean, var = mean_var_with_update()
+ print(mean)
+ print(var)
+ self.outputs = act(tf.nn.batch_normalization(self.inputs, mean, var, beta, gamma, epsilon))
+ else:
+ self.outputs = act(tf.nn.batch_normalization(self.inputs, moving_mean, moving_variance, beta, gamma, epsilon))
+ variables = [beta, gamma, moving_mean, moving_variance]
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+ self.all_layers.extend([self.outputs])
+ self.all_params.extend(variables)
+
+
+def Conv2d(
+ net,
+ n_filter=32,
+ filter_size=(3, 3),
+ strides=(1, 1),
+ act=None,
+ padding='SAME',
+ W_init=tf.truncated_normal_initializer(stddev=0.02),
+ b_init=tf.constant_initializer(value=0.0),
+ W_init_args={},
+ b_init_args={},
+ use_cudnn_on_gpu=None,
+ data_format=None,
+ name='conv2d',
+):
+ """Wrapper for :class:`Conv2dLayer`, if you don't understand how to use :class:`Conv2dLayer`, this function may be easier.
+
+ Parameters
+ ----------
+ net : TensorLayer layer.
+ n_filter : number of filter.
+ filter_size : tuple (height, width) for filter size.
+ strides : tuple (height, width) for strides.
+ act : None or activation function.
+ others : see :class:`Conv2dLayer`.
+
+ Examples
+ --------
+ >>> w_init = tf.truncated_normal_initializer(stddev=0.01)
+ >>> b_init = tf.constant_initializer(value=0.0)
+ >>> inputs = InputLayer(x, name='inputs')
+ >>> conv1 = Conv2d(inputs, 64, (3, 3), act=tf.nn.relu, padding='SAME', W_init=w_init, b_init=b_init, name='conv1_1')
+ >>> conv1 = Conv2d(conv1, 64, (3, 3), act=tf.nn.relu, padding='SAME', W_init=w_init, b_init=b_init, name='conv1_2')
+ >>> pool1 = MaxPool2d(conv1, (2, 2), padding='SAME', name='pool1')
+ >>> conv2 = Conv2d(pool1, 128, (3, 3), act=tf.nn.relu, padding='SAME', W_init=w_init, b_init=b_init, name='conv2_1')
+ >>> conv2 = Conv2d(conv2, 128, (3, 3), act=tf.nn.relu, padding='SAME', W_init=w_init, b_init=b_init, name='conv2_2')
+ >>> pool2 = MaxPool2d(conv2, (2, 2), padding='SAME', name='pool2')
+ """
+ assert len(strides) == 2, "len(strides) should be 2, Conv2d and Conv2dLayer are different."
+ if act is None:
+ act = tf.identity
+
+ try:
+ pre_channel = int(net.outputs.get_shape()[-1])
+ except: # if pre_channel is ?, it happens when using Spatial Transformer Net
+ pre_channel = 1
+ print("[warnings] unknow input channels, set to 1")
+ net = Conv2dLayer(
+ net,
+ act=act,
+ shape=[filter_size[0], filter_size[1], pre_channel, n_filter], # 32 features for each 5x5 patch
+ strides=[1, strides[0], strides[1], 1],
+ padding=padding,
+ W_init=W_init,
+ W_init_args=W_init_args,
+ b_init=b_init,
+ b_init_args=b_init_args,
+ use_cudnn_on_gpu=use_cudnn_on_gpu,
+ data_format=data_format,
+ name=name)
+ return net
+
+
+class Conv2dLayer(Layer):
+ """
+ The :class:`Conv2dLayer` class is a 2D CNN layer, see `tf.nn.conv2d `_.
+
+ Parameters
+ ----------
+ layer : a :class:`Layer` instance
+ The `Layer` class feeding into this layer.
+ act : activation function
+ The function that is applied to the layer activations.
+ shape : list of shape
+ shape of the filters, [filter_height, filter_width, in_channels, out_channels].
+ strides : a list of ints.
+ The stride of the sliding window for each dimension of input.\n
+ It Must be in the same order as the dimension specified with format.
+ padding : a string from: "SAME", "VALID".
+ The type of padding algorithm to use.
+ W_init : weights initializer
+ The initializer for initializing the weight matrix.
+ b_init : biases initializer or None
+ The initializer for initializing the bias vector. If None, skip biases.
+ W_init_args : dictionary
+ The arguments for the weights tf.get_variable().
+ b_init_args : dictionary
+ The arguments for the biases tf.get_variable().
+ use_cudnn_on_gpu : bool, default is None.
+ data_format : string "NHWC" or "NCHW", default is "NHWC"
+ name : a string or None
+ An optional name to attach to this layer.
+
+ Notes
+ ------
+ - shape = [h, w, the number of output channel of previous layer, the number of output channels]
+ - the number of output channel of a layer is its last dimension.
+
+ Examples
+ --------
+ >>> x = tf.placeholder(tf.float32, shape=[None, 28, 28, 1])
+ >>> network = tl.layers.InputLayer(x, name='input_layer')
+ >>> network = tl.layers.Conv2dLayer(network,
+ ... act = tf.nn.relu,
+ ... shape = [5, 5, 1, 32], # 32 features for each 5x5 patch
+ ... strides=[1, 1, 1, 1],
+ ... padding='SAME',
+ ... W_init=tf.truncated_normal_initializer(stddev=5e-2),
+ ... W_init_args={},
+ ... b_init = tf.constant_initializer(value=0.0),
+ ... b_init_args = {},
+ ... name ='cnn_layer1') # output: (?, 28, 28, 32)
+ >>> network = tl.layers.PoolLayer(network,
+ ... ksize=[1, 2, 2, 1],
+ ... strides=[1, 2, 2, 1],
+ ... padding='SAME',
+ ... pool = tf.nn.max_pool,
+ ... name ='pool_layer1',) # output: (?, 14, 14, 32)
+
+ >>> Without TensorLayer, you can implement 2d convolution as follow.
+ >>> W = tf.Variable(W_init(shape=[5, 5, 1, 32], ), name='W_conv')
+ >>> b = tf.Variable(b_init(shape=[32], ), name='b_conv')
+ >>> outputs = tf.nn.relu( tf.nn.conv2d(inputs, W,
+ ... strides=[1, 1, 1, 1],
+ ... padding='SAME') + b )
+ """
+
+ def __init__(
+ self,
+ layer=None,
+ act=tf.identity,
+ shape=[5, 5, 1, 100],
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ W_init=tf.truncated_normal_initializer(stddev=0.02),
+ b_init=tf.constant_initializer(value=0.0),
+ W_init_args={},
+ b_init_args={},
+ use_cudnn_on_gpu=None,
+ data_format=None,
+ name='cnn_layer',
+ ):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ print(" [TL] Conv2dLayer %s: shape:%s strides:%s pad:%s act:%s" % (self.name, str(shape), str(strides), padding, act.__name__))
+
+ with tf.variable_scope(name) as vs:
+ with tf.device('/cpu:0'):
+ W = tf.get_variable(name='W_conv2d', shape=shape, initializer=W_init, dtype=D_TYPE, **W_init_args)
+ if b_init:
+ with tf.device('/cpu:0'):
+ b = tf.get_variable(name='b_conv2d', shape=(shape[-1]), initializer=b_init, dtype=D_TYPE, **b_init_args)
+ self.outputs = act(
+ tf.nn.conv2d(self.inputs, W, strides=strides, padding=padding, use_cudnn_on_gpu=use_cudnn_on_gpu, data_format=data_format) + b)
+ else:
+ self.outputs = act(tf.nn.conv2d(self.inputs, W, strides=strides, padding=padding, use_cudnn_on_gpu=use_cudnn_on_gpu, data_format=data_format))
+
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+ self.all_layers.extend([self.outputs])
+ if b_init:
+ self.all_params.extend([W, b])
+ else:
+ self.all_params.extend([W])
+
+
+## Special activation
+class PReluLayer(Layer):
+ """
+ The :class:`PReluLayer` class is Parametric Rectified Linear layer.
+
+ Parameters
+ ----------
+ x : A `Tensor` with type `float`, `double`, `int32`, `int64`, `uint8`,
+ `int16`, or `int8`.
+ channel_shared : `bool`. Single weight is shared by all channels
+ a_init : alpha initializer, default zero constant.
+ The initializer for initializing the alphas.
+ a_init_args : dictionary
+ The arguments for the weights initializer.
+ name : A name for this activation op (optional).
+
+ References
+ -----------
+ - `Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification `_
+ """
+
+ def __init__(
+ self,
+ layer=None,
+ channel_shared=False,
+ a_init=tf.constant_initializer(value=0.0),
+ a_init_args={},
+ # restore = True,
+ name="prelu_layer"):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ print(" [TL] PReluLayer %s: channel_shared:%s" % (self.name, channel_shared))
+ if channel_shared:
+ w_shape = (1, )
+ else:
+ w_shape = int(self.inputs.get_shape()[-1])
+
+ # with tf.name_scope(name) as scope:
+ with tf.variable_scope(name) as vs:
+ with tf.device('/cpu:0'):
+ alphas = tf.get_variable(name='alphas', shape=w_shape, initializer=a_init, dtype=D_TYPE, **a_init_args)
+ try: ## TF 1.0
+ self.outputs = tf.nn.relu(self.inputs) + tf.multiply(alphas, (self.inputs - tf.abs(self.inputs))) * 0.5
+ except: ## TF 0.12
+ self.outputs = tf.nn.relu(self.inputs) + tf.mul(alphas, (self.inputs - tf.abs(self.inputs))) * 0.5
+
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+
+ self.all_layers.extend([self.outputs])
+ self.all_params.extend([alphas])
+
+
+## Dense layer
+class DenseLayer(Layer):
+ """
+ The :class:`DenseLayer` class is a fully connected layer.
+
+ Parameters
+ ----------
+ layer : a :class:`Layer` instance
+ The `Layer` class feeding into this layer.
+ n_units : int
+ The number of units of the layer.
+ act : activation function
+ The function that is applied to the layer activations.
+ W_init : weights initializer
+ The initializer for initializing the weight matrix.
+ b_init : biases initializer or None
+ The initializer for initializing the bias vector. If None, skip biases.
+ W_init_args : dictionary
+ The arguments for the weights tf.get_variable.
+ b_init_args : dictionary
+ The arguments for the biases tf.get_variable.
+ name : a string or None
+ An optional name to attach to this layer.
+
+ Examples
+ --------
+ >>> network = tl.layers.InputLayer(x, name='input_layer')
+ >>> network = tl.layers.DenseLayer(
+ ... network,
+ ... n_units=800,
+ ... act = tf.nn.relu,
+ ... W_init=tf.truncated_normal_initializer(stddev=0.1),
+ ... name ='relu_layer'
+ ... )
+
+ >>> Without TensorLayer, you can do as follow.
+ >>> W = tf.Variable(
+ ... tf.random_uniform([n_in, n_units], -1.0, 1.0), name='W')
+ >>> b = tf.Variable(tf.zeros(shape=[n_units]), name='b')
+ >>> y = tf.nn.relu(tf.matmul(inputs, W) + b)
+
+ Notes
+ -----
+ If the input to this layer has more than two axes, it need to flatten the
+ input by using :class:`FlattenLayer` in this case.
+ """
+
+ def __init__(
+ self,
+ layer=None,
+ n_units=100,
+ act=tf.identity,
+ W_init=tf.truncated_normal_initializer(stddev=0.1),
+ b_init=tf.constant_initializer(value=0.0),
+ W_init_args={},
+ b_init_args={},
+ name='dense_layer',
+ ):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ if self.inputs.get_shape().ndims != 2:
+ raise Exception("The input dimension must be rank 2, please reshape or flatten it")
+
+ n_in = int(self.inputs.get_shape()[-1])
+ self.n_units = n_units
+ print(" [TL] DenseLayer %s: %d %s" % (self.name, self.n_units, act.__name__))
+ with tf.variable_scope(name) as vs:
+ with tf.device('/cpu:0'):
+ W = tf.get_variable(name='W', shape=(n_in, n_units), initializer=W_init, dtype=D_TYPE, **W_init_args)
+ if b_init is not None:
+ try:
+ with tf.device('/cpu:0'):
+ b = tf.get_variable(name='b', shape=(n_units), initializer=b_init, dtype=D_TYPE, **b_init_args)
+ except: # If initializer is a constant, do not specify shape.
+ with tf.device('/cpu:0'):
+ b = tf.get_variable(name='b', initializer=b_init, dtype=D_TYPE, **b_init_args)
+ self.outputs = act(tf.matmul(self.inputs, W) + b)
+ else:
+ self.outputs = act(tf.matmul(self.inputs, W))
+
+ # Hint : list(), dict() is pass by value (shallow), without them, it is
+ # pass by reference.
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+ self.all_layers.extend([self.outputs])
+ if b_init is not None:
+ self.all_params.extend([W, b])
+ else:
+ self.all_params.extend([W])
+
+
+def get_shape(input_tensor):
+ static_shape = input_tensor.get_shape().as_list()
+ dynamic_shape = tf.unstack(tf.shape(input_tensor))
+ final_shapes = [shapes[1] if shapes[0] is None else shapes[0] for shapes in zip(static_shape, dynamic_shape)]
+ return final_shapes
+
+
+## Group Batch Normalization layer
+class GroupNormLayer(Layer):
+ """The :class:`GroupNormLayer` class is a for instance normalization.
+ The implementation is based on paper [Group Normalization](https://arxiv.org/pdf/1803.08494.pdf)
+ Parameters
+ -----------
+ layer : a :class:`Layer` instance
+ The `Layer` class feeding into this layer.
+ act : activation function.
+ epsilon : float
+ A small float number.
+ scale_init : beta initializer
+ The initializer for initializing beta
+ offset_init : gamma initializer
+ The initializer for initializing gamma
+ G: the group number
+ name : a string or None
+ An optional name to attach to this layer.
+ """
+ def __init__(
+ self,
+ layer=None,
+ act=tf.identity,
+ epsilon=1e-5,
+ scale_init=tf.constant_initializer(1.0),
+ offset_init=tf.constant_initializer(0.0),
+ G=32,
+ name='group_norm',
+ ):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ print(" [TL] GroupNormLayer %s: epsilon:%f act:%s" % (self.name, epsilon, act.__name__))
+ inputs_shape = get_shape(layer.outputs)
+ G = tf.minimum(G, inputs_shape[-1])
+ # [N, H, W, C] to [N, C, H, W]
+ temp_input = tf.transpose(self.inputs, [0, 3, 1, 2])
+ temp_input = tf.reshape(temp_input, [inputs_shape[0], G, inputs_shape[-1]//G, inputs_shape[1], inputs_shape[2]],
+ name='group_reshape1')
+ with tf.variable_scope(name) as vs:
+ mean, var = tf.nn.moments(temp_input, [2, 3, 4], keep_dims=True)
+ scale = tf.get_variable('scale', shape=[1, inputs_shape[-1], 1, 1], initializer=scale_init, dtype=D_TYPE)
+ offset = tf.get_variable('offset', shape=[1, inputs_shape[-1], 1, 1], initializer=offset_init, dtype=D_TYPE)
+ temp_input = (temp_input - mean) / tf.sqrt(var + epsilon)
+ temp_input = tf.reshape(temp_input, shape=[inputs_shape[0], inputs_shape[-1], inputs_shape[1], inputs_shape[2]],
+ name='group_reshape2')
+ self.outputs = scale * temp_input + offset
+ self.outputs = tf.transpose(self.outputs, [0, 2, 3, 1])
+ self.outputs = act(self.outputs)
+ variables = tf.get_collection(TF_GRAPHKEYS_VARIABLES, scope=vs.name)
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+ self.all_layers.extend([self.outputs])
+ self.all_params.extend(variables)
+
+
+if __name__ == '__main__':
+ import numpy as np
+ a = np.random.randn(5, 32, 32, 64)
+ b = tf.placeholder(dtype=tf.float32, name='placeholder_b', shape=(None, 32, 32, 64))
+ inputs_tl = tl.layers.InputLayer(b, name='inputs_layer')
+ gp = GroupNormLayer(layer=inputs_tl)
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/vgg16.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/vgg16.py
new file mode 100644
index 0000000000000000000000000000000000000000..550527910cd3e10ecec1dd3a064ba30eda09e5d7
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/vgg16.py
@@ -0,0 +1,209 @@
+#! /usr/bin/python
+# -*- coding: utf-8 -*-
+"""
+VGG-16 for ImageNet.
+Introduction
+----------------
+VGG is a convolutional neural network model proposed by K. Simonyan and A. Zisserman
+from the University of Oxford in the paper “Very Deep Convolutional Networks for
+Large-Scale Image Recognition” . The model achieves 92.7% top-5 test accuracy in ImageNet,
+which is a dataset of over 14 million images belonging to 1000 classes.
+Download Pre-trained Model
+----------------------------
+- Model weights in this example - vgg16_weights.npz : http://www.cs.toronto.edu/~frossard/post/vgg16/
+- Caffe VGG 16 model : https://gist.github.com/ksimonyan/211839e770f7b538e2d8#file-readme-md
+- Tool to convert the Caffe models to TensorFlow's : https://github.com/ethereon/caffe-tensorflow
+Note
+------
+- For simplified CNN layer see "Convolutional layer (Simplified)"
+in read the docs website.
+- When feeding other images to the model be sure to properly resize or crop them
+beforehand. Distorted images might end up being misclassified. One way of safely
+feeding images of multiple sizes is by doing center cropping, as shown in the
+
+The input image type is
+ from scipy.misc import imread, imresize
+ img1 = imread('data/laska.png', mode='RGB') # test data in github
+ img1 = imresize(img1, (224, 224))
+So the input image is three channels, and is RGB.
+
+following snippet:
+# >>> image_h, image_w, _ = np.shape(img)
+# >>> shorter_side = min(image_h, image_w)
+# >>> scale = 224. / shorter_side
+# >>> image_h, image_w = np.ceil([scale * image_h, scale * image_w]).astype('int32')
+# >>> img = imresize(img, (image_h, image_w))
+# >>> crop_x = (image_w - 224) / 2
+# >>> crop_y = (image_h - 224) / 2
+# >>> img = img[crop_y:crop_y+224,crop_x:crop_x+224,:]
+"""
+from npu_bridge.npu_init import *
+
+import tensorlayer as tl
+from tensorlayer.layers import *
+from scipy.misc import imread, imresize
+from nets.imagenet_classes import *
+import os
+
+
+def _conv_layers(net_in):
+ with tf.name_scope('preprocess'):
+ # Notice that we include a preprocessing layer that takes the RGB image
+ # with pixels values in the range of 0-255 and subtracts the mean image
+ # values (calculated over the entire ImageNet training set).
+ mean = tf.constant([123.68, 116.779, 103.939], dtype=tf.float32, shape=[1, 1, 1, 3], name='img_mean')
+ net_in.outputs = net_in.outputs - mean
+
+ # conv1
+ network = Conv2dLayer(
+ net_in,
+ act=tf.nn.relu,
+ shape=[3, 3, 3, 64], # 64 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv1_1')
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 64, 64], # 64 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv1_2')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool1')
+
+ # conv2
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 64, 128], # 128 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv2_1')
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 128, 128], # 128 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv2_2')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool2')
+
+ # conv3
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 128, 256], # 256 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv3_1')
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 256, 256], # 256 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv3_2')
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 256, 256], # 256 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv3_3')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool3')
+
+ # conv4
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 256, 512], # 512 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv4_1')
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 512, 512], # 512 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv4_2')
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 512, 512], # 512 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv4_3')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool4')
+
+ # conv5
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 512, 512], # 512 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv5_1')
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 512, 512], # 512 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv5_2')
+ network = Conv2dLayer(
+ network,
+ act=tf.nn.relu,
+ shape=[3, 3, 512, 512], # 512 features for each 3x3 patch
+ strides=[1, 1, 1, 1],
+ padding='SAME',
+ name='conv5_3')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool5')
+ return network
+
+
+def _fc_layers(net):
+ network = FlattenLayer(net, name='flatten')
+ network = DenseLayer(network, n_units=4096, act=tf.nn.relu, name='fc1_relu')
+ network = DenseLayer(network, n_units=4096, act=tf.nn.relu, name='fc2_relu')
+ network = DenseLayer(network, n_units=1000, act=tf.identity, name='fc3_relu')
+ return network
+
+
+def get_vgg16(x, sess=None, pretrained=True):
+ net_in = InputLayer(x, name='input')
+ net_cnn = _conv_layers(net_in) # simplified CNN APIs
+ network = _fc_layers(net_cnn)
+
+ if pretrained:
+ npz = np.load('../model_weights/vgg16_weights.npz')
+ params = []
+ for val in sorted(npz.items()):
+ print(" Loading %s" % str(val[1].shape))
+ params.append(val[1])
+ tl.files.assign_params(sess, params, network)
+ return network
+ else:
+ tl.layers.initialize_global_variables(sess)
+ return network
+
+
+if __name__ == '__main__':
+ DATA_PATH = '/home/aurora/workspaces2/PycharmProjects/tensorflow/tensorlayer/example/data'
+
+ x = tf.placeholder(tf.float32, [None, 224, 224, 3])
+ tfconfig = tf.ConfigProto(allow_soft_placement=True)
+ with tf.Session(config=npu_config_proto(config_proto=tfconfig)) as sess:
+ network = get_vgg16(x, sess, pretrained=True)
+ y = network.outputs
+ network.print_params()
+ network.print_layers()
+ img1 = imread(os.path.join(DATA_PATH, 'laska.png'), mode='RGB') # test data in github
+ img1 = imresize(img1, (224, 224))
+ probs = tf.nn.softmax(y)
+ start_time = time.time()
+ prob = sess.run(probs, feed_dict={x: [img1]})[0]
+ print(" End time : %.5ss" % (time.time() - start_time))
+ preds = (np.argsort(prob)[::-1])[0:5]
+ for p in preds:
+ print(class_names[p], prob[p])
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/nets/vgg19.py b/TensorFlow/contrib/cv/InsightFace_TF/nets/vgg19.py
new file mode 100644
index 0000000000000000000000000000000000000000..a5646cd648b229d3b7f1b1ab7211acdba055855c
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/nets/vgg19.py
@@ -0,0 +1,240 @@
+#! /usr/bin/python
+# -*- coding: utf-8 -*-
+"""
+VGG-19 for ImageNet.
+Pre-trained model in this example - VGG19 NPZ and
+trainable examples of VGG16/19 in TensorFlow can be found here:
+https://github.com/machrisaa/tensorflow-vgg
+For simplified CNN layer see "Convolutional layer (Simplified)"
+in read the docs website.
+"""
+from npu_bridge.npu_init import *
+
+import os
+import time
+
+import numpy as np
+import skimage
+import skimage.io
+import skimage.transform
+import tensorflow as tf
+from scipy.misc import imread, imresize
+import tensorlayer as tl
+from tensorlayer.layers import *
+from imagenet_classes import *
+
+
+DATA_PATH = '/home/aurora/workspaces2/PycharmProjects/tensorflow/tensorlayer/example/data'
+VGG_MEAN = [103.939, 116.779, 123.68]
+
+
+def load_image(path):
+ # load image
+ img = skimage.io.imread(path)
+ img = img / 255.0
+ if ((0 <= img).all() and (img <= 1.0).all()) is False:
+ raise Exception("image value should be [0, 1]")
+ # print "Original Image Shape: ", img.shape
+ # we crop image from center
+ short_edge = min(img.shape[:2])
+ yy = int((img.shape[0] - short_edge) / 2)
+ xx = int((img.shape[1] - short_edge) / 2)
+ crop_img = img[yy:yy + short_edge, xx:xx + short_edge]
+ # resize to 224, 224
+ resized_img = skimage.transform.resize(crop_img, (224, 224))
+ return resized_img
+
+
+def print_prob(prob):
+ synset = class_names
+ # print prob
+ pred = np.argsort(prob)[::-1]
+ # Get top1 label
+ top1 = synset[pred[0]]
+ print("Top1: ", top1, prob[pred[0]])
+ # Get top5 label
+ top5 = [(synset[pred[i]], prob[pred[i]]) for i in range(5)]
+ print("Top5: ", top5)
+ return top1
+
+
+def _Vgg19(rgb):
+ """
+ Build the VGG 19 Model
+ Parameters
+ -----------
+ rgb : rgb image placeholder [batch, height, width, 3] values scaled [0, 1]
+ """
+ start_time = time.time()
+ print("build model started")
+ rgb_scaled = rgb * 255.0
+ # Convert RGB to BGR
+ if tf.__version__ <= '0.11':
+ red, green, blue = tf.split(3, 3, rgb_scaled)
+ else: # TF 1.0
+ print(rgb_scaled)
+ red, green, blue = tf.split(rgb_scaled, 3, 3)
+ if red.get_shape().as_list()[1:] != [224, 224, 1]:
+ raise Exception("image size unmatch")
+ if green.get_shape().as_list()[1:] != [224, 224, 1]:
+ raise Exception("image size unmatch")
+ if blue.get_shape().as_list()[1:] != [224, 224, 1]:
+ raise Exception("image size unmatch")
+ if tf.__version__ <= '0.11':
+ bgr = tf.concat(3, [
+ blue - VGG_MEAN[0],
+ green - VGG_MEAN[1],
+ red - VGG_MEAN[2],
+ ])
+ else:
+ bgr = tf.concat(
+ [
+ blue - VGG_MEAN[0],
+ green - VGG_MEAN[1],
+ red - VGG_MEAN[2],
+ ], axis=3)
+ if bgr.get_shape().as_list()[1:] != [224, 224, 3]:
+ raise Exception("image size unmatch")
+ # input layer
+ net_in = InputLayer(bgr, name='input')
+ # conv1
+ network = Conv2dLayer(net_in, act=tf.nn.relu, shape=[3, 3, 3, 64], strides=[1, 1, 1, 1], padding='SAME', name='conv1_1')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 64, 64], strides=[1, 1, 1, 1], padding='SAME', name='conv1_2')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool1')
+ # conv2
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 64, 128], strides=[1, 1, 1, 1], padding='SAME', name='conv2_1')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 128, 128], strides=[1, 1, 1, 1], padding='SAME', name='conv2_2')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool2')
+ # conv3
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 128, 256], strides=[1, 1, 1, 1], padding='SAME', name='conv3_1')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 256, 256], strides=[1, 1, 1, 1], padding='SAME', name='conv3_2')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 256, 256], strides=[1, 1, 1, 1], padding='SAME', name='conv3_3')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 256, 256], strides=[1, 1, 1, 1], padding='SAME', name='conv3_4')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool3')
+ # conv4
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 256, 512], strides=[1, 1, 1, 1], padding='SAME', name='conv4_1')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 512, 512], strides=[1, 1, 1, 1], padding='SAME', name='conv4_2')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 512, 512], strides=[1, 1, 1, 1], padding='SAME', name='conv4_3')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 512, 512], strides=[1, 1, 1, 1], padding='SAME', name='conv4_4')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool4')
+ # conv5
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 512, 512], strides=[1, 1, 1, 1], padding='SAME', name='conv5_1')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 512, 512], strides=[1, 1, 1, 1], padding='SAME', name='conv5_2')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 512, 512], strides=[1, 1, 1, 1], padding='SAME', name='conv5_3')
+ network = Conv2dLayer(network, act=tf.nn.relu, shape=[3, 3, 512, 512], strides=[1, 1, 1, 1], padding='SAME', name='conv5_4')
+ network = PoolLayer(network, ksize=[1, 2, 2, 1], strides=[1, 2, 2, 1], padding='SAME', pool=tf.nn.max_pool, name='pool5')
+ # fc 6~8
+ network = FlattenLayer(network, name='flatten')
+ network = DenseLayer(network, n_units=4096, act=tf.nn.relu, name='fc6')
+ network = DenseLayer(network, n_units=4096, act=tf.nn.relu, name='fc7')
+ network = DenseLayer(network, n_units=1000, act=tf.identity, name='fc8')
+ print("build model finished: %fs" % (time.time() - start_time))
+ return network
+
+
+def _Vgg19_simple_api(rgb):
+ """
+ Build the VGG 19 Model
+ Parameters
+ -----------
+ rgb : rgb image placeholder [batch, height, width, 3] values scaled [0, 1]
+ """
+ start_time = time.time()
+ print("build model started")
+ rgb_scaled = rgb * 255.0
+ # Convert RGB to BGR
+ if tf.__version__ <= '0.11':
+ red, green, blue = tf.split(3, 3, rgb_scaled)
+ else: # TF 1.0
+ print(rgb_scaled)
+ red, green, blue = tf.split(rgb_scaled, 3, 3)
+ if red.get_shape().as_list()[1:] != [224, 224, 1]:
+ raise Exception("image size unmatch")
+ if green.get_shape().as_list()[1:] != [224, 224, 1]:
+ raise Exception("image size unmatch")
+ if blue.get_shape().as_list()[1:] != [224, 224, 1]:
+ raise Exception("image size unmatch")
+ if tf.__version__ <= '0.11':
+ bgr = tf.concat(3, [
+ blue - VGG_MEAN[0],
+ green - VGG_MEAN[1],
+ red - VGG_MEAN[2],
+ ])
+ else:
+ bgr = tf.concat(
+ [
+ blue - VGG_MEAN[0],
+ green - VGG_MEAN[1],
+ red - VGG_MEAN[2],
+ ], axis=3)
+ if bgr.get_shape().as_list()[1:] != [224, 224, 3]:
+ raise Exception("image size unmatch")
+ # input layer
+ net_in = InputLayer(bgr, name='input')
+ # conv1
+ network = Conv2d(net_in, n_filter=64, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv1_1')
+ network = Conv2d(network, n_filter=64, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv1_2')
+ network = MaxPool2d(network, filter_size=(2, 2), strides=(2, 2), padding='SAME', name='pool1')
+ # conv2
+ network = Conv2d(network, n_filter=128, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv2_1')
+ network = Conv2d(network, n_filter=128, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv2_2')
+ network = MaxPool2d(network, filter_size=(2, 2), strides=(2, 2), padding='SAME', name='pool2')
+ # conv3
+ network = Conv2d(network, n_filter=256, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv3_1')
+ network = Conv2d(network, n_filter=256, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv3_2')
+ network = Conv2d(network, n_filter=256, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv3_3')
+ network = Conv2d(network, n_filter=256, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv3_4')
+ network = MaxPool2d(network, filter_size=(2, 2), strides=(2, 2), padding='SAME', name='pool3')
+ # conv4
+ network = Conv2d(network, n_filter=512, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv4_1')
+ network = Conv2d(network, n_filter=512, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv4_2')
+ network = Conv2d(network, n_filter=512, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv4_3')
+ network = Conv2d(network, n_filter=512, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv4_4')
+ network = MaxPool2d(network, filter_size=(2, 2), strides=(2, 2), padding='SAME', name='pool4')
+ # conv5
+ network = Conv2d(network, n_filter=512, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv5_1')
+ network = Conv2d(network, n_filter=512, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv5_2')
+ network = Conv2d(network, n_filter=512, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv5_3')
+ network = Conv2d(network, n_filter=512, filter_size=(3, 3), strides=(1, 1), act=tf.nn.relu, padding='SAME', name='conv5_4')
+ network = MaxPool2d(network, filter_size=(2, 2), strides=(2, 2), padding='SAME', name='pool5')
+ # fc 6~8
+ network = FlattenLayer(network, name='flatten')
+ network = DenseLayer(network, n_units=4096, act=tf.nn.relu, name='fc6')
+ network = DenseLayer(network, n_units=4096, act=tf.nn.relu, name='fc7')
+ network = DenseLayer(network, n_units=1000, act=tf.identity, name='fc8')
+ print("build model finished: %fs" % (time.time() - start_time))
+ return network
+
+
+def get_vgg19(inputs, sess=None, pretrained=True):
+ network = _Vgg19(inputs)
+ if pretrained:
+ vgg19_npy_path = "../model_weights/vgg19.npy"
+ npz = np.load(vgg19_npy_path, encoding='latin1').item()
+ params = []
+ for val in sorted(npz.items()):
+ W = np.asarray(val[1][0])
+ b = np.asarray(val[1][1])
+ print(" Loading %s: %s, %s" % (val[0], W.shape, b.shape))
+ params.extend([W, b])
+ print("Restoring model from npz file")
+ tl.files.assign_params(sess, params, network)
+ return network
+ else:
+ tl.layers.initialize_global_variables(sess)
+ return network
+
+
+if __name__ == '__main__':
+ sess = tf.InteractiveSession(config=npu_config_proto())
+ x = tf.placeholder("float", [None, 224, 224, 3])
+ network = get_vgg19(x, sess)
+ y = network.outputs
+ probs = tf.nn.softmax(y, name="prob")
+ img1 = load_image(os.path.join(DATA_PATH, "tiger.jpeg")) # test data in github
+ img1 = img1.reshape((1, 224, 224, 3))
+ start_time = time.time()
+ prob = sess.run(probs, feed_dict={x: img1})
+ print("End time : %.5ss" % (time.time() - start_time))
+
+ print_prob(prob[0])
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/requirements.txt b/TensorFlow/contrib/cv/InsightFace_TF/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..74202e087c0bac8702f38da9163d52ad98bdfae6
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/requirements.txt
@@ -0,0 +1,92 @@
+absl-py==0.12.0
+astor==0.8.1
+astunparse==1.6.3
+attrs==20.3.0
+Automat==20.2.0
+Babel==2.9.1
+cached-property==1.5.2
+cachetools==4.2.2
+certifi==2020.12.5
+cffi==1.14.5
+chardet==4.0.0
+constantly==15.1.0
+cryptography==3.4.6
+cssselect==1.1.0
+cycler==0.10.0
+decorator==5.1.0
+easygui===0.98.0-unreleased
+Faker==8.2.1
+flatbuffers==1.12
+gast==0.3.3
+google-auth==1.30.0
+google-auth-oauthlib==0.4.4
+google-pasta==0.2.0
+grpcio==1.32.0
+h5py==2.10.0
+hyperlink==21.0.0
+idna==2.10
+importlib-metadata==4.0.1
+incremental==17.5.0
+itemadapter==0.2.0
+itemloaders==1.0.4
+jmespath==0.10.0
+joblib==1.0.1
+Keras==2.2.5
+Keras-Applications==1.0.8
+Keras-Preprocessing==1.1.2
+keras-resnet==0.2.0
+kiwisolver==1.3.1
+lxml==4.6.2
+Markdown==3.3.4
+mxnet==1.9.0
+numpy==1.19.3
+oauthlib==3.1.0
+opencv-python==4.6.0.66
+opt-einsum==3.3.0
+paddlepaddle==2.2.0
+
+parsel==1.6.0
+Pillow==8.1.1
+Protego==0.1.16
+protobuf==3.15.8
+pyasn1==0.4.8
+pyasn1-modules==0.2.8
+pycparser==2.20
+PyDispatcher==2.0.5
+PyHamcrest==2.0.2
+pyOpenSSL==20.0.1
+pyparsing==2.4.7
+python-dateutil==2.8.1
+pytz==2021.1
+PyYAML==5.4.1
+queuelib==1.5.0
+requests==2.25.1
+requests-oauthlib==1.3.0
+rsa==4.7.2
+scikit-learn==0.24.2
+scipy==1.6.3
+Scrapy==2.4.1
+service-identity==18.1.0
+six==1.15.0
+sklearn==0.0
+tensorboard==1.14.0
+tensorboard-data-server==0.6.0
+tensorboard-plugin-wit==1.8.0
+tensorflow==1.14.0
+tensorflow-estimator==1.14.0
+tensorflow-gpu==1.14.0
+tensorlayer==1.7.0
+termcolor==1.1.0
+text-unidecode==1.3
+threadpoolctl==2.1.0
+torch==1.8.1
+torchtext==0.9.1
+tqdm==4.61.0
+Twisted==20.3.0
+typing-extensions==3.7.4.3
+urllib3==1.26.4
+w3lib==1.22.0
+Werkzeug==1.0.1
+wrapt==1.12.1
+zipp==3.4.1
+zope.interface==5.2.0
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/README.md b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..3b4a360a7b0e1ea8593259b1fd0a6d58b5cf1da9
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/README.md
@@ -0,0 +1 @@
+##### Test max batch size
\ No newline at end of file
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/__init__.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/gluon_batchsize_test.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/gluon_batchsize_test.py
new file mode 100644
index 0000000000000000000000000000000000000000..761fc8a7e73de2817de7734a3025bc0bc24d0feb
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/gluon_batchsize_test.py
@@ -0,0 +1,62 @@
+from npu_bridge.npu_init import *
+from mxnet import gluon
+import mxnet as mx
+from mxnet import ndarray as nd
+import utils_final as utils
+import mxnet.gluon.nn as nn
+from mxnet import init
+import os
+from mxnet import initializer
+from mxnet.gluon.block import HybridBlock
+
+
+def prelu():
+ pass
+
+
+def inference():
+ net = gluon.nn.Sequential()
+ with net.name_scope():
+ net.add(nn.Conv2D(channels=64, kernel_size=3, padding=1))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+ # net.add(mx.sym.LeakyReLU(data=net, act_type='prelu', name='prelu1'))
+ net.add(nn.Conv2D(channels=64, kernel_size=3, padding=1))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+ net.add(nn.Conv2D(channels=64, kernel_size=3, padding=1, strides=2))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+
+ net.add(nn.Conv2D(channels=128, kernel_size=3, padding=1))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+ net.add(nn.Conv2D(channels=128, kernel_size=3, padding=1))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+ net.add(nn.Conv2D(channels=128, kernel_size=3, padding=1, strides=2))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+
+ net.add(nn.Conv2D(channels=256, kernel_size=3, padding=1))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+ net.add(nn.Conv2D(channels=256, kernel_size=3, padding=1))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+ net.add(nn.Conv2D(channels=256, kernel_size=3, padding=1, strides=2))
+ net.add(nn.BatchNorm(axis=1, center=True, scale=True))
+
+ net.add(nn.Flatten())
+ net.add(nn.Dense(10))
+ return net
+
+
+if __name__ == '__main__':
+ # without prelu and bn 7000< max batch size <8000
+ # with bn only 3000< max batch size <4000
+ os.environ["CUDA_VISIBLE_DEVICES"] = "0"
+ batch_size = 3000
+ train_data, test_data = utils.load_data_mnist(batch_size=batch_size)
+ ctx = utils.try_gpu()
+ net = inference()
+ print(net)
+ net.initialize(ctx=ctx, init=init.Xavier())
+ softmax_cross_entropy = gluon.loss.SoftmaxCrossEntropyLoss()
+ trainer = gluon.Trainer(net.collect_params(), 'sgd', {'learning_rate': 0.01})
+ utils.train(train_data, test_data, net, softmax_cross_entropy, trainer, ctx, num_epochs=10)
+
+
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/mxnet_batchsize_test.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/mxnet_batchsize_test.py
new file mode 100644
index 0000000000000000000000000000000000000000..592c3e41a763ccbaed755dc48cd717ed2d876564
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/mxnet_batchsize_test.py
@@ -0,0 +1,68 @@
+from npu_bridge.npu_init import *
+import mxnet as mx
+import mxnet.ndarray as nd
+import os
+
+
+if __name__ == '__main__':
+ # without bn and prelu max batchsize (40000, 50000)
+ # with bn max batchsize (20000, 30000)
+ # with prelu batchsize (20000, 30000)
+ # with bn and prelu max batchsize (10000, 20000)
+ os.environ["CUDA_VISIBLE_DEVICES"] = "0"
+ batch_size = 10000
+ mnist = mx.test_utils.get_mnist()
+ print(mnist['train_data'].shape)
+ train_iter = mx.io.NDArrayIter(mnist['train_data'], mnist['train_label'], batch_size, shuffle=True)
+
+ # inference
+ data = mx.sym.var('data')
+ # first conv layer
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), num_filter=64)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn1')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul1')
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), num_filter=64)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn2')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul2')
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), stride=(2, 2), num_filter=64)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn3')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul3')
+
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), num_filter=128)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn4')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul4')
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), num_filter=128)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn5')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul5')
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), stride=(2, 2), num_filter=128)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn6')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul6')
+
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), num_filter=256)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn7')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul7')
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), num_filter=256)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn8')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul8')
+ net = mx.sym.Convolution(data=data, kernel=(3, 3), stride=(2, 2), num_filter=256)
+ net = mx.sym.BatchNorm(data=net, fix_gamma=False, eps=2e-5, name='_bn9')
+ net = mx.sym.LeakyReLU(data=net, act_type='prelu', name='_preul9')
+
+ flatten = mx.sym.flatten(data=net)
+ # MNIST has 10 classes
+ fc3 = mx.sym.FullyConnected(data=flatten, num_hidden=10)
+ # Softmax with cross entropy loss
+ mlp = mx.sym.SoftmaxOutput(data=fc3, name='softmax')
+
+ import logging
+
+ logging.getLogger().setLevel(logging.DEBUG) # logging to stdout
+ # create a trainable module on GPU
+ mlp_model = mx.mod.Module(symbol=mlp, context=mx.gpu())
+ mlp_model.fit(train_iter, # train data
+ optimizer='sgd', # use SGD to train
+ optimizer_params={'learning_rate': 0.1}, # use fixed learning rate
+ eval_metric='acc', # report accuracy during training
+ batch_end_callback=mx.callback.Speedometer(batch_size, 100),
+ # output progress for each 100 data batches
+ num_epoch=10) # train for at most 10 dataset passes
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/resnet_slim_benchmark.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/resnet_slim_benchmark.py
new file mode 100644
index 0000000000000000000000000000000000000000..017aae37a60499fdfc8e4223d1356737a21dca60
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/resnet_slim_benchmark.py
@@ -0,0 +1,28 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorflow.contrib.slim.nets as nets
+import numpy as np
+
+
+slim = tf.contrib.slim
+resnet = nets.resnet_v1
+
+if __name__ == '__main__':
+ output_shape = 85164
+ batch_size = 64
+ image = tf.placeholder(name='input_x', shape=[None, 224, 224, 3], dtype=tf.float32)
+ labels = tf.placeholder(name='input_label', shape=[None, output_shape], dtype=tf.float32)
+ with slim.arg_scope(nets.resnet_utils.resnet_arg_scope()):
+ resnet_50, end_points = resnet.resnet_v1_50(inputs=image, num_classes=output_shape, scope='resnet_v1_50')
+ prob = tf.squeeze(resnet_50, axis=[1, 2])
+ probabilities = tf.reduce_mean(tf.nn.softmax(prob, dim=-1))
+ losses = tf.norm(tf.subtract(probabilities, labels))
+ train_op = tf.train.AdamOptimizer(learning_rate=0.0001).minimize(losses)
+ sess = tf.Session(config=npu_config_proto())
+ saver = tf.train.Saver()
+ sess.run(tf.global_variables_initializer())
+ while True:
+ datasets = np.random.randn(batch_size, 224, 224, 3).astype(np.float32)
+ datasets_labels = np.random.randn(batch_size, output_shape).astype(np.float32)
+ losses_val, _ = sess.run([losses, train_op], feed_dict={image: datasets, labels: datasets_labels})
+ print(losses_val)
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/resnet_tl_benchmark.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/resnet_tl_benchmark.py
new file mode 100644
index 0000000000000000000000000000000000000000..7c9afe3fa52347cec5b0959771b26cfe842e98d1
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/resnet_tl_benchmark.py
@@ -0,0 +1,30 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorflow.contrib.slim.nets as nets
+import numpy as np
+from nets.resnet import get_resnet
+
+
+slim = tf.contrib.slim
+resnet = nets.resnet_v1
+
+if __name__ == '__main__':
+ output_shape = 85164
+ batch_size = 128
+ image = tf.placeholder(name='input_x', shape=[None, 224, 224, 3], dtype=tf.float32)
+ labels = tf.placeholder(name='input_label', shape=[None, output_shape], dtype=tf.float32)
+ with slim.arg_scope(nets.resnet_utils.resnet_arg_scope()):
+ nets = get_resnet(image, output_shape, 50, type='resnet', sess=None, pretrained=False)
+ print(nets.outputs)
+ probabilities = tf.reduce_mean(tf.nn.softmax(nets.outputs, dim=-1))
+ print(probabilities)
+ losses = tf.norm(tf.subtract(probabilities, labels))
+ train_op = tf.train.AdamOptimizer(learning_rate=0.0001).minimize(losses)
+ sess = tf.Session(config=npu_config_proto())
+ saver = tf.train.Saver()
+ sess.run(tf.global_variables_initializer())
+ while True:
+ datasets = np.random.randn(batch_size, 224, 224, 3).astype(np.float32)
+ datasets_labels = np.random.randn(batch_size, output_shape).astype(np.float32)
+ losses_val, _ = sess.run([losses, train_op], feed_dict={image: datasets, labels: datasets_labels})
+ print(losses_val)
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/tensorlayer_batchsize_test.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/tensorlayer_batchsize_test.py
new file mode 100644
index 0000000000000000000000000000000000000000..7b9a3883653fc75266a3a8e6d519f5fb4eb6d428
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/tensorlayer_batchsize_test.py
@@ -0,0 +1,89 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+import os
+
+
+def inference(x):
+ w_init_method = tf.contrib.layers.xavier_initializer(uniform=True)
+ # define the network
+ network = tl.layers.InputLayer(x, name='input')
+ network = tl.layers.Conv2d(network, n_filter=64, filter_size=(3, 3), strides=(1, 1), padding='SAME', act=None,
+ W_init=w_init_method, name='conv1_1')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn1')
+ network = tl.layers.PReluLayer(network, name='prelu1')
+ network = tl.layers.Conv2d(network, n_filter=64, filter_size=(3, 3), strides=(1, 1), padding='SAME', act=None,
+ W_init=w_init_method, name='conv1_2')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn2')
+ network = tl.layers.PReluLayer(network, name='prelu2')
+ network = tl.layers.Conv2d(network, n_filter=64, filter_size=(3, 3), strides=(2, 2), padding='SAME', act=None,
+ W_init=w_init_method, name='conv1_3')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn3')
+ network = tl.layers.PReluLayer(network, name='prelu3')
+
+ network = tl.layers.Conv2d(network, n_filter=128, filter_size=(3, 3), strides=(1, 1), padding='SAME', act=None,
+ W_init=w_init_method, name='conv2_1')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn4')
+ network = tl.layers.PReluLayer(network, name='prelu4')
+
+ network = tl.layers.Conv2d(network, n_filter=128, filter_size=(3, 3), strides=(1, 1), padding='SAME', act=None,
+ W_init=w_init_method, name='conv2_2')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn5')
+ network = tl.layers.PReluLayer(network, name='prelu5')
+ network = tl.layers.Conv2d(network, n_filter=128, filter_size=(3, 3), strides=(2, 2), padding='SAME', act=None,
+ W_init=w_init_method, name='conv2_3')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn6')
+ network = tl.layers.PReluLayer(network, name='prelu6')
+
+ network = tl.layers.Conv2d(network, n_filter=256, filter_size=(3, 3), strides=(1, 1), padding='SAME', act=None,
+ W_init=w_init_method, name='conv3_1')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn7')
+ network = tl.layers.PReluLayer(network, name='prelu7')
+ network = tl.layers.Conv2d(network, n_filter=256, filter_size=(3, 3), strides=(1, 1), padding='SAME', act=None,
+ W_init=w_init_method, name='conv3_2')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn8')
+ network = tl.layers.PReluLayer(network, name='prelu8')
+ network = tl.layers.Conv2d(network, n_filter=256, filter_size=(3, 3), strides=(2, 2), padding='SAME', act=None,
+ W_init=w_init_method, name='conv3_3')
+ network = tl.layers.BatchNormLayer(network, act=tf.identity, is_train=True, name='bn9')
+ network = tl.layers.PReluLayer(network, name='prelu9')
+
+ network = tl.layers.FlattenLayer(network, name='flatten')
+ network = tl.layers.DenseLayer(network, 10)
+
+ return network.outputs
+
+
+if __name__ == '__main__':
+ # without bn prelu 8000< max batch size <9000
+ # with bn only 5000< max batch size <6000
+ # with prelu only 3000< max batch size <4000
+ # with bn and prelu 2000< max batch size <3000
+ os.environ["CUDA_VISIBLE_DEVICES"] = "0"
+ batch_size = 2000
+ n_epoch = 10
+ # prepare data
+ X_train, y_train, X_val, y_val, X_test, y_test = tl.files.load_mnist_dataset(shape=(-1, 28, 28, 1))
+ # define placeholder
+ x = tf.placeholder(tf.float32, shape=[None, 28, 28, 1], name='x')
+ y_ = tf.placeholder(tf.int64, shape=[None], name='y_')
+
+ output = inference(x)
+ cost = tl.cost.cross_entropy(output, y_, 'cost')
+ train_op = tf.train.GradientDescentOptimizer(learning_rate=0.1).minimize(cost)
+
+ sess = tf.Session(config=npu_config_proto())
+ tl.layers.initialize_global_variables(sess)
+
+ correct_prediction = tf.equal(tf.argmax(output, 1), y_)
+ acc = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))
+
+ for epoch in range(n_epoch):
+ train_loss, train_acc, n_batch = 0, 0, 0
+ for X_train_a, y_train_a in tl.iterate.minibatches(X_train, y_train, batch_size, shuffle=True):
+ feed_dict = {x: X_train_a, y_: y_train_a}
+ _, err, ac = sess.run([train_op, cost, acc], feed_dict=feed_dict)
+ train_loss += err
+ train_acc += ac
+ n_batch += 1
+ print("epoch %d, train acc: %f" % (epoch, (train_acc / n_batch)))
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/utils_final.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/utils_final.py
new file mode 100644
index 0000000000000000000000000000000000000000..bb65184e4bb7d837f8dd108a30eac834392b7b03
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/utils_final.py
@@ -0,0 +1,386 @@
+from npu_bridge.npu_init import *
+from math import exp
+from mxnet import gluon
+from mxnet import autograd
+from mxnet import nd
+from mxnet import image
+from mxnet.gluon import nn
+import mxnet as mx
+import numpy as np
+from time import time
+import matplotlib.pyplot as plt
+import random
+
+
+class DataLoader(object):
+ """similiar to gluon.data.DataLoader, but might be faster.
+
+ The main difference this data loader tries to read more exmaples each
+ time. But the limits are 1) all examples in dataset have the same shape, 2)
+ data transfomer needs to process multiple examples at each time
+ """
+
+ def __init__(self, dataset, batch_size, shuffle, transform=None):
+ self.dataset = dataset
+ self.batch_size = batch_size
+ self.shuffle = shuffle
+ self.transform = transform
+
+ def __iter__(self):
+ data = self.dataset[:]
+ X = data[0]
+ y = nd.array(data[1])
+ n = X.shape[0]
+ if self.shuffle:
+ idx = np.arange(n)
+ np.random.shuffle(idx)
+ X = nd.array(X.asnumpy()[idx])
+ y = nd.array(y.asnumpy()[idx])
+
+ for i in range(n // self.batch_size):
+ if self.transform is not None:
+ yield self.transform(X[i * self.batch_size:(i + 1) * self.batch_size],
+ y[i * self.batch_size:(i + 1) * self.batch_size])
+ else:
+ yield (X[i * self.batch_size:(i + 1) * self.batch_size],
+ y[i * self.batch_size:(i + 1) * self.batch_size])
+
+ def __len__(self):
+ return len(self.dataset) // self.batch_size
+
+
+def load_data_fashion_mnist(batch_size, resize=None, root="~/.mxnet/datasets/fashion-mnist"):
+ """download the fashion mnist dataest and then load into memory"""
+
+ def transform_mnist(data, label):
+ # Transform a batch of examples.
+ if resize:
+ n = data.shape[0]
+ new_data = nd.zeros((n, resize, resize, data.shape[3]))
+ for i in range(n):
+ new_data[i] = image.imresize(data[i], resize, resize)
+ data = new_data
+ # change data from batch x height x width x channel to batch x channel x height x width
+ return nd.transpose(data.astype('float32'), (0, 3, 1, 2)) / 255, label.astype('float32')
+
+ mnist_train = gluon.data.vision.FashionMNIST(root=root, train=True, transform=None)
+ mnist_test = gluon.data.vision.FashionMNIST(root=root, train=False, transform=None)
+ # Transform later to avoid memory explosion.
+ train_data = DataLoader(mnist_train, batch_size, shuffle=True, transform=transform_mnist)
+ test_data = DataLoader(mnist_test, batch_size, shuffle=False, transform=transform_mnist)
+ return (train_data, test_data)
+
+
+def load_data_mnist(batch_size, resize=None, root="~/.mxnet/datasets/mnist"):
+ """download the fashion mnist dataest and then load into memory"""
+
+ def transform_mnist(data, label):
+ # Transform a batch of examples.
+ if resize:
+ n = data.shape[0]
+ new_data = nd.zeros((n, resize, resize, data.shape[3]))
+ for i in range(n):
+ new_data[i] = image.imresize(data[i], resize, resize)
+ data = new_data
+ # change data from batch x height x width x channel to batch x channel x height x width
+ return nd.transpose(data.astype('float32'), (0, 3, 1, 2)) / 255, label.astype('float32')
+
+ mnist_train = gluon.data.vision.MNIST(root=root, train=True, transform=None)
+ mnist_test = gluon.data.vision.MNIST(root=root, train=False, transform=None)
+ # Transform later to avoid memory explosion.
+ train_data = DataLoader(mnist_train, batch_size, shuffle=True, transform=transform_mnist)
+ test_data = DataLoader(mnist_test, batch_size, shuffle=False, transform=transform_mnist)
+ return (train_data, test_data)
+
+
+def try_gpu():
+ """If GPU is available, return mx.gpu(0); else return mx.cpu()"""
+ try:
+ ctx = mx.gpu()
+ _ = nd.array([0], ctx=ctx)
+ except:
+ ctx = mx.cpu()
+ return ctx
+
+
+def try_all_gpus():
+ """Return all available GPUs, or [mx.gpu()] if there is no GPU"""
+ ctx_list = []
+ try:
+ for i in range(16):
+ ctx = mx.gpu(i)
+ _ = nd.array([0], ctx=ctx)
+ ctx_list.append(ctx)
+ except:
+ pass
+ if not ctx_list:
+ ctx_list = [mx.cpu()]
+ return ctx_list
+
+
+def SGD(params, lr):
+ for param in params:
+ param[:] = param - lr * param.grad
+
+
+def accuracy(output, label):
+ return nd.mean(output.argmax(axis=1) == label).asscalar()
+
+
+def _get_batch(batch, ctx):
+ """return data and label on ctx"""
+ if isinstance(batch, mx.io.DataBatch):
+ data = batch.data[0]
+ label = batch.label[0]
+ else:
+ data, label = batch
+ return (gluon.utils.split_and_load(data, ctx),
+ gluon.utils.split_and_load(label, ctx),
+ data.shape[0])
+
+
+def evaluate_accuracy(data_iterator, net, ctx=[mx.cpu()]):
+ if isinstance(ctx, mx.Context):
+ ctx = [ctx]
+ acc = nd.array([0])
+ n = 0.
+ if isinstance(data_iterator, mx.io.MXDataIter):
+ data_iterator.reset()
+ for batch in data_iterator:
+ data, label, batch_size = _get_batch(batch, ctx)
+ for X, y in zip(data, label):
+ acc += nd.sum(net(X).argmax(axis=1) == y).copyto(mx.cpu())
+ n += y.size
+ acc.wait_to_read() # don't push too many operators into backend
+ return acc.asscalar() / n
+
+
+def train(train_data, test_data, net, loss, trainer, ctx, num_epochs, print_batches=None):
+ """Train a network"""
+ print("Start training on ", ctx)
+ if isinstance(ctx, mx.Context):
+ ctx = [ctx]
+ for epoch in range(num_epochs):
+ train_loss, train_acc, n, m = 0.0, 0.0, 0.0, 0.0
+ if isinstance(train_data, mx.io.MXDataIter):
+ train_data.reset()
+ start = time()
+ for i, batch in enumerate(train_data):
+ data, label, batch_size = _get_batch(batch, ctx)
+ losses = []
+ with autograd.record():
+ outputs = [net(X) for X in data]
+ losses = [loss(yhat, y) for yhat, y in zip(outputs, label)]
+ for l in losses:
+ l.backward()
+ train_acc += sum([(yhat.argmax(axis=1) == y).sum().asscalar()
+ for yhat, y in zip(outputs, label)])
+ train_loss += sum([l.sum().asscalar() for l in losses])
+ trainer.step(batch_size)
+ n += batch_size
+ m += sum([y.size for y in label])
+ if print_batches and (i + 1) % print_batches == 0:
+ print("Batch %d. Loss: %f, Train acc %f" % (
+ n, train_loss / n, train_acc / m
+ ))
+
+ test_acc = evaluate_accuracy(test_data, net, ctx)
+ print("Epoch %d. Loss: %.3f, Train acc %.2f, Test acc %.2f, Time %.1f sec" % (
+ epoch, train_loss / n, train_acc / m, test_acc, time() - start
+ ))
+
+
+class Residual(nn.HybridBlock):
+ def __init__(self, channels, same_shape=True, **kwargs):
+ super(Residual, self).__init__(**kwargs)
+ self.same_shape = same_shape
+ with self.name_scope():
+ strides = 1 if same_shape else 2
+ self.conv1 = nn.Conv2D(channels, kernel_size=3, padding=1,
+ strides=strides)
+ self.bn1 = nn.BatchNorm()
+ self.conv2 = nn.Conv2D(channels, kernel_size=3, padding=1)
+ self.bn2 = nn.BatchNorm()
+ if not same_shape:
+ self.conv3 = nn.Conv2D(channels, kernel_size=1,
+ strides=strides)
+
+ def hybrid_forward(self, F, x):
+ out = F.relu(self.bn1(self.conv1(x)))
+ out = self.bn2(self.conv2(out))
+ if not self.same_shape:
+ x = self.conv3(x)
+ return F.relu(out + x)
+
+
+def resnet18(num_classes):
+ net = nn.HybridSequential()
+ with net.name_scope():
+ net.add(
+ nn.BatchNorm(),
+ nn.Conv2D(64, kernel_size=3, strides=1),
+ nn.MaxPool2D(pool_size=3, strides=2),
+ Residual(64),
+ Residual(64),
+ Residual(128, same_shape=False),
+ Residual(128),
+ Residual(256, same_shape=False),
+ Residual(256),
+ nn.GlobalAvgPool2D(),
+ nn.Dense(num_classes)
+ )
+ return net
+
+
+def show_images(imgs, nrows, ncols, figsize=None):
+ """plot a list of images"""
+ if not figsize:
+ figsize = (ncols, nrows)
+ _, figs = plt.subplots(nrows, ncols, figsize=figsize)
+ for i in range(nrows):
+ for j in range(ncols):
+ figs[i][j].imshow(imgs[i * ncols + j].asnumpy())
+ figs[i][j].axes.get_xaxis().set_visible(False)
+ figs[i][j].axes.get_yaxis().set_visible(False)
+ plt.show()
+
+
+def data_iter_random(corpus_indices, batch_size, num_steps, ctx=None):
+ """Sample mini-batches in a random order from sequential data."""
+ # Subtract 1 because label indices are corresponding input indices + 1.
+ num_examples = (len(corpus_indices) - 1) // num_steps
+ epoch_size = num_examples // batch_size
+ # Randomize samples.
+ example_indices = list(range(num_examples))
+ random.shuffle(example_indices)
+
+ def _data(pos):
+ return corpus_indices[pos: pos + num_steps]
+
+ for i in range(epoch_size):
+ # Read batch_size random samples each time.
+ i = i * batch_size
+ batch_indices = example_indices[i: i + batch_size]
+ data = nd.array(
+ [_data(j * num_steps) for j in batch_indices], ctx=ctx)
+ label = nd.array(
+ [_data(j * num_steps + 1) for j in batch_indices], ctx=ctx)
+ yield data, label
+
+
+def data_iter_consecutive(corpus_indices, batch_size, num_steps, ctx=None):
+ """Sample mini-batches in a consecutive order from sequential data."""
+ corpus_indices = nd.array(corpus_indices, ctx=ctx)
+ data_len = len(corpus_indices)
+ batch_len = data_len // batch_size
+
+ indices = corpus_indices[0: batch_size * batch_len].reshape((
+ batch_size, batch_len))
+ # Subtract 1 because label indices are corresponding input indices + 1.
+ epoch_size = (batch_len - 1) // num_steps
+
+ for i in range(epoch_size):
+ i = i * num_steps
+ data = indices[:, i: i + num_steps]
+ label = indices[:, i + 1: i + num_steps + 1]
+ yield data, label
+
+
+def grad_clipping(params, clipping_norm, ctx):
+ """Gradient clipping."""
+ if clipping_norm is not None:
+ norm = nd.array([0.0], ctx)
+ for p in params:
+ norm += nd.sum(p.grad ** 2)
+ norm = nd.sqrt(norm).asscalar()
+ if norm > clipping_norm:
+ for p in params:
+ p.grad[:] *= clipping_norm / norm
+
+
+def predict_rnn(rnn, prefix, num_chars, params, hidden_dim, ctx, idx_to_char,
+ char_to_idx, get_inputs, is_lstm=False):
+ """Predict the next chars given the prefix."""
+ prefix = prefix.lower()
+ state_h = nd.zeros(shape=(1, hidden_dim), ctx=ctx)
+ if is_lstm:
+ state_c = nd.zeros(shape=(1, hidden_dim), ctx=ctx)
+ output = [char_to_idx[prefix[0]]]
+ for i in range(num_chars + len(prefix)):
+ X = nd.array([output[-1]], ctx=ctx)
+ if is_lstm:
+ Y, state_h, state_c = rnn(get_inputs(X), state_h, state_c, *params)
+ else:
+ Y, state_h = rnn(get_inputs(X), state_h, *params)
+ if i < len(prefix) - 1:
+ next_input = char_to_idx[prefix[i + 1]]
+ else:
+ next_input = int(Y[0].argmax(axis=1).asscalar())
+ output.append(next_input)
+ return ''.join([idx_to_char[i] for i in output])
+
+
+def train_and_predict_rnn(rnn, is_random_iter, epochs, num_steps, hidden_dim,
+ learning_rate, clipping_norm, batch_size,
+ pred_period, pred_len, seqs, get_params, get_inputs,
+ ctx, corpus_indices, idx_to_char, char_to_idx,
+ is_lstm=False):
+ """Train an RNN model and predict the next item in the sequence."""
+ if is_random_iter:
+ data_iter = data_iter_random
+ else:
+ data_iter = data_iter_consecutive
+ params = get_params()
+
+ softmax_cross_entropy = gluon.loss.SoftmaxCrossEntropyLoss()
+
+ for e in range(1, epochs + 1):
+ # If consecutive sampling is used, in the same epoch, the hidden state
+ # is initialized only at the beginning of the epoch.
+ if not is_random_iter:
+ state_h = nd.zeros(shape=(batch_size, hidden_dim), ctx=ctx)
+ if is_lstm:
+ state_c = nd.zeros(shape=(batch_size, hidden_dim), ctx=ctx)
+ train_loss, num_examples = 0, 0
+ for data, label in data_iter(corpus_indices, batch_size, num_steps,
+ ctx):
+ # If random sampling is used, the hidden state has to be
+ # initialized for each mini-batch.
+ if is_random_iter:
+ state_h = nd.zeros(shape=(batch_size, hidden_dim), ctx=ctx)
+ if is_lstm:
+ state_c = nd.zeros(shape=(batch_size, hidden_dim), ctx=ctx)
+ with autograd.record():
+ # outputs shape: (batch_size, vocab_size)
+ if is_lstm:
+ outputs, state_h, state_c = rnn(get_inputs(data), state_h,
+ state_c, *params)
+ else:
+ outputs, state_h = rnn(get_inputs(data), state_h, *params)
+ # Let t_ib_j be the j-th element of the mini-batch at time i.
+ # label shape: (batch_size * num_steps)
+ # label = [t_0b_0, t_0b_1, ..., t_1b_0, t_1b_1, ..., ].
+ label = label.T.reshape((-1,))
+ # Concatenate outputs:
+ # shape: (batch_size * num_steps, vocab_size).
+ outputs = nd.concat(*outputs, dim=0)
+ # Now outputs and label are aligned.
+ loss = softmax_cross_entropy(outputs, label)
+ loss.backward()
+
+ grad_clipping(params, clipping_norm, ctx)
+ SGD(params, learning_rate)
+
+ train_loss += nd.sum(loss).asscalar()
+ num_examples += loss.size
+
+ if e % pred_period == 0:
+ print("Epoch %d. Training perplexity %f" % (e,
+ exp(train_loss / num_examples)))
+ for seq in seqs:
+ print(' - ', predict_rnn(rnn, seq, pred_len, params,
+ hidden_dim, ctx, idx_to_char, char_to_idx, get_inputs,
+ is_lstm))
+ print()
+
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/vgg19_slim_benchmark.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/vgg19_slim_benchmark.py
new file mode 100644
index 0000000000000000000000000000000000000000..e3cdc64537f2d6fdd347bb05e797c284df6999ff
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/vgg19_slim_benchmark.py
@@ -0,0 +1,25 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorflow.contrib.slim.nets as nets
+import numpy as np
+
+slim = tf.contrib.slim
+
+if __name__ == '__main__':
+ output_shape = 1000
+ batch_size = 128
+ image = tf.placeholder(name='input_x', shape=[None, 224, 224, 3], dtype=tf.float32)
+ labels = tf.placeholder(name='input_label', shape=[None, output_shape], dtype=tf.float32)
+ with slim.arg_scope(nets.vgg.vgg_arg_scope()):
+ vgg_19, end_points = nets.vgg.vgg_19(inputs=image, num_classes=output_shape, scope='vgg_19')
+ probabilities = tf.reduce_mean(tf.nn.softmax(vgg_19, dim=-1))
+ losses = tf.norm(tf.subtract(probabilities, labels))
+ train_op = tf.train.AdamOptimizer(learning_rate=0.0001).minimize(losses)
+ sess = tf.Session(config=npu_config_proto())
+ saver = tf.train.Saver()
+ sess.run(tf.global_variables_initializer())
+ while True:
+ datasets = np.random.randn(batch_size, 224, 224, 3).astype(np.float32)
+ datasets_labels = np.random.randn(batch_size, output_shape).astype(np.float32)
+ losses_val, _ = sess.run([losses, train_op], feed_dict={image: datasets, labels: datasets_labels})
+ print(losses_val)
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/vgg19_tl_benchmark.py b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/vgg19_tl_benchmark.py
new file mode 100644
index 0000000000000000000000000000000000000000..7ca1ac5512d546df17fb52a9462026bd15bc40fb
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/benchmark/vgg19_tl_benchmark.py
@@ -0,0 +1,23 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+from nets.vgg19 import get_vgg19
+import numpy as np
+
+
+if __name__ == '__main__':
+ sess = tf.Session(config=npu_config_proto())
+ x = tf.placeholder(name="inputs_x", shape=[None, 224, 224, 3], dtype=tf.float32)
+ y = tf.placeholder(name='inputs_y', shape=[None, 1000], dtype=tf.float32)
+ network = get_vgg19(x, sess, pretrained=False)
+ outputs_y = network.outputs
+ probs = tf.nn.softmax(outputs_y, name="prob")
+ loss = tf.reduce_mean(tf.subtract(probs, y))
+
+ while True:
+ batch_size = 128
+ datasets_x = np.random.randn(batch_size, 224, 224, 3).astype(np.float32)
+ datasets_y = np.random.randn(batch_size, 1000).astype(np.float32)
+ feed_dict = {x: datasets_x, y: datasets_y}
+ loss_val = sess.run(loss, feed_dict=feed_dict)
+ print('batch size %d, loss value is %.2f' % (batch_size, loss_val))
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/memory_usage_test.py b/TensorFlow/contrib/cv/InsightFace_TF/test/memory_usage_test.py
new file mode 100644
index 0000000000000000000000000000000000000000..2fb4aaf7bdbe6847f5701efe013c5dc683bc40cb
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/memory_usage_test.py
@@ -0,0 +1,124 @@
+from npu_bridge.npu_init import *
+import mxnet as mx
+import argparse
+import PIL.Image
+import io
+import numpy as np
+import cv2
+import tensorflow as tf
+import os
+import sys
+
+
+def parse_args():
+ parser = argparse.ArgumentParser(
+ formatter_class=argparse.ArgumentDefaultsHelpFormatter,
+ description='data path information'
+ )
+ parser.add_argument('--bin_path', default='../datasets/faces_ms1m_112x112/train.rec', type=str,
+ help='path to the binary image file')
+ parser.add_argument('--idx_path', default='../datasets/faces_ms1m_112x112/train.idx', type=str,
+ help='path to the image index path')
+ parser.add_argument('--tfrecords_file_path', default='../datasets/tfrecords', type=str,
+ help='path to the output of tfrecords file path')
+ args = parser.parse_args()
+ return args
+
+
+def mx2tfrecords_mem_test(imgidx, imgrec, args):
+ output_path = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ writer = tf.python_io.TFRecordWriter(output_path)
+ for i in imgidx:
+ img_info = imgrec.read_idx(i)
+ header, img = mx.recordio.unpack(img_info)
+ print(type(img))
+ print(img)
+ print(sys.getsizeof(img))
+ print('#####################')
+ img_mx = mx.image.imdecode(img)
+ print(type(img_mx))
+ print(sys.getsizeof(img_mx))
+ print(img_mx.size)
+ print(img_mx.dtype)
+ print(img_mx.context)
+ print(img_mx.stype)
+ print(img_mx)
+ print('#####################')
+ img_mx_np = img_mx.asnumpy()
+ print(type(img_mx_np))
+ print(sys.getsizeof(img_mx_np))
+ print('#####################')
+ back_mx_ndarray = mx.nd.array(img_mx_np)
+ print(type(back_mx_ndarray))
+ print(sys.getsizeof(back_mx_ndarray))
+ encoded_jpg_io = io.BytesIO(img)
+ print(sys.getsizeof(encoded_jpg_io))
+ image = PIL.Image.open(encoded_jpg_io)
+ np_img = np.array(image)
+ img = cv2.cvtColor(np_img, cv2.COLOR_RGB2BGR)
+ print(sys.getsizeof(img))
+ print('#####################')
+ img_raw = img.tobytes()
+ print(sys.getsizeof(img))
+ print('#####################')
+ writer.close()
+
+
+def mx2tfrecords(imgidx, imgrec, args):
+ output_path = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ writer = tf.python_io.TFRecordWriter(output_path)
+ for i in imgidx:
+ img_info = imgrec.read_idx(i)
+ header, img = mx.recordio.unpack(img_info)
+ # encoded_jpg_io = io.BytesIO(img)
+ # image = PIL.Image.open(encoded_jpg_io)
+ # np_img = np.array(image)
+ # img = cv2.cvtColor(np_img, cv2.COLOR_RGB2BGR)
+ # img_raw = img.tobytes()
+ # images = tf.image.decode_jpeg(img)
+ # images = tf.reshape(images, shape=(112, 112, 3))
+ # r, g, b = tf.split(images, num_or_size_splits=3, axis=-1)
+ # images = tf.concat([b, g, r], axis=-1)
+ # sess = tf.Session()
+ # np_images = sess.run(images)
+ # print(images.shape)
+ # print(type(np_images))
+ # print(sys.getsizeof(np_images))
+ # cv2.imshow('test', np_images)
+ # cv2.waitKey(0)
+ label = int(header.label)
+ example = tf.train.Example(features=tf.train.Features(feature={
+ 'image_raw': tf.train.Feature(bytes_list=tf.train.BytesList(value=[img])),
+ "label": tf.train.Feature(int64_list=tf.train.Int64List(value=[label]))
+ }))
+ writer.write(example.SerializeToString()) # Serialize To String
+ if i % 10000 == 0:
+ print('%d num image processed' % i)
+ writer.close()
+
+
+if __name__ == '__main__':
+ # define parameters
+ id2range = {}
+ data_shape = (3, 112, 112)
+ args = parse_args()
+ imgrec = mx.recordio.MXIndexedRecordIO(args.idx_path, args.bin_path, 'r')
+ s = imgrec.read_idx(0)
+ header, _ = mx.recordio.unpack(s)
+ print(header.label)
+ imgidx = list(range(1, int(header.label[0])))
+ seq_identity = range(int(header.label[0]), int(header.label[1]))
+ for identity in seq_identity:
+ s = imgrec.read_idx(identity)
+ header, _ = mx.recordio.unpack(s)
+ a, b = int(header.label[0]), int(header.label[1])
+ id2range[identity] = (a, b)
+ print('id2range', len(id2range))
+
+ # generate tfrecords
+ mx2tfrecords_mem_test(imgidx, imgrec, args)
+
+
+
+
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/multiple_gpu_test/__init__.py b/TensorFlow/contrib/cv/InsightFace_TF/test/multiple_gpu_test/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/multiple_gpu_test/test_mgpu_mnist.py b/TensorFlow/contrib/cv/InsightFace_TF/test/multiple_gpu_test/test_mgpu_mnist.py
new file mode 100644
index 0000000000000000000000000000000000000000..cabed8c75c8a34a23fd7aec9ffd5de48678927c9
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/multiple_gpu_test/test_mgpu_mnist.py
@@ -0,0 +1,225 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+import os
+
+Layer = tl.layers.Layer
+D_TYPE = tf.float32
+
+class DenseLayer(Layer):
+ """
+ The :class:`DenseLayer` class is a fully connected layer.
+
+ Parameters
+ ----------
+ layer : a :class:`Layer` instance
+ The `Layer` class feeding into this layer.
+ n_units : int
+ The number of units of the layer.
+ act : activation function
+ The function that is applied to the layer activations.
+ W_init : weights initializer
+ The initializer for initializing the weight matrix.
+ b_init : biases initializer or None
+ The initializer for initializing the bias vector. If None, skip biases.
+ W_init_args : dictionary
+ The arguments for the weights tf.get_variable.
+ b_init_args : dictionary
+ The arguments for the biases tf.get_variable.
+ name : a string or None
+ An optional name to attach to this layer.
+
+ Examples
+ --------
+ >>> network = tl.layers.InputLayer(x, name='input_layer')
+ >>> network = tl.layers.DenseLayer(
+ ... network,
+ ... n_units=800,
+ ... act = tf.nn.relu,
+ ... W_init=tf.truncated_normal_initializer(stddev=0.1),
+ ... name ='relu_layer'
+ ... )
+
+ >>> Without TensorLayer, you can do as follow.
+ >>> W = tf.Variable(
+ ... tf.random_uniform([n_in, n_units], -1.0, 1.0), name='W')
+ >>> b = tf.Variable(tf.zeros(shape=[n_units]), name='b')
+ >>> y = tf.nn.relu(tf.matmul(inputs, W) + b)
+
+ Notes
+ -----
+ If the input to this layer has more than two axes, it need to flatten the
+ input by using :class:`FlattenLayer` in this case.
+ """
+
+ def __init__(
+ self,
+ layer=None,
+ n_units=100,
+ act=tf.identity,
+ W_init=tf.truncated_normal_initializer(stddev=0.1),
+ b_init=tf.constant_initializer(value=0.0),
+ W_init_args={},
+ b_init_args={},
+ name='dense_layer',
+ ):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ if self.inputs.get_shape().ndims != 2:
+ raise Exception("The input dimension must be rank 2, please reshape or flatten it")
+
+ n_in = int(self.inputs.get_shape()[-1])
+ self.n_units = n_units
+ print(" [TL] DenseLayer %s: %d %s" % (self.name, self.n_units, act.__name__))
+ with tf.variable_scope(name) as vs:
+ with tf.device('/cpu:0'):
+ W = tf.get_variable(name='W', shape=(n_in, n_units), initializer=W_init, dtype=D_TYPE, **W_init_args)
+ if b_init is not None:
+ try:
+ with tf.device('/cpu:0'):
+ b = tf.get_variable(name='b', shape=(n_units), initializer=b_init, dtype=D_TYPE, **b_init_args)
+ except: # If initializer is a constant, do not specify shape.
+ with tf.device('/cpu:0'):
+ b = tf.get_variable(name='b', initializer=b_init, dtype=D_TYPE, **b_init_args)
+ self.outputs = act(tf.matmul(self.inputs, W) + b)
+ else:
+ self.outputs = act(tf.matmul(self.inputs, W))
+
+ # Hint : list(), dict() is pass by value (shallow), without them, it is
+ # pass by reference.
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+ self.all_layers.extend([self.outputs])
+ if b_init is not None:
+ self.all_params.extend([W, b])
+ else:
+ self.all_params.extend([W])
+
+
+def inference(x):
+ network = tl.layers.InputLayer(x, name='input')
+ network = tl.layers.DropoutLayer(network, keep=0.8, name='drop1')
+ network = DenseLayer(network, n_units=800, act=tf.nn.relu, name='relu1')
+ network = tl.layers.DropoutLayer(network, keep=0.5, name='drop2')
+ network = DenseLayer(network, n_units=800, act=tf.nn.relu, name='relu2')
+ network = tl.layers.DropoutLayer(network, keep=0.5, name='drop3')
+ network = DenseLayer(network, n_units=10, act=tf.identity, name='output')
+ y = network.outputs
+ return y
+
+
+def load_data():
+ X_train, y_train, X_val, y_val, X_test, y_test = \
+ tl.files.load_mnist_dataset(shape=(-1, 784), path='/home/aurora/workspaces/data')
+ print('X_train.shape', X_train.shape)
+ print('y_train.shape', y_train.shape)
+ print('X_val.shape', X_val.shape)
+ print('y_val.shape', y_val.shape)
+ print('X_test.shape', X_test.shape)
+ print('y_test.shape', y_test.shape)
+ print('X %s y %s' % (X_test.dtype, y_test.dtype))
+ return X_train, y_train
+
+
+def tower_losses(inputs, labels):
+ logit = inference(inputs)
+ loss = tf.nn.sparse_softmax_cross_entropy_with_logits(logits=logit, labels=labels, name='cross_entropy')
+ return loss
+
+
+def average_gradients(tower_grads):
+ """Calculate the average gradient for each shared variable across all towers.
+
+ Note that this function provides a synchronization point across all towers.
+
+ Args:
+ tower_grads: List of lists of (gradient, variable) tuples. The outer list
+ is over individual gradients. The inner list is over the gradient
+ calculation for each tower.
+ Returns:
+ List of pairs of (gradient, variable) where the gradient has been averaged
+ across all towers.
+ """
+ average_grads = []
+
+ for grad_and_vars in zip(*tower_grads):
+ # Note that each grad_and_vars looks like the following:
+ # ((grad0_gpu0, var0_gpu0), ... , (grad0_gpuN, var0_gpuN))
+ grads = []
+ for g, g1 in grad_and_vars:
+ # Add 0 dimension to the gradients to represent the tower.
+ expanded_g = tf.expand_dims(g, 0)
+
+ # Append on a 'tower' dimension which we will average over below.
+ grads.append(expanded_g)
+
+ # Average over the 'tower' dimension.
+ grad = tf.concat(axis=0, values=grads)
+ grad = tf.reduce_mean(grad, 0)
+
+ # Keep in mind that the Variables are redundant because they are shared
+ # across towers. So .. we will just return the first tower's pointer to
+ # the Variable.
+ v = grad_and_vars[0][1]
+ grad_and_var = (grad, v)
+ average_grads.append(grad_and_var)
+ return average_grads
+
+
+def train():
+ with tf.Graph().as_default(), tf.device('/cpu:0'):
+ global_step = tf.get_variable(
+ 'global_step', [],
+ initializer=tf.constant_initializer(0), trainable=False)
+ # Decay the learning rate exponentially based on the number of steps.
+ lr = tf.train.exponential_decay(0.01,
+ global_step,
+ 10000,
+ 0.99,
+ staircase=True)
+ # Create an optimizer that performs gradient descent.
+ opt = tf.train.GradientDescentOptimizer(lr)
+ tower_grads = []
+ x = tf.placeholder(tf.float32, shape=[None, 784], name='x')
+ y_ = tf.placeholder(tf.int64, shape=[None, ], name='y_')
+ with tf.variable_scope(tf.get_variable_scope()):
+ for i in range(1):
+ with tf.device('/cpu:0'):
+ with tf.name_scope('%s_%d' % ('tower', i)) as scope:
+ tl.layers.set_name_reuse(True)
+ # Dequeues one batch for the GPU
+ # Calculate the loss for one tower of the CIFAR model. This function
+ # constructs the entire CIFAR model but shares the variables across
+ # all towers.
+ summaries = tf.get_collection(tf.GraphKeys.SUMMARIES, scope)
+ loss = tower_losses(x, y_)
+ # Reuse variables for the next tower.
+ tf.get_variable_scope().reuse_variables()
+ # Calculate the gradients for the batch of data on this CIFAR tower.
+ grads = opt.compute_gradients(loss)
+ # Keep track of the gradients across all towers.
+ tower_grads.append(grads)
+ # We must calculate the mean of each gradient. Note that this is the
+ # synchronization point across all towers.
+ grads = average_gradients(tower_grads)
+ apply_gradient_op = opt.apply_gradients(grads, global_step=global_step)
+
+ # Track the moving averages of all trainable variables.
+ variable_averages = tf.train.ExponentialMovingAverage(0.9999, global_step)
+ variables_averages_op = variable_averages.apply(tf.trainable_variables())
+
+ train_op = tf.group(apply_gradient_op, variables_averages_op)
+ # Build an initialization operation to run below.
+ init = tf.global_variables_initializer()
+ sess = tf.Session(config=npu_config_proto(config_proto=tf.ConfigProto(
+ allow_soft_placement=True,
+ log_device_placement=True)))
+ sess.run(init)
+
+
+if __name__ == '__main__':
+ os.environ["CUDA_VISIBLE_DEVICES"] = "1"
+ train()
+
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/multiple_gpu_test/test_tensorlayer.py b/TensorFlow/contrib/cv/InsightFace_TF/test/multiple_gpu_test/test_tensorlayer.py
new file mode 100644
index 0000000000000000000000000000000000000000..73483480df9c16afa9612068cac9f59f7d5a0105
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/multiple_gpu_test/test_tensorlayer.py
@@ -0,0 +1,119 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+
+Layer = tl.layers.Layer
+D_TYPE = tf.float32
+
+
+class DenseLayer(Layer):
+ """
+ The :class:`DenseLayer` class is a fully connected layer.
+
+ Parameters
+ ----------
+ layer : a :class:`Layer` instance
+ The `Layer` class feeding into this layer.
+ n_units : int
+ The number of units of the layer.
+ act : activation function
+ The function that is applied to the layer activations.
+ W_init : weights initializer
+ The initializer for initializing the weight matrix.
+ b_init : biases initializer or None
+ The initializer for initializing the bias vector. If None, skip biases.
+ W_init_args : dictionary
+ The arguments for the weights tf.get_variable.
+ b_init_args : dictionary
+ The arguments for the biases tf.get_variable.
+ name : a string or None
+ An optional name to attach to this layer.
+
+ Examples
+ --------
+ >>> network = tl.layers.InputLayer(x, name='input_layer')
+ >>> network = tl.layers.DenseLayer(
+ ... network,
+ ... n_units=800,
+ ... act = tf.nn.relu,
+ ... W_init=tf.truncated_normal_initializer(stddev=0.1),
+ ... name ='relu_layer'
+ ... )
+
+ >>> Without TensorLayer, you can do as follow.
+ >>> W = tf.Variable(
+ ... tf.random_uniform([n_in, n_units], -1.0, 1.0), name='W')
+ >>> b = tf.Variable(tf.zeros(shape=[n_units]), name='b')
+ >>> y = tf.nn.relu(tf.matmul(inputs, W) + b)
+
+ Notes
+ -----
+ If the input to this layer has more than two axes, it need to flatten the
+ input by using :class:`FlattenLayer` in this case.
+ """
+
+ def __init__(
+ self,
+ layer=None,
+ n_units=100,
+ act=tf.identity,
+ W_init=tf.truncated_normal_initializer(stddev=0.1),
+ b_init=tf.constant_initializer(value=0.0),
+ W_init_args={},
+ b_init_args={},
+ name='dense_layer',
+ ):
+ Layer.__init__(self, name=name)
+ self.inputs = layer.outputs
+ if self.inputs.get_shape().ndims != 2:
+ raise Exception("The input dimension must be rank 2, please reshape or flatten it")
+
+ n_in = int(self.inputs.get_shape()[-1])
+ self.n_units = n_units
+ print(" [TL] DenseLayer %s: %d %s" % (self.name, self.n_units, act.__name__))
+ with tf.variable_scope(name) as vs:
+ with tf.device('/cpu:0'):
+ W = tf.get_variable(name='W', shape=(n_in, n_units), initializer=W_init, dtype=D_TYPE, **W_init_args)
+ if b_init is not None:
+ try:
+ with tf.device('/cpu:0'):
+ b = tf.get_variable(name='b', shape=(n_units), initializer=b_init, dtype=D_TYPE, **b_init_args)
+ except: # If initializer is a constant, do not specify shape.
+ with tf.device('/cpu:0'):
+ b = tf.get_variable(name='b', initializer=b_init, dtype=D_TYPE, **b_init_args)
+ self.outputs = act(tf.matmul(self.inputs, W) + b)
+ else:
+ self.outputs = act(tf.matmul(self.inputs, W))
+
+ # Hint : list(), dict() is pass by value (shallow), without them, it is
+ # pass by reference.
+ self.all_layers = list(layer.all_layers)
+ self.all_params = list(layer.all_params)
+ self.all_drop = dict(layer.all_drop)
+ self.all_layers.extend([self.outputs])
+ if b_init is not None:
+ self.all_params.extend([W, b])
+ else:
+ self.all_params.extend([W])
+
+
+def inference():
+ x = tf.placeholder(tf.float32, shape=[None, 784], name='x')
+ network = tl.layers.InputLayer(x, name='input')
+ network = tl.layers.DropoutLayer(network, keep=0.8, name='drop1')
+ network = DenseLayer(network, n_units=800, act=tf.nn.relu, name='relu1')
+ network = tl.layers.DropoutLayer(network, keep=0.5, name='drop2')
+ network = DenseLayer(network, n_units=800, act=tf.nn.relu, name='relu2')
+ network = tl.layers.DropoutLayer(network, keep=0.5, name='drop3')
+ network = DenseLayer(network, n_units=10, act=tf.identity, name='output')
+ return network
+
+
+if __name__ == '__main__':
+ with tf.device('/cpu:0'):
+ network = inference()
+ network.print_layers()
+ sess = tf.Session(config=npu_config_proto(config_proto=tf.ConfigProto(
+ allow_soft_placement=True,
+ log_device_placement=True)))
+ tl.layers.initialize_global_variables(sess)
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/resnet_test_static.py b/TensorFlow/contrib/cv/InsightFace_TF/test/resnet_test_static.py
new file mode 100644
index 0000000000000000000000000000000000000000..fd577b914fc834cd87bbeedfd4213743e6d69681
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/resnet_test_static.py
@@ -0,0 +1,56 @@
+from npu_bridge.npu_init import *
+from resnet import get_resnet
+import tensorflow as tf
+from nets_utils import get_tensor_static_val
+import numpy as np
+
+
+def resnet_diff_test(layers_num):
+ ckpt_file_path = '../model_weights/resnet_v1_'+str(layers_num)+'.ckpt'
+ x = tf.placeholder(dtype=tf.float32, shape=[1, 224, 224, 3], name='input_place')
+ tfconfig = tf.ConfigProto(allow_soft_placement=True)
+ sess = tf.Session(config=npu_config_proto(config_proto=tfconfig))
+ nets = get_resnet(x, 1000, layers_num, sess)
+ ckpt_static = get_tensor_static_val(ckpt_file_path, all_tensors=True, all_tensor_names=True)
+
+ print('###########'*30)
+ vars = tf.get_collection(tf.GraphKeys.GLOBAL_VARIABLES)
+
+ total_count = 0
+ mean_avg = 0.0
+ median_avg = 0.0
+ std_avg = 0.0
+
+ for var in vars:
+ var_name = var.op.name
+ var_name_new = var_name
+ if '_bn' in var_name:
+ var_name_new = var_name_new.replace('_bn', '')
+ if 'W_conv2d' in var_name:
+ var_name_new = var_name_new.replace('W_conv2d', 'weights')
+ if 'b_conv2d' in var_name:
+ var_name_new = var_name_new.replace('b_conv2d', 'biases')
+ if 'shortcut_conv' in var_name:
+ var_name_new = var_name_new.replace('shortcut_conv', 'shortcut')
+
+ if var_name_new in ckpt_static:
+ print(var_name_new, end=', ')
+ total_count += 1
+ ckpt_s = ckpt_static[var_name_new]
+ var_val = sess.run(var)
+ mean_diff = np.mean(var_val) - ckpt_s.mean
+ mean_avg += mean_diff
+ median_diff = np.median(var_val) - ckpt_s.median
+ median_avg += median_diff
+ std_diff = np.std(var_val) - ckpt_s.std
+ std_avg += std_diff
+ print('mean_diff: ', mean_diff, 'median_diff: ', median_diff, 'std_diff: ', std_diff)
+
+ print('total_mean_diff', mean_avg/total_count, 'total_mean_diff', median_avg/total_count,
+ 'total_std_diff', std_avg/total_count)
+
+
+if __name__ == '__main__':
+ with tf.device('/cpu:0'):
+ resnet_diff_test(50)
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/test/test_losses.py b/TensorFlow/contrib/cv/InsightFace_TF/test/test_losses.py
new file mode 100644
index 0000000000000000000000000000000000000000..9d8dc4c10d83379c66a05ae6a9f2da7700b209b2
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/test/test_losses.py
@@ -0,0 +1,138 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import numpy as np
+from losses.face_losses import cosineface_losses
+import mxnet as mx
+import math
+
+
+def arcface_loss_val(embedding, labels, weights, out_num, s=64., m=0.5):
+ '''
+ :param embedding: the input embedding vectors
+ :param labels: the input labels, the shape should be eg: (batch_size, 1)
+ :param s: scalar value default is 64
+ :param out_num: output class num
+ :param m: the margin value, default is 0.5
+ :return: the final cacualted output, this output is send into the tf.nn.softmax directly
+ '''
+ cos_m = math.cos(m)
+ sin_m = math.sin(m)
+ mm = sin_m * m * s
+ threshold = math.cos(math.pi - m)
+ with tf.variable_scope('arcface_loss'):
+ # inputs and weights norm
+ embedding_norm = tf.norm(embedding, axis=1, keep_dims=True)
+ embedding = tf.div(embedding, embedding_norm, name='norm_embedding')
+ weights_norm = tf.norm(weights, axis=0, keep_dims=True)
+ weights = tf.div(weights, weights_norm, name='norm_weights')
+ # cos(theta+m)
+ cos_t = tf.matmul(embedding, weights, name='cos_t')
+ cos_t2 = tf.square(cos_t, name='cos_2')
+ sin_t2 = tf.subtract(1., cos_t2, name='sin_2')
+ sin_t = tf.sqrt(sin_t2, name='sin_t')
+ cos_mt = s * tf.subtract(tf.multiply(cos_t, cos_m), tf.multiply(sin_t, sin_m), name='cos_mt')
+
+ # this condition controls the theta+m should in range [0, pi]
+ # 0<=theta+m<=pi
+ # -m<=theta<=pi-m
+ cond_v = cos_t - threshold
+ cond = tf.cast(tf.nn.relu(cond_v, name='if_else'), dtype=tf.bool)
+
+ keep_val = s * (cos_t - mm)
+ cos_mt_temp = tf.where(cond, cos_mt, keep_val)
+
+ mask = tf.one_hot(labels, depth=out_num, name='one_hot_mask')
+ inv_mask = tf.subtract(1., mask, name='inverse_mask')
+
+ s_cos_t = tf.multiply(s, cos_t, name='scalar_cos_t')
+
+ output = tf.add(tf.multiply(s_cos_t, inv_mask), tf.multiply(cos_mt_temp, mask), name='arcface_loss_output')
+ return output
+
+
+def test_arcface_losses(np_embedding, np_weights):
+ tf_embedding = tf.constant(np_embedding, name='embedding', dtype=tf.float32)
+ labels = tf.constant([1, 3, 2, 1, 1], name='input_labels', dtype=tf.int64)
+ print(labels)
+ tf_weights = tf.constant(np_weights, name='weights')
+ output = arcface_loss_val(embedding=tf_embedding, labels=labels, out_num=10, weights=tf_weights)
+ print(output)
+ sess = tf.Session(config=npu_config_proto())
+ sess.run(tf.global_variables_initializer())
+ results1 = sess.run(output)
+ print(results1)
+ return results1
+
+
+def test_cosineface_losses():
+ np_embedding = np.random.randn(5, 512).astype(dtype=np.float32)
+ tf_embedding = tf.constant(np_embedding, name='embedding', dtype=tf.float32)
+ labels = tf.constant([1, 3, 2, 1, 1], name='input_labels', dtype=tf.int64)
+ output = cosineface_losses(embedding=tf_embedding, labels=labels, out_num=10)
+ sess = tf.Session(config=npu_config_proto())
+ sess.run(tf.global_variables_initializer())
+ print(sess.run(output))
+
+
+def test_mxnet_losses(np_embedding, np_weights):
+ labels = np.array([1, 3, 2, 1, 1]).astype(dtype=np.float32)
+ return mxnet_arcface_val(np_embedding, labels, np_weights)
+
+
+def mxnet_arcface_val(embedding, gt_label, weights):
+ s = 64
+ m = 0.5
+ _weight = mx.symbol.Variable("fc7_weight", shape=(10, 512), lr_mult=1.0)
+ _weight = mx.symbol.L2Normalization(_weight, mode='instance')
+ _embedding = mx.symbol.Variable('mx_embedding', shape=(5, 512), lr_mult=1.0)
+ nembedding = mx.symbol.L2Normalization(_embedding, mode='instance', name='fc1n')*s
+ fc7 = mx.sym.FullyConnected(data=nembedding, weight=_weight, no_bias=True, num_hidden=10, name='fc7')
+
+ _labels = mx.symbol.Variable('labels', shape=(5, ), lr_mult=1.0)
+ zy = mx.sym.pick(fc7, _labels, axis=1)
+ cos_t = zy/s
+
+ cos_m = math.cos(m)
+ sin_m = math.sin(m)
+ mm = math.sin(math.pi - m) * m
+ # threshold = 0.0
+ threshold = math.cos(math.pi - m)
+
+ cond_v = cos_t - threshold
+ cond = mx.symbol.Activation(data=cond_v, act_type='relu')
+
+ body = cos_t * cos_t
+ body = 1.0 - body
+ sin_t = mx.sym.sqrt(body)
+ new_zy = cos_t * cos_m
+ b = sin_t * sin_m
+ new_zy = new_zy - b
+ new_zy = new_zy * s
+
+ zy_keep = zy - s * mm
+ new_zy = mx.sym.where(cond, new_zy, zy_keep)
+
+ diff = new_zy - zy
+ diff = mx.sym.expand_dims(diff, 1)
+ gt_one_hot = mx.sym.one_hot(_labels, depth = 10, on_value = 1.0, off_value = 0.0)
+ body = mx.sym.broadcast_mul(gt_one_hot, diff)
+ fc7 = fc7+body
+ executor = fc7.bind(mx.cpu(), {'fc7_weight': mx.nd.array(weights.T), 'mx_embedding': mx.nd.array(embedding),
+ 'labels': mx.nd.array(gt_label)})
+ output = executor.forward()
+ print(output)
+ return output
+
+
+if __name__ == '__main__':
+ np_embedding = np.random.randn(5, 512).astype(dtype=np.float32)
+ np_weights = np.random.randn(512, 10).astype(dtype=np.float32)
+ # test arcface_losses output
+ result1 = test_arcface_losses(np_embedding, np_weights)
+ # print('########'*30)
+ print('################')
+ result2 = test_mxnet_losses(np_embedding, np_weights)
+ print(len(result2[0]))
+ print(type(result1))
+ print(type(result2[0].asnumpy()))
+ print(np.mean(result1 - result2[0].asnumpy())) # 1.26362e-07
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/train_nets.py b/TensorFlow/contrib/cv/InsightFace_TF/train_nets.py
new file mode 100644
index 0000000000000000000000000000000000000000..665318472771791501302c9652cc95bf46c430e3
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/train_nets.py
@@ -0,0 +1,229 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+import argparse
+from data.mx2tfrecords import parse_function
+import os
+# from nets.L_Resnet_E_IR import get_resnet
+# from nets.L_Resnet_E_IR_GBN import get_resnet
+from nets.L_Resnet_E_IR_fix_issue9 import get_resnet
+from losses.face_losses import arcface_loss
+from tensorflow.core.protobuf import config_pb2
+import time
+from data.eval_data_reader import load_bin
+from verification import ver_test
+
+
+def get_parser():
+ parser = argparse.ArgumentParser(description='parameters to train net')
+ parser.add_argument('--net_depth', default=100, help='resnet depth, default is 50')
+ parser.add_argument('--epoch', default=200, help='epoch to train the network')
+ parser.add_argument('--batch_size', default=32, help='batch size to train network')
+ parser.add_argument('--lr_steps', default=[40000, 60000, 80000], help='learning rate to train network')
+ parser.add_argument('--momentum', default=0.9, help='learning alg momentum')
+ parser.add_argument('--weight_deacy', default=5e-4, help='learning alg momentum')
+ # parser.add_argument('--eval_datasets', default=['lfw', 'cfp_ff', 'cfp_fp', 'agedb_30'], help='evluation datasets')
+ parser.add_argument('--eval_datasets', default=['lfw'], help='evluation datasets')
+ parser.add_argument('--eval_db_path', default='./datasets/faces_ms1m_112x112', help='evluate datasets base path')
+ parser.add_argument('--image_size', default=[112, 112], help='the image size')
+ parser.add_argument('--num_output', default=85164, help='the image size')
+ parser.add_argument('--tfrecords_file_path', default='./datasets/tfrecords_webface', type=str,
+ help='path to the output of tfrecords file path')
+ parser.add_argument('--summary_path', default='./output/summary', help='the summary file save path')
+ parser.add_argument('--ckpt_path', default='./output/ckpt', help='the ckpt file save path')
+ parser.add_argument('--log_file_path', default='./output/logs', help='the ckpt file save path')
+ parser.add_argument('--saver_maxkeep', default=100, help='tf.train.Saver max keep ckpt files')
+ parser.add_argument('--buffer_size', default=10000, help='tf dataset api buffer size')
+ parser.add_argument('--log_device_mapping', default=False, help='show device placement log')
+ parser.add_argument('--summary_interval', default=300, help='interval to save summary')
+ parser.add_argument('--ckpt_interval', default=10000, help='intervals to save ckpt file')
+ parser.add_argument('--validate_interval', default=2000, help='intervals to save ckpt file')
+ parser.add_argument('--show_info_interval', default=20, help='intervals to save ckpt file')
+ args = parser.parse_args()
+ return args
+
+
+if __name__ == '__main__':
+ os.environ["CUDA_VISIBLE_DEVICES"] = "0"
+ # 1. define global parameters
+ args = get_parser()
+ global_step = tf.Variable(name='global_step', initial_value=0, trainable=False)
+ inc_op = tf.assign_add(global_step, 1, name='increment_global_step')
+ images = tf.placeholder(name='img_inputs', shape=[None, *args.image_size, 3], dtype=tf.float32)
+ labels = tf.placeholder(name='img_labels', shape=[None, ], dtype=tf.int64)
+ # trainable = tf.placeholder(name='trainable_bn', dtype=tf.bool)
+ dropout_rate = tf.placeholder(name='dropout_rate', dtype=tf.float32)
+ # 2 prepare train datasets and test datasets by using tensorflow dataset api
+ # 2.1 train datasets
+ # the image is substracted 127.5 and multiplied 1/128.
+ # random flip left right
+ tfrecords_f = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ print("tfrecords_f path : " + tfrecords_f)
+ dataset = tf.data.TFRecordDataset(tfrecords_f)
+ dataset = dataset.map(parse_function)
+ print("parse_function :" , parse_function)
+ dataset = dataset.shuffle(buffer_size=args.buffer_size)
+ dataset = dataset.batch(args.batch_size, drop_remainder=True)
+ iterator = dataset.make_initializable_iterator()
+ next_element = iterator.get_next()
+ print("let look at the next_element:")
+ print(iterator.__sizeof__())
+ # 2.2 prepare validate datasets
+ ver_list = []
+ ver_name_list = []
+ for db in args.eval_datasets:
+ print('begin db %s convert.' % db)
+ data_set = load_bin(db, args.image_size, args)
+ ver_list.append(data_set)
+ ver_name_list.append(db)
+ # 3. define network, loss, optimize method, learning rate schedule, summary writer, saver
+ # 3.1 inference phase
+ w_init_method = tf.contrib.layers.xavier_initializer(uniform=False)
+ net = get_resnet(images, args.net_depth, type='ir', w_init=w_init_method, trainable=True, keep_rate=dropout_rate)
+ # 3.2 get arcface loss
+ logit = arcface_loss(embedding=net.outputs, labels=labels, w_init=w_init_method, out_num=args.num_output)
+ # test net because of batch normal layer
+ tl.layers.set_name_reuse(True)
+ test_net = get_resnet(images, args.net_depth, type='ir', w_init=w_init_method, trainable=False, reuse=True, keep_rate=dropout_rate)
+ embedding_tensor = test_net.outputs
+ # 3.3 define the cross entropy
+ inference_loss = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(logits=logit, labels=labels))
+ # inference_loss_avg = tf.reduce_mean(inference_loss)
+ # 3.4 define weight deacy losses
+ # for var in tf.trainable_variables():
+ # print(var.name)
+ # print('##########'*30)
+ wd_loss = 0
+ for weights in tl.layers.get_variables_with_name('W_conv2d', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(weights)
+ for W in tl.layers.get_variables_with_name('resnet_v1_50/E_DenseLayer/W', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(W)
+ for weights in tl.layers.get_variables_with_name('embedding_weights', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(weights)
+ for gamma in tl.layers.get_variables_with_name('gamma', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(gamma)
+ # for beta in tl.layers.get_variables_with_name('beta', True, True):
+ # wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(beta)
+ for alphas in tl.layers.get_variables_with_name('alphas', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(alphas)
+ # for bias in tl.layers.get_variables_with_name('resnet_v1_50/E_DenseLayer/b', True, True):
+ # wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(bias)
+
+ # 3.5 total losses
+ total_loss = inference_loss + wd_loss
+ # 3.6 define the learning rate schedule
+ p = int(512.0/args.batch_size)
+ lr_steps = [p*val for val in args.lr_steps]
+ print(lr_steps)
+ lr = tf.train.piecewise_constant(global_step, boundaries=lr_steps, values=[0.001, 0.0005, 0.0003, 0.0001], name='lr_schedule')
+ # 3.7 define the optimize method
+ opt = tf.train.MomentumOptimizer(learning_rate=lr, momentum=args.momentum)
+ # 3.8 get train op
+ grads = opt.compute_gradients(total_loss)
+ update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)
+ with tf.control_dependencies(update_ops):
+ train_op = opt.apply_gradients(grads, global_step=global_step)
+ # train_op = opt.minimize(total_loss, global_step=global_step)
+ # 3.9 define the inference accuracy used during validate or test
+ pred = tf.nn.softmax(logit)
+ acc = tf.reduce_mean(tf.cast(tf.equal(tf.argmax(pred, axis=1), labels), dtype=tf.float32))
+ # 3.10 define sess
+ config = tf.ConfigProto(allow_soft_placement=True, log_device_placement=args.log_device_mapping)
+ config.gpu_options.allow_growth = True
+
+ sess = tf.Session(config=npu_config_proto(config_proto=config))
+ # 3.11 summary writer
+ summary = tf.summary.FileWriter(args.summary_path, sess.graph)
+ summaries = []
+ # # 3.11.1 add grad histogram op
+ for grad, var in grads:
+ if grad is not None:
+ summaries.append(tf.summary.histogram(var.op.name + '/gradients', grad))
+ # 3.11.2 add trainabel variable gradients
+ for var in tf.trainable_variables():
+ summaries.append(tf.summary.histogram(var.op.name, var))
+ # 3.11.3 add loss summary
+ summaries.append(tf.summary.scalar('inference_loss', inference_loss))
+ summaries.append(tf.summary.scalar('wd_loss', wd_loss))
+ summaries.append(tf.summary.scalar('total_loss', total_loss))
+ # 3.11.4 add learning rate
+ summaries.append(tf.summary.scalar('leraning_rate', lr))
+ summary_op = tf.summary.merge(summaries)
+ # 3.12 saver
+ saver = tf.train.Saver(max_to_keep=args.saver_maxkeep)
+ # 3.13 init all variables
+ sess.run(tf.global_variables_initializer())
+
+ # restore_saver = tf.train.Saver()
+ # restore_saver.restore(sess, '/home/aurora/workspaces2018/InsightFace_TF/output/ckpt/InsightFace_iter_1110000.ckpt')
+ # 4 begin iteration
+ if not os.path.exists(args.log_file_path):
+ os.makedirs(args.log_file_path)
+ log_file_path = args.log_file_path + '/train' + time.strftime('_%Y-%m-%d-%H-%M', time.localtime(time.time())) + '.log'
+ log_file = open(log_file_path, 'w')
+ # 4 begin iteration
+ count = 0
+ total_accuracy = {}
+
+ for i in range(args.epoch):
+ sess.run(iterator.initializer)
+ while True:
+ try:
+ images_train, labels_train = sess.run(next_element)
+ feed_dict = {images: images_train, labels: labels_train, dropout_rate: 0.4}
+ feed_dict.update(net.all_drop)
+ start = time.time()
+ _, total_loss_val, inference_loss_val, wd_loss_val, _, acc_val = \
+ sess.run([train_op, total_loss, inference_loss, wd_loss, inc_op, acc],
+ feed_dict=feed_dict,
+ options=config_pb2.RunOptions(report_tensor_allocations_upon_oom=True))
+ end = time.time()
+ pre_sec = args.batch_size/(end - start)
+ # print training information
+ if count > 0 and count % args.show_info_interval == 0:
+ print('epoch %d, total_step %d, total loss is %.2f , inference loss is %.2f, weight deacy '
+ 'loss is %.2f, training accuracy is %.6f, time %.3f samples/sec' %
+ (i, count, total_loss_val, inference_loss_val, wd_loss_val, acc_val, pre_sec))
+ count += 1
+
+ # save summary
+ if count > 0 and count % args.summary_interval == 0:
+ feed_dict = {images: images_train, labels: labels_train, dropout_rate: 0.4}
+ feed_dict.update(net.all_drop)
+ summary_op_val = sess.run(summary_op, feed_dict=feed_dict)
+ summary.add_summary(summary_op_val, count)
+
+ # save ckpt files
+ if count > 0 and count % args.ckpt_interval == 0:
+ filename = 'InsightFace_iter_{:d}'.format(count) + '.ckpt'
+ filename = os.path.join(args.ckpt_path, filename)
+ saver.save(sess, filename)
+
+ # validate
+ if count > 0 and count % args.validate_interval == 0:
+ feed_dict_test ={dropout_rate: 1.0}
+ feed_dict_test.update(tl.utils.dict_to_one(net.all_drop))
+ results = ver_test(ver_list=ver_list, ver_name_list=ver_name_list, nbatch=count, sess=sess,
+ embedding_tensor=embedding_tensor, batch_size=args.batch_size, feed_dict=feed_dict_test,
+ input_placeholder=images)
+ print('test accuracy is: ', str(results[0]))
+ total_accuracy[str(count)] = results[0]
+ log_file.write('########'*10+'\n')
+ log_file.write(','.join(list(total_accuracy.keys())) + '\n')
+ log_file.write(','.join([str(val) for val in list(total_accuracy.values())])+'\n')
+ log_file.flush()
+ if max(results) > 0.996:
+ print('best accuracy is %.5f' % max(results))
+ filename = 'InsightFace_iter_best_{:d}'.format(count) + '.ckpt'
+ filename = os.path.join(args.ckpt_path, filename)
+ saver.save(sess, filename)
+ log_file.write('######Best Accuracy######'+'\n')
+ log_file.write(str(max(results))+'\n')
+ log_file.write(filename+'\n')
+
+ log_file.flush()
+ except tf.errors.OutOfRangeError:
+ print("End of epoch %d" % i)
+ break
+ log_file.write('\n')
+ log_file.close()
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/train_nets_mgpu.py b/TensorFlow/contrib/cv/InsightFace_TF/train_nets_mgpu.py
new file mode 100644
index 0000000000000000000000000000000000000000..0f500a7aeccb386e1b81812d6d0e537f13e80d5e
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/train_nets_mgpu.py
@@ -0,0 +1,261 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+import argparse
+from data.mx2tfrecords import parse_function
+import os
+from nets.L_Resnet_E_IR_MGPU import get_resnet
+from losses.face_losses import arcface_loss
+import time
+from data.eval_data_reader import load_bin
+from verification import ver_test
+
+
+def get_parser():
+ parser = argparse.ArgumentParser(description='parameters to train net')
+ parser.add_argument('--net_depth', default=50, help='resnet depth, default is 50')
+ parser.add_argument('--epoch', default=100000, help='epoch to train the network')
+ parser.add_argument('--batch_size', default=32, help='batch size to train network')
+ parser.add_argument('--lr_steps', default=[40000, 60000, 80000], help='learning rate to train network')
+ parser.add_argument('--momentum', default=0.9, help='learning alg momentum')
+ parser.add_argument('--weight_deacy', default=5e-4, help='learning alg momentum')
+ # parser.add_argument('--eval_datasets', default=['lfw', 'cfp_ff', 'cfp_fp', 'agedb_30'], help='evluation datasets')
+ parser.add_argument('--eval_datasets', default=['lfw', 'cfp_fp'], help='evluation datasets')
+ parser.add_argument('--eval_db_path', default='./datasets/faces_ms1m_112x112', help='evluate datasets base path')
+ parser.add_argument('--image_size', default=[112, 112], help='the image size')
+ parser.add_argument('--num_output', default=85164, help='the image size')
+ parser.add_argument('--tfrecords_file_path', default='./datasets/tfrecords', type=str,
+ help='path to the output of tfrecords file path')
+ parser.add_argument('--summary_path', default='./output/summary', help='the summary file save path')
+ parser.add_argument('--ckpt_path', default='./output/ckpt', help='the ckpt file save path')
+ parser.add_argument('--saver_maxkeep', default=100, help='tf.train.Saver max keep ckpt files')
+ parser.add_argument('--buffer_size', default=50000, help='tf dataset api buffer size')
+ parser.add_argument('--log_device_mapping', default=False, help='show device placement log')
+ parser.add_argument('--summary_interval', default=300, help='interval to save summary')
+ parser.add_argument('--ckpt_interval', default=5000, help='intervals to save ckpt file')
+ parser.add_argument('--validate_interval', default=2000, help='intervals to save ckpt file')
+ parser.add_argument('--show_info_interval', default=20, help='intervals to show information')
+ parser.add_argument('--num_gpus', default=2, help='the num of gpus')
+ parser.add_argument('--tower_name', default='tower', help='tower name')
+ args = parser.parse_args()
+ return args
+
+
+def average_gradients(tower_grads):
+ """Calculate the average gradient for each shared variable across all towers.
+
+ Note that this function provides a synchronization point across all towers.
+
+ Args:
+ tower_grads: List of lists of (gradient, variable) tuples. The outer list
+ is over individual gradients. The inner list is over the gradient
+ calculation for each tower.
+ Returns:
+ List of pairs of (gradient, variable) where the gradient has been averaged
+ across all towers.
+ """
+ average_grads = []
+ for grad_and_vars in zip(*tower_grads):
+ # Note that each grad_and_vars looks like the following:
+ # ((grad0_gpu0, var0_gpu0), ... , (grad0_gpuN, var0_gpuN))
+ grads = []
+ for g, _ in grad_and_vars:
+ # Add 0 dimension to the gradients to represent the tower.
+ expanded_g = tf.expand_dims(g, 0)
+
+ # Append on a 'tower' dimension which we will average over below.
+ grads.append(expanded_g)
+
+ # Average over the 'tower' dimension.
+ grad = tf.concat(axis=0, values=grads)
+ grad = tf.reduce_mean(grad, 0)
+
+ # Keep in mind that the Variables are redundant because they are shared
+ # across towers. So .. we will just return the first tower's pointer to
+ # the Variable.
+ v = grad_and_vars[0][1]
+ grad_and_var = (grad, v)
+ average_grads.append(grad_and_var)
+ return average_grads
+
+
+if __name__ == '__main__':
+ # os.environ["CUDA_VISIBLE_DEVICES"] = "0"
+ # 1. define global parameters
+ args = get_parser()
+ global_step = tf.Variable(name='global_step', initial_value=0, trainable=False)
+ inc_op = tf.assign_add(global_step, 1, name='increment_global_step')
+ trainable = tf.placeholder(name='trainable_bn', dtype=tf.bool)
+ images = tf.placeholder(name='img_inputs', shape=[None, *args.image_size, 3], dtype=tf.float32)
+ labels = tf.placeholder(name='img_labels', shape=[None, ], dtype=tf.int64)
+ # splits input to different gpu
+ images_s = tf.split(images, num_or_size_splits=args.num_gpus, axis=0)
+ labels_s = tf.split(labels, num_or_size_splits=args.num_gpus, axis=0)
+ # 2 prepare train datasets and test datasets by using tensorflow dataset api
+ # 2.1 train datasets
+ # the image is substracted 127.5 and multiplied 1/128.
+ # random flip left right
+ tfrecords_f = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ dataset = tf.data.TFRecordDataset(tfrecords_f)
+ dataset = dataset.map(parse_function)
+ dataset = dataset.shuffle(buffer_size=args.buffer_size)
+ dataset = dataset.batch(args.batch_size, drop_remainder=True)
+ iterator = dataset.make_initializable_iterator()
+ next_element = iterator.get_next()
+ # 2.2 prepare validate datasets
+ ver_list = []
+ ver_name_list = []
+ for db in args.eval_datasets:
+ print('begin db %s convert.' % db)
+ data_set = load_bin(db, args.image_size, args)
+ ver_list.append(data_set)
+ ver_name_list.append(db)
+
+ # 3. define network, loss, optimize method, learning rate schedule, summary writer, saver
+ # 3.1 inference phase
+ w_init_method = tf.contrib.layers.xavier_initializer(uniform=False)
+ # 3.2 define the learning rate schedule
+ p = int(512.0/args.batch_size)
+ lr_steps = [p*val for val in args.lr_steps]
+ print('learning rate steps: ', lr_steps)
+ lr = tf.train.piecewise_constant(global_step, boundaries=lr_steps, values=[0.001, 0.0001, 0.00005, 0.00001], name='lr_schedule')
+ # 3.3 define the optimize method
+ opt = tf.train.MomentumOptimizer(learning_rate=lr, momentum=args.momentum)
+
+ # Calculate the gradients for each model tower.
+ tower_grads = []
+ tl.layers.set_name_reuse(True)
+ loss_dict = {}
+ drop_dict = {}
+ loss_keys = []
+ with tf.variable_scope(tf.get_variable_scope()):
+ for i in range(args.num_gpus):
+ with tf.device('/cpu:0'):
+ with tf.name_scope('%s_%d' % (args.tower_name, i)) as scope:
+ net = get_resnet(images_s[i], args.net_depth, type='ir', w_init=w_init_method, trainable=trainable)
+ logit = arcface_loss(embedding=net.outputs, labels=labels_s[i], w_init=w_init_method, out_num=args.num_output)
+ # Reuse variables for the next tower.
+ tf.get_variable_scope().reuse_variables()
+ # define the cross entropy
+ inference_loss = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(logits=logit, labels=labels_s[i]))
+ # define weight deacy losses
+ wd_loss = 0
+ for weights in tl.layers.get_variables_with_name('W_conv2d', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(weights)
+ for W in tl.layers.get_variables_with_name('resnet_v1_50/E_DenseLayer/W', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(W)
+ for weights in tl.layers.get_variables_with_name('embedding_weights', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(weights)
+ for gamma in tl.layers.get_variables_with_name('gamma', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(gamma)
+ for beta in tl.layers.get_variables_with_name('beta', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(beta)
+ for alphas in tl.layers.get_variables_with_name('alphas', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(alphas)
+ for bias in tl.layers.get_variables_with_name('resnet_v1_50/E_DenseLayer/b', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(bias)
+ total_loss = inference_loss + wd_loss
+
+ loss_dict[('inference_loss_%s_%d' % ('gpu', i))] = inference_loss
+ loss_keys.append(('inference_loss_%s_%d' % ('gpu', i)))
+ loss_dict[('wd_loss_%s_%d' % ('gpu', i))] = wd_loss
+ loss_keys.append(('wd_loss_%s_%d' % ('gpu', i)))
+ loss_dict[('total_loss_%s_%d' % ('gpu', i))] = total_loss
+ loss_keys.append(('total_loss_%s_%d' % ('gpu', i)))
+ grads = opt.compute_gradients(total_loss)
+ tower_grads.append(grads)
+ drop_dict.update(net.all_drop)
+ if i == 0:
+ update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)
+ pred = tf.nn.softmax(logit)
+ acc = tf.reduce_mean(tf.cast(tf.equal(tf.argmax(pred, axis=1), labels_s[i]), dtype=tf.int64))
+ embedding_tensor_gpu0 = net.outputs
+
+ grads = average_gradients(tower_grads)
+ with tf.control_dependencies(update_ops):
+ # Apply the gradients to adjust the shared variables.
+ train_op = opt.apply_gradients(grads, global_step=global_step)
+
+ config = tf.ConfigProto(allow_soft_placement=True, log_device_placement=args.log_device_mapping)
+ config.gpu_options.allow_growth = True
+ sess = tf.Session(config=npu_config_proto(config_proto=config))
+ # summary writer
+ summary = tf.summary.FileWriter(args.summary_path, sess.graph)
+ summaries = []
+ # add grad histogram op
+ for grad, var in grads:
+ if grad is not None:
+ summaries.append(tf.summary.histogram(var.op.name + '/gradients', grad))
+ # add trainabel variable gradients
+ for var in tf.trainable_variables():
+ summaries.append(tf.summary.histogram(var.op.name, var))
+ # add loss summary
+ for keys, val in loss_dict.items():
+ summaries.append(tf.summary.scalar(keys, val))
+ # add learning rate
+ summaries.append(tf.summary.scalar('leraning_rate', lr))
+ summary_op = tf.summary.merge(summaries)
+
+ # Create a saver.
+ saver = tf.train.Saver(tf.global_variables())
+ # init all variables
+ sess.run(tf.global_variables_initializer())
+
+ drop_dict_test = {keys: 1 for keys in drop_dict.keys()}
+ # begin iteration
+ count = 0
+ for i in range(args.epoch):
+ sess.run(iterator.initializer)
+ while True:
+ try:
+ images_train, labels_train = sess.run(next_element)
+ feed_dict = {images: images_train, labels: labels_train, trainable: True}
+ feed_dict.update(drop_dict)
+ start = time.time()
+ _, _, inference_loss_val_gpu_1, wd_loss_val_gpu_1, total_loss_gpu_1, inference_loss_val_gpu_2, \
+ wd_loss_val_gpu_2, total_loss_gpu_2, acc_val = sess.run([train_op, inc_op, loss_dict[loss_keys[0]],
+ loss_dict[loss_keys[1]],
+ loss_dict[loss_keys[2]],
+ loss_dict[loss_keys[3]],
+ loss_dict[loss_keys[4]],
+ loss_dict[loss_keys[5]], acc],
+ feed_dict=feed_dict)
+ end = time.time()
+ pre_sec = args.batch_size/(end - start)
+ # print training information
+ if count > 0 and count % args.show_info_interval == 0:
+ print('epoch %d, total_step %d, total loss gpu 1 is %.2f , inference loss gpu 1 is %.2f, weight deacy '
+ 'loss gpu 1 is %.2f, total loss gpu 2 is %.2f , inference loss gpu 2 is %.2f, weight deacy '
+ 'loss gpu 2 is %.2f, training accuracy is %.6f, time %.3f samples/sec' %
+ (i, count, total_loss_gpu_1, inference_loss_val_gpu_1, wd_loss_val_gpu_1, total_loss_gpu_2,
+ inference_loss_val_gpu_2, wd_loss_val_gpu_2, acc_val, pre_sec))
+ count += 1
+
+ # save summary
+ if count > 0 and count % args.summary_interval == 0:
+ feed_dict = {images: images_train, labels: labels_train, trainable: True}
+ feed_dict.update(drop_dict)
+ summary_op_val = sess.run(summary_op, feed_dict=feed_dict)
+ summary.add_summary(summary_op_val, count)
+
+ # save ckpt files
+ if count > 0 and count % args.ckpt_interval == 0:
+ filename = 'InsightFace_iter_{:d}'.format(count) + '.ckpt'
+ filename = os.path.join(args.ckpt_path, filename)
+ saver.save(sess, filename)
+ # # validate
+ if count > 0 and count % args.validate_interval == 0:
+ feed_dict_test ={trainable: False}
+ feed_dict_test.update(drop_dict_test)
+ results = ver_test(ver_list=ver_list, ver_name_list=ver_name_list, nbatch=count, sess=sess,
+ embedding_tensor=embedding_tensor_gpu0, batch_size=args.batch_size//args.num_gpus, feed_dict=feed_dict_test,
+ input_placeholder=images_s[0])
+ if max(results) > 0.99:
+ print('best accuracy is %.5f' % max(results))
+ filename = 'InsightFace_iter_best_{:d}'.format(count) + '.ckpt'
+ filename = os.path.join(args.ckpt_path, filename)
+ saver.save(sess, filename)
+ except tf.errors.OutOfRangeError:
+ print("End of epoch %d" % i)
+ break
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/train_nets_mgpu_new.py b/TensorFlow/contrib/cv/InsightFace_TF/train_nets_mgpu_new.py
new file mode 100644
index 0000000000000000000000000000000000000000..0670cb9a7274a8bb1c7cf6de31fd01a76ff926c2
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/train_nets_mgpu_new.py
@@ -0,0 +1,258 @@
+from npu_bridge.npu_init import *
+import tensorflow as tf
+import tensorlayer as tl
+import argparse
+from data.mx2tfrecords import parse_function
+import os
+from nets.L_Resnet_E_IR_MGPU import get_resnet
+from losses.face_losses import arcface_loss
+import time
+from data.eval_data_reader import load_bin
+from verification import ver_test
+
+
+def get_parser():
+ parser = argparse.ArgumentParser(description='parameters to train net')
+ parser.add_argument('--net_depth', default=100, help='resnet depth, default is 50')
+ parser.add_argument('--epoch', default=100000, help='epoch to train the network')
+ parser.add_argument('--batch_size', default=64, help='batch size to train network')
+ parser.add_argument('--lr_steps', default=[40000, 60000, 80000], help='learning rate to train network')
+ parser.add_argument('--momentum', default=0.9, help='learning alg momentum')
+ parser.add_argument('--weight_deacy', default=5e-4, help='learning alg momentum')
+ # parser.add_argument('--eval_datasets', default=['lfw', 'cfp_ff', 'cfp_fp', 'agedb_30'], help='evluation datasets')
+ parser.add_argument('--eval_datasets', default=['lfw', 'cfp_fp'], help='evluation datasets')
+ parser.add_argument('--eval_db_path', default='./datasets/faces_ms1m_112x112', help='evluate datasets base path')
+ parser.add_argument('--image_size', default=[112, 112], help='the image size')
+ parser.add_argument('--num_output', default=85164, help='the image size')
+ parser.add_argument('--tfrecords_file_path', default='./datasets/tfrecords', type=str,
+ help='path to the output of tfrecords file path')
+ parser.add_argument('--summary_path', default='./output/summary', help='the summary file save path')
+ parser.add_argument('--ckpt_path', default='./output/ckpt', help='the ckpt file save path')
+ parser.add_argument('--saver_maxkeep', default=100, help='tf.train.Saver max keep ckpt files')
+ parser.add_argument('--buffer_size', default=100000, help='tf dataset api buffer size')
+ parser.add_argument('--log_device_mapping', default=False, help='show device placement log')
+ parser.add_argument('--summary_interval', default=300, help='interval to save summary')
+ parser.add_argument('--ckpt_interval', default=5000, help='intervals to save ckpt file')
+ parser.add_argument('--validate_interval', default=2000, help='intervals to save ckpt file')
+ parser.add_argument('--show_info_interval', default=20, help='intervals to show information')
+ parser.add_argument('--num_gpus', default=2, help='the num of gpus')
+ parser.add_argument('--tower_name', default='tower', help='tower name')
+ args = parser.parse_args()
+ return args
+
+
+def average_gradients(tower_grads):
+ """Calculate the average gradient for each shared variable across all towers.
+
+ Note that this function provides a synchronization point across all towers.
+
+ Args:
+ tower_grads: List of lists of (gradient, variable) tuples. The outer list
+ is over individual gradients. The inner list is over the gradient
+ calculation for each tower.
+ Returns:
+ List of pairs of (gradient, variable) where the gradient has been averaged
+ across all towers.
+ """
+ average_grads = []
+ for grad_and_vars in zip(*tower_grads):
+ # Note that each grad_and_vars looks like the following:
+ # ((grad0_gpu0, var0_gpu0), ... , (grad0_gpuN, var0_gpuN))
+ grads = []
+ for g, _ in grad_and_vars:
+ # Add 0 dimension to the gradients to represent the tower.
+ expanded_g = tf.expand_dims(g, 0)
+
+ # Append on a 'tower' dimension which we will average over below.
+ grads.append(expanded_g)
+
+ # Average over the 'tower' dimension.
+ grad = tf.concat(axis=0, values=grads)
+ grad = tf.reduce_mean(grad, 0)
+
+ # Keep in mind that the Variables are redundant because they are shared
+ # across towers. So .. we will just return the first tower's pointer to
+ # the Variable.
+ v = grad_and_vars[0][1]
+ grad_and_var = (grad, v)
+ average_grads.append(grad_and_var)
+ return average_grads
+
+
+if __name__ == '__main__':
+ # os.environ["CUDA_VISIBLE_DEVICES"] = "0"
+ # 1. define global parameters
+ args = get_parser()
+ global_step = tf.Variable(name='global_step', initial_value=0, trainable=False)
+ inc_op = tf.assign_add(global_step, 1, name='increment_global_step')
+ images = tf.placeholder(name='img_inputs', shape=[None, *args.image_size, 3], dtype=tf.float32)
+ images_test = tf.placeholder(name='img_inputs', shape=[None, *args.image_size, 3], dtype=tf.float32)
+ labels = tf.placeholder(name='img_labels', shape=[None, ], dtype=tf.int64)
+ dropout_rate = tf.placeholder(name='dropout_rate', dtype=tf.float32)
+ # splits input to different gpu
+ images_s = tf.split(images, num_or_size_splits=args.num_gpus, axis=0)
+ labels_s = tf.split(labels, num_or_size_splits=args.num_gpus, axis=0)
+ # 2 prepare train datasets and test datasets by using tensorflow dataset api
+ # 2.1 train datasets
+ # the image is substracted 127.5 and multiplied 1/128.
+ # random flip left right
+ tfrecords_f = os.path.join(args.tfrecords_file_path, 'tran.tfrecords')
+ dataset = tf.data.TFRecordDataset(tfrecords_f)
+ dataset = dataset.map(parse_function)
+ dataset = dataset.shuffle(buffer_size=args.buffer_size)
+ dataset = dataset.batch(args.batch_size, drop_remainder=True)
+ iterator = dataset.make_initializable_iterator()
+ next_element = iterator.get_next()
+ # 2.2 prepare validate datasets
+ ver_list = []
+ ver_name_list = []
+ for db in args.eval_datasets:
+ print('begin db %s convert.' % db)
+ data_set = load_bin(db, args.image_size, args)
+ ver_list.append(data_set)
+ ver_name_list.append(db)
+ # 3. define network, loss, optimize method, learning rate schedule, summary writer, saver
+ # 3.1 inference phase
+ w_init_method = tf.contrib.layers.xavier_initializer(uniform=False)
+ # 3.2 define the learning rate schedule
+ p = int(512.0/args.batch_size)
+ lr_steps = [p*val for val in args.lr_steps]
+ print('learning rate steps: ', lr_steps)
+ lr = tf.train.piecewise_constant(global_step, boundaries=lr_steps, values=[0.001, 0.0005, 0.0003, 0.0001],
+ name='lr_schedule')
+ # 3.3 define the optimize method
+ opt = tf.train.MomentumOptimizer(learning_rate=lr, momentum=args.momentum)
+
+ # Calculate the gradients for each model tower.
+ tower_grads = []
+ tl.layers.set_name_reuse(True)
+ loss_dict = {}
+ drop_dict = {}
+ loss_keys = []
+ with tf.variable_scope(tf.get_variable_scope()):
+ for i in range(args.num_gpus):
+ with tf.device('/cpu:0'):
+ with tf.name_scope('%s_%d' % (args.tower_name, i)) as scope:
+ net = get_resnet(images_s[i], args.net_depth, type='ir', w_init=w_init_method, trainable=True, keep_rate=dropout_rate)
+ logit = arcface_loss(embedding=net.outputs, labels=labels_s[i], w_init=w_init_method, out_num=args.num_output)
+ # Reuse variables for the next tower.
+ tf.get_variable_scope().reuse_variables()
+ # define the cross entropy
+ inference_loss = tf.reduce_mean(tf.nn.sparse_softmax_cross_entropy_with_logits(logits=logit, labels=labels_s[i]))
+ # define weight deacy losses
+ wd_loss = 0
+ for weights in tl.layers.get_variables_with_name('W_conv2d', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(weights)
+ for W in tl.layers.get_variables_with_name('resnet_v1_50/E_DenseLayer/W', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(W)
+ for weights in tl.layers.get_variables_with_name('embedding_weights', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(weights)
+ for gamma in tl.layers.get_variables_with_name('gamma', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(gamma)
+ for alphas in tl.layers.get_variables_with_name('alphas', True, True):
+ wd_loss += tf.contrib.layers.l2_regularizer(args.weight_deacy)(alphas)
+ total_loss = inference_loss + wd_loss
+
+ loss_dict[('inference_loss_%s_%d' % ('gpu', i))] = inference_loss
+ loss_keys.append(('inference_loss_%s_%d' % ('gpu', i)))
+ loss_dict[('wd_loss_%s_%d' % ('gpu', i))] = wd_loss
+ loss_keys.append(('wd_loss_%s_%d' % ('gpu', i)))
+ loss_dict[('total_loss_%s_%d' % ('gpu', i))] = total_loss
+ loss_keys.append(('total_loss_%s_%d' % ('gpu', i)))
+ grads = opt.compute_gradients(total_loss)
+ tower_grads.append(grads)
+ if i == 0:
+ test_net = get_resnet(images_test, args.net_depth, type='ir', w_init=w_init_method, trainable=False, keep_rate=dropout_rate)
+ embedding_tensor = test_net.outputs
+ update_ops = tf.get_collection(tf.GraphKeys.UPDATE_OPS)
+ pred = tf.nn.softmax(logit)
+ acc = tf.reduce_mean(tf.cast(tf.equal(tf.argmax(pred, axis=1), labels_s[i]), dtype=tf.float32))
+
+ grads = average_gradients(tower_grads)
+ with tf.control_dependencies(update_ops):
+ # Apply the gradients to adjust the shared variables.
+ train_op = opt.apply_gradients(grads, global_step=global_step)
+
+ config = tf.ConfigProto(allow_soft_placement=True, log_device_placement=args.log_device_mapping)
+ config.gpu_options.allow_growth = True
+ sess = tf.Session(config=npu_config_proto(config_proto=config))
+ # summary writer
+ summary = tf.summary.FileWriter(args.summary_path, sess.graph)
+ summaries = []
+ # add grad histogram op
+ for grad, var in grads:
+ if grad is not None:
+ summaries.append(tf.summary.histogram(var.op.name + '/gradients', grad))
+ # add trainabel variable gradients
+ for var in tf.trainable_variables():
+ summaries.append(tf.summary.histogram(var.op.name, var))
+ # add loss summary
+ for keys, val in loss_dict.items():
+ summaries.append(tf.summary.scalar(keys, val))
+ # add learning rate
+ summaries.append(tf.summary.scalar('leraning_rate', lr))
+ summary_op = tf.summary.merge(summaries)
+
+ # Create a saver.
+ saver = tf.train.Saver(tf.global_variables())
+ # init all variables
+ sess.run(tf.global_variables_initializer())
+ # begin iteration
+ count = 0
+ for i in range(args.epoch):
+ sess.run(iterator.initializer)
+ while True:
+ try:
+ images_train, labels_train = sess.run(next_element)
+ feed_dict = {images: images_train, labels: labels_train, dropout_rate: 0.4}
+ start = time.time()
+ _, _, inference_loss_val_gpu_1, wd_loss_val_gpu_1, total_loss_gpu_1, inference_loss_val_gpu_2, \
+ wd_loss_val_gpu_2, total_loss_gpu_2, acc_val = sess.run([train_op, inc_op, loss_dict[loss_keys[0]],
+ loss_dict[loss_keys[1]],
+ loss_dict[loss_keys[2]],
+ loss_dict[loss_keys[3]],
+ loss_dict[loss_keys[4]],
+ loss_dict[loss_keys[5]], acc],
+ feed_dict=feed_dict)
+ end = time.time()
+ pre_sec = args.batch_size/(end - start)
+ # print training information
+ if count > 0 and count % args.show_info_interval == 0:
+ # print('epoch %d, total_step %d, total loss gpu 1 is %.2f , inference loss gpu 1 is %.2f, weight deacy '
+ # 'loss gpu 1 is %.2f, total loss gpu 2 is %.2f , inference loss gpu 2 is %.2f, weight deacy '
+ # 'loss gpu 2 is %.2f, training accuracy is %.6f, time %.3f samples/sec' %
+ # (i, count, total_loss_gpu_1, inference_loss_val_gpu_1, wd_loss_val_gpu_1, total_loss_gpu_2,
+ # inference_loss_val_gpu_2, wd_loss_val_gpu_2, acc_val, pre_sec))
+
+ print('epoch %d, total_step %d, total loss: [%.2f, %.2f], inference loss: [%.2f, %.2f], weight deacy '
+ 'loss: [%.2f, %.2f], training accuracy is %.6f, time %.3f samples/sec' %
+ (i, count, total_loss_gpu_1, total_loss_gpu_2, inference_loss_val_gpu_1, inference_loss_val_gpu_2,
+ wd_loss_val_gpu_1, wd_loss_val_gpu_2, acc_val, pre_sec))
+ count += 1
+
+ # save summary
+ if count > 0 and count % args.summary_interval == 0:
+ feed_dict = {images: images_train, labels: labels_train, dropout_rate: 0.4}
+ summary_op_val = sess.run(summary_op, feed_dict=feed_dict)
+ summary.add_summary(summary_op_val, count)
+
+ # save ckpt files
+ if count > 0 and count % args.ckpt_interval == 0:
+ filename = 'InsightFace_iter_{:d}'.format(count) + '.ckpt'
+ filename = os.path.join(args.ckpt_path, filename)
+ saver.save(sess, filename)
+ # # validate
+ if count >= 0 and count % args.validate_interval == 0:
+ feed_dict_test ={dropout_rate: 1.0}
+ results = ver_test(ver_list=ver_list, ver_name_list=ver_name_list, nbatch=count, sess=sess,
+ embedding_tensor=embedding_tensor, batch_size=args.batch_size//args.num_gpus, feed_dict=feed_dict_test,
+ input_placeholder=images_test)
+ if max(results) > 0.998:
+ print('best accuracy is %.5f' % max(results))
+ filename = 'InsightFace_iter_best_{:d}'.format(count) + '.ckpt'
+ filename = os.path.join(args.ckpt_path, filename)
+ saver.save(sess, filename)
+ except tf.errors.OutOfRangeError:
+ print("End of epoch %d" % i)
+ break
+
diff --git a/TensorFlow/contrib/cv/InsightFace_TF/verification.py b/TensorFlow/contrib/cv/InsightFace_TF/verification.py
new file mode 100644
index 0000000000000000000000000000000000000000..8d1fa2e99a605633ab4db30c6c97bec90b235499
--- /dev/null
+++ b/TensorFlow/contrib/cv/InsightFace_TF/verification.py
@@ -0,0 +1,252 @@
+"""Helper for evaluation on the Labeled Faces in the Wild dataset
+"""
+from npu_bridge.npu_init import *
+
+# MIT License
+#
+# Copyright (c) 2016 David Sandberg
+#
+# Permission is hereby granted, free of charge, to any person obtaining a copy
+# of this software and associated documentation files (the "Software"), to deal
+# in the Software without restriction, including without limitation the rights
+# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
+# copies of the Software, and to permit persons to whom the Software is
+# furnished to do so, subject to the following conditions:
+#
+# The above copyright notice and this permission notice shall be included in all
+# copies or substantial portions of the Software.
+#
+# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
+# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
+# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+# SOFTWARE.
+
+import tensorflow as tf
+import numpy as np
+from sklearn.model_selection import KFold
+from sklearn.decomposition import PCA
+import sklearn
+from scipy import interpolate
+import datetime
+import mxnet as mx
+
+
+def calculate_roc(thresholds, embeddings1, embeddings2, actual_issame, nrof_folds=10, pca=0):
+ assert (embeddings1.shape[0] == embeddings2.shape[0])
+ assert (embeddings1.shape[1] == embeddings2.shape[1])
+ nrof_pairs = min(len(actual_issame), embeddings1.shape[0])
+ nrof_thresholds = len(thresholds)
+ k_fold = KFold(n_splits=nrof_folds, shuffle=False)
+
+ tprs = np.zeros((nrof_folds, nrof_thresholds))
+ fprs = np.zeros((nrof_folds, nrof_thresholds))
+ accuracy = np.zeros((nrof_folds))
+ indices = np.arange(nrof_pairs)
+ # print('pca', pca)
+
+ if pca == 0:
+ diff = np.subtract(embeddings1, embeddings2)
+ dist = np.sum(np.square(diff), 1)
+
+ for fold_idx, (train_set, test_set) in enumerate(k_fold.split(indices)):
+ # print('train_set', train_set)
+ # print('test_set', test_set)
+ if pca > 0:
+ print('doing pca on', fold_idx)
+ embed1_train = embeddings1[train_set]
+ embed2_train = embeddings2[train_set]
+ _embed_train = np.concatenate((embed1_train, embed2_train), axis=0)
+ # print(_embed_train.shape)
+ pca_model = PCA(n_components=pca)
+ pca_model.fit(_embed_train)
+ embed1 = pca_model.transform(embeddings1)
+ embed2 = pca_model.transform(embeddings2)
+ embed1 = sklearn.preprocessing.normalize(embed1)
+ embed2 = sklearn.preprocessing.normalize(embed2)
+ # print(embed1.shape, embed2.shape)
+ diff = np.subtract(embed1, embed2)
+ dist = np.sum(np.square(diff), 1)
+
+ # Find the best threshold for the fold
+ acc_train = np.zeros((nrof_thresholds))
+ for threshold_idx, threshold in enumerate(thresholds):
+ _, _, acc_train[threshold_idx] = calculate_accuracy(threshold, dist[train_set], actual_issame[train_set])
+ best_threshold_index = np.argmax(acc_train)
+ print('best_threshold_index', best_threshold_index, acc_train[best_threshold_index])
+ for threshold_idx, threshold in enumerate(thresholds):
+ tprs[fold_idx, threshold_idx], fprs[fold_idx, threshold_idx], _ = calculate_accuracy(threshold,
+ dist[test_set],
+ actual_issame[
+ test_set])
+ _, _, accuracy[fold_idx] = calculate_accuracy(thresholds[best_threshold_index], dist[test_set],
+ actual_issame[test_set])
+
+ tpr = np.mean(tprs, 0)
+ fpr = np.mean(fprs, 0)
+ return tpr, fpr, accuracy
+
+
+def calculate_accuracy(threshold, dist, actual_issame):
+ predict_issame = np.less(dist, threshold)
+ tp = np.sum(np.logical_and(predict_issame, actual_issame))
+ fp = np.sum(np.logical_and(predict_issame, np.logical_not(actual_issame)))
+ tn = np.sum(np.logical_and(np.logical_not(predict_issame), np.logical_not(actual_issame)))
+ fn = np.sum(np.logical_and(np.logical_not(predict_issame), actual_issame))
+
+ tpr = 0 if (tp + fn == 0) else float(tp) / float(tp + fn)
+ fpr = 0 if (fp + tn == 0) else float(fp) / float(fp + tn)
+ acc = float(tp + tn) / dist.size
+ return tpr, fpr, acc
+
+
+def calculate_val(thresholds, embeddings1, embeddings2, actual_issame, far_target, nrof_folds=10):
+ '''
+ Copy from [insightface](https://github.com/deepinsight/insightface)
+ :param thresholds:
+ :param embeddings1:
+ :param embeddings2:
+ :param actual_issame:
+ :param far_target:
+ :param nrof_folds:
+ :return:
+ '''
+ assert (embeddings1.shape[0] == embeddings2.shape[0])
+ assert (embeddings1.shape[1] == embeddings2.shape[1])
+ nrof_pairs = min(len(actual_issame), embeddings1.shape[0])
+ nrof_thresholds = len(thresholds)
+ k_fold = KFold(n_splits=nrof_folds, shuffle=False)
+
+ val = np.zeros(nrof_folds)
+ far = np.zeros(nrof_folds)
+
+ diff = np.subtract(embeddings1, embeddings2)
+ dist = np.sum(np.square(diff), 1)
+ indices = np.arange(nrof_pairs)
+
+ for fold_idx, (train_set, test_set) in enumerate(k_fold.split(indices)):
+
+ # Find the threshold that gives FAR = far_target
+ far_train = np.zeros(nrof_thresholds)
+ for threshold_idx, threshold in enumerate(thresholds):
+ _, far_train[threshold_idx] = calculate_val_far(threshold, dist[train_set], actual_issame[train_set])
+ if np.max(far_train) >= far_target:
+ f = interpolate.interp1d(far_train, thresholds, kind='slinear')
+ threshold = f(far_target)
+ else:
+ threshold = 0.0
+
+ val[fold_idx], far[fold_idx] = calculate_val_far(threshold, dist[test_set], actual_issame[test_set])
+
+ val_mean = np.mean(val)
+ far_mean = np.mean(far)
+ val_std = np.std(val)
+ return val_mean, val_std, far_mean
+
+
+def calculate_val_far(threshold, dist, actual_issame):
+ predict_issame = np.less(dist, threshold)
+ true_accept = np.sum(np.logical_and(predict_issame, actual_issame))
+ false_accept = np.sum(np.logical_and(predict_issame, np.logical_not(actual_issame)))
+ n_same = np.sum(actual_issame)
+ n_diff = np.sum(np.logical_not(actual_issame))
+ val = float(true_accept) / float(n_same)
+ far = float(false_accept) / float(n_diff)
+ return val, far
+
+
+def evaluate(embeddings, actual_issame, nrof_folds=10, pca=0):
+ # Calculate evaluation metrics
+ thresholds = np.arange(0, 4, 0.01)
+ embeddings1 = embeddings[0::2]
+ embeddings2 = embeddings[1::2]
+ tpr, fpr, accuracy = calculate_roc(thresholds, embeddings1, embeddings2,
+ np.asarray(actual_issame), nrof_folds=nrof_folds, pca=pca)
+ thresholds = np.arange(0, 4, 0.001)
+ val, val_std, far = calculate_val(thresholds, embeddings1, embeddings2,
+ np.asarray(actual_issame), 1e-3, nrof_folds=nrof_folds)
+ return tpr, fpr, accuracy, val, val_std, far
+
+
+def data_iter(datasets, batch_size):
+ data_num = datasets.shape[0]
+ for i in range(0, data_num, batch_size):
+ yield datasets[i:min(i+batch_size, data_num), ...]
+
+
+def test(data_set, sess, embedding_tensor, batch_size, label_shape=None, feed_dict=None, input_placeholder=None):
+ '''
+ referenc official implementation [insightface](https://github.com/deepinsight/insightface)
+ :param data_set:
+ :param sess:
+ :param embedding_tensor:
+ :param batch_size:
+ :param label_shape:
+ :param feed_dict:
+ :param input_placeholder:
+ :return:
+ '''
+ print('testing verification..')
+ data_list = data_set[0]
+ issame_list = data_set[1]
+ embeddings_list = []
+ time_consumed = 0.0
+ for i in range(len(data_list)):
+ datas = data_list[i]
+ embeddings = None
+ feed_dict.setdefault(input_placeholder, None)
+ for idx, data in enumerate(data_iter(datas, batch_size)):
+ data_tmp = data.copy() # fix issues #4
+ data_tmp -= 127.5
+ data_tmp *= 0.0078125
+ feed_dict[input_placeholder] = data_tmp
+ time0 = datetime.datetime.now()
+ _embeddings = sess.run(embedding_tensor, feed_dict)
+ time_now = datetime.datetime.now()
+ diff = time_now - time0
+ time_consumed += diff.total_seconds()
+ if embeddings is None:
+ embeddings = np.zeros((datas.shape[0], _embeddings.shape[1]))
+ try:
+ embeddings[idx*batch_size:min((idx+1)*batch_size, datas.shape[0]), ...] = _embeddings
+ except ValueError:
+ print('idx*batch_size value is %d min((idx+1)*batch_size, datas.shape[0]) %d, batch_size %d, data.shape[0] %d' %
+ (idx*batch_size, min((idx+1)*batch_size, datas.shape[0]), batch_size, datas.shape[0]))
+ print('embedding shape is ', _embeddings.shape)
+ embeddings_list.append(embeddings)
+
+ _xnorm = 0.0
+ _xnorm_cnt = 0
+ for embed in embeddings_list:
+ for i in range(embed.shape[0]):
+ _em = embed[i]
+ _norm = np.linalg.norm(_em)
+ # print(_em.shape, _norm)
+ _xnorm += _norm
+ _xnorm_cnt += 1
+ _xnorm /= _xnorm_cnt
+
+ acc1 = 0.0
+ std1 = 0.0
+ embeddings = embeddings_list[0] + embeddings_list[1]
+ embeddings = sklearn.preprocessing.normalize(embeddings)
+ print(embeddings.shape)
+ print('infer time', time_consumed)
+ _, _, accuracy, val, val_std, far = evaluate(embeddings, issame_list, nrof_folds=10)
+ acc2, std2 = np.mean(accuracy), np.std(accuracy)
+ return acc1, std1, acc2, std2, _xnorm, embeddings_list
+
+
+def ver_test(ver_list, ver_name_list, nbatch, sess, embedding_tensor, batch_size, feed_dict, input_placeholder):
+ results = []
+ for i in range(len(ver_list)):
+ acc1, std1, acc2, std2, xnorm, embeddings_list = test(data_set=ver_list[i], sess=sess, embedding_tensor=embedding_tensor,
+ batch_size=batch_size, feed_dict=feed_dict,
+ input_placeholder=input_placeholder)
+ print('[%s][%d]XNorm: %f' % (ver_name_list[i], nbatch, xnorm))
+ print('[%s][%d]Accuracy-Flip: %1.5f+-%1.5f' % (ver_name_list[i], nbatch, acc2, std2))
+ results.append(acc2)
+ return results
+