Python基于keras训练实现微笑识别的示例详解
作者:醉意丶千层梦 发布时间:2022-06-14 10:02:07
标签:Python,keras,微笑识别
一、数据预处理
实验数据来自genki4k
提取含有完整人脸的图片
def init_file():
num = 0
bar = tqdm(os.listdir(read_path))
for file_name in bar:
bar.desc = "预处理图片: "
# a图片的全路径
img_path = (read_path + "/" + file_name)
# 读入的图片的路径中含非英文
img = cv2.imdecode(np.fromfile(img_path, dtype=np.uint8), cv2.IMREAD_UNCHANGED)
# 获取图片的宽高
img_shape = img.shape
img_height = img_shape[0]
img_width = img_shape[1]
# 用来存储生成的单张人脸的路径
# dlib检测
dets = detector(img, 1)
for k, d in enumerate(dets):
if len(dets) > 1:
continue
num += 1
# 计算矩形大小
# (x,y), (宽度width, 高度height)
# pos_start = tuple([d.left(), d.top()])
# pos_end = tuple([d.right(), d.bottom()])
# 计算矩形框大小
height = d.bottom() - d.top()
width = d.right() - d.left()
# 根据人脸大小生成空的图像
img_blank = np.zeros((height, width, 3), np.uint8)
for i in range(height):
if d.top() + i >= img_height: # 防止越界
continue
for j in range(width):
if d.left() + j >= img_width: # 防止越界
continue
img_blank[i][j] = img[d.top() + i][d.left() + j]
img_blank = cv2.resize(img_blank, (200, 200), interpolation=cv2.INTER_CUBIC)
# 保存图片
cv2.imencode('.jpg', img_blank)[1].tofile(save_path + "/" + "file" + str(num) + ".jpg")
logging.info("一共", len(os.listdir(read_path)), "个样本")
logging.info("有效样本", num)
二、训练模型
创建模型
# 创建网络
def create_model():
model = models.Sequential()
model.add(layers.Conv2D(32, (3, 3), activation='relu', input_shape=(150, 150, 3)))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Conv2D(64, (3, 3), activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Conv2D(128, (3, 3), activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Conv2D(128, (3, 3), activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Flatten())
model.add(layers.Dropout(0.5))
model.add(layers.Dense(512, activation='relu'))
model.add(layers.Dense(1, activation='sigmoid'))
model.compile(loss='binary_crossentropy',
optimizer=optimizers.RMSprop(lr=1e-4),
metrics=['acc'])
return model
训练模型
# 训练模型
def train_model(model):
# 归一化处理
train_datagen = ImageDataGenerator(
rescale=1. / 255,
rotation_range=40,
width_shift_range=0.2,
height_shift_range=0.2,
shear_range=0.2,
zoom_range=0.2,
horizontal_flip=True, )
test_datagen = ImageDataGenerator(rescale=1. / 255)
train_generator = train_datagen.flow_from_directory(
# This is the target directory
train_dir,
# All images will be resized to 150x150
target_size=(150, 150),
batch_size=32,
# Since we use binary_crossentropy loss, we need binary labels
class_mode='binary')
validation_generator = test_datagen.flow_from_directory(
validation_dir,
target_size=(150, 150),
batch_size=32,
class_mode='binary')
history = model.fit_generator(
train_generator,
steps_per_epoch=60,
epochs=12,
validation_data=validation_generator,
validation_steps=30)
# 保存模型
save_path = "../output/model"
if not os.path.exists(save_path):
os.makedirs(save_path)
model.save(save_path + "/smileDetect.h5")
return history
训练结果
准确率
丢失率
训练过程
三、预测
通过读取摄像头内容进行预测
def rec(img):
gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
dets = detector(gray, 1)
if dets is not None:
for face in dets:
left = face.left()
top = face.top()
right = face.right()
bottom = face.bottom()
cv2.rectangle(img, (left, top), (right, bottom), (0, 255, 0), 2)
img1 = cv2.resize(img[top:bottom, left:right], dsize=(150, 150))
img1 = cv2.cvtColor(img1, cv2.COLOR_BGR2RGB)
img1 = np.array(img1) / 255.
img_tensor = img1.reshape(-1, 150, 150, 3)
prediction = model.predict(img_tensor)
if prediction[0][0] > 0.5:
result = 'unsmile'
else:
result = 'smile'
cv2.putText(img, result, (left, top), font, 2, (0, 255, 0), 2, cv2.LINE_AA)
cv2.imshow('Video', img)
while video.isOpened():
res, img_rd = video.read()
if not res:
break
rec(img_rd)
if cv2.waitKey(1) & 0xFF == ord('q'):
break
效果
四、源代码
pretreatment.py
import dlib # 人脸识别的库dlib
import numpy as np # 数据处理的库numpy
import cv2 # 图像处理的库OpenCv
import os
import shutil
from tqdm import tqdm
import logging
# dlib预测器
detector = dlib.get_frontal_face_detector()
predictor = dlib.shape_predictor('../resources/shape_predictor_68_face_landmarks.dat')
# 原图片路径
read_path = "../resources/genki4k/files"
# 提取人脸存储路径
save_path = "../output/genki4k/files"
if not os.path.exists(save_path):
os.makedirs(save_path)
# 新的数据集
data_dir = '../resources/data'
if not os.path.exists(data_dir):
os.makedirs(data_dir)
# 训练集
train_dir = data_dir + "/train"
if not os.path.exists(train_dir):
os.makedirs(train_dir)
# 验证集
validation_dir = os.path.join(data_dir, 'validation')
if not os.path.exists(validation_dir):
os.makedirs(validation_dir)
# 测试集
test_dir = os.path.join(data_dir, 'test')
if not os.path.exists(test_dir):
os.makedirs(test_dir)
# 初始化训练数据
def init_data(file_list):
# 如果不存在文件夹则新建
for file_path in file_list:
if not os.path.exists(file_path):
os.makedirs(file_path)
# 存在则清空里面所有数据
else:
for i in os.listdir(file_path):
path = os.path.join(file_path, i)
if os.path.isfile(path):
os.remove(path)
def init_file():
num = 0
bar = tqdm(os.listdir(read_path))
for file_name in bar:
bar.desc = "预处理图片: "
# a图片的全路径
img_path = (read_path + "/" + file_name)
# 读入的图片的路径中含非英文
img = cv2.imdecode(np.fromfile(img_path, dtype=np.uint8), cv2.IMREAD_UNCHANGED)
# 获取图片的宽高
img_shape = img.shape
img_height = img_shape[0]
img_width = img_shape[1]
# 用来存储生成的单张人脸的路径
# dlib检测
dets = detector(img, 1)
for k, d in enumerate(dets):
if len(dets) > 1:
continue
num += 1
# 计算矩形大小
# (x,y), (宽度width, 高度height)
# pos_start = tuple([d.left(), d.top()])
# pos_end = tuple([d.right(), d.bottom()])
# 计算矩形框大小
height = d.bottom() - d.top()
width = d.right() - d.left()
# 根据人脸大小生成空的图像
img_blank = np.zeros((height, width, 3), np.uint8)
for i in range(height):
if d.top() + i >= img_height: # 防止越界
continue
for j in range(width):
if d.left() + j >= img_width: # 防止越界
continue
img_blank[i][j] = img[d.top() + i][d.left() + j]
img_blank = cv2.resize(img_blank, (200, 200), interpolation=cv2.INTER_CUBIC)
# 保存图片
cv2.imencode('.jpg', img_blank)[1].tofile(save_path + "/" + "file" + str(num) + ".jpg")
logging.info("一共", len(os.listdir(read_path)), "个样本")
logging.info("有效样本", num)
# 划分数据集
def divide_data(file_path, message, begin, end):
files = ['file{}.jpg'.format(i) for i in range(begin, end)]
bar = tqdm(files)
bar.desc = message
for file in bar:
src = os.path.join(save_path, file)
dst = os.path.join(file_path, file)
shutil.copyfile(src, dst)
if __name__ == "__main__":
init_file()
positive_train_dir = os.path.join(train_dir, 'smile')
negative_train_dir = os.path.join(train_dir, 'unSmile')
positive_validation_dir = os.path.join(validation_dir, 'smile')
negative_validation_dir = os.path.join(validation_dir, 'unSmile')
positive_test_dir = os.path.join(test_dir, 'smile')
negative_test_dir = os.path.join(test_dir, 'unSmile')
file_list = [positive_train_dir, positive_validation_dir, positive_test_dir,
negative_train_dir, negative_validation_dir, negative_test_dir]
init_data(file_list)
divide_data(positive_train_dir, "划分训练集正样本", 1, 1001)
divide_data(negative_train_dir, "划分训练集负样本", 2200, 3200)
divide_data(positive_validation_dir, "划分验证集正样本", 1000, 1500)
divide_data(negative_validation_dir, "划分验证集负样本", 3000, 3500)
divide_data(positive_test_dir, "划分测试集正样本", 1500, 2000)
divide_data(negative_test_dir, "划分测试集负样本", 2800, 3500)
train.py
import os
from keras import layers
from keras import models
from tensorflow import optimizers
import matplotlib.pyplot as plt
from keras.preprocessing.image import ImageDataGenerator
train_dir = "../resources/data/train"
validation_dir = "../resources/data/validation"
# 创建网络
def create_model():
model = models.Sequential()
model.add(layers.Conv2D(32, (3, 3), activation='relu', input_shape=(150, 150, 3)))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Conv2D(64, (3, 3), activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Conv2D(128, (3, 3), activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Conv2D(128, (3, 3), activation='relu'))
model.add(layers.MaxPooling2D((2, 2)))
model.add(layers.Flatten())
model.add(layers.Dropout(0.5))
model.add(layers.Dense(512, activation='relu'))
model.add(layers.Dense(1, activation='sigmoid'))
model.compile(loss='binary_crossentropy',
optimizer=optimizers.RMSprop(lr=1e-4),
metrics=['acc'])
return model
# 训练模型
def train_model(model):
# 归一化处理
train_datagen = ImageDataGenerator(
rescale=1. / 255,
rotation_range=40,
width_shift_range=0.2,
height_shift_range=0.2,
shear_range=0.2,
zoom_range=0.2,
horizontal_flip=True, )
test_datagen = ImageDataGenerator(rescale=1. / 255)
train_generator = train_datagen.flow_from_directory(
# This is the target directory
train_dir,
# All images will be resized to 150x150
target_size=(150, 150),
batch_size=32,
# Since we use binary_crossentropy loss, we need binary labels
class_mode='binary')
validation_generator = test_datagen.flow_from_directory(
validation_dir,
target_size=(150, 150),
batch_size=32,
class_mode='binary')
history = model.fit_generator(
train_generator,
steps_per_epoch=60,
epochs=12,
validation_data=validation_generator,
validation_steps=30)
# 保存模型
save_path = "../output/model"
if not os.path.exists(save_path):
os.makedirs(save_path)
model.save(save_path + "/smileDetect.h5")
return history
# 展示训练结果
def show_results(history):
# 数据增强过后的训练集与验证集的精确度与损失度的图形
acc = history.history['acc']
val_acc = history.history['val_acc']
loss = history.history['loss']
val_loss = history.history['val_loss']
# 绘制结果
epochs = range(len(acc))
plt.plot(epochs, acc, 'bo', label='Training acc')
plt.plot(epochs, val_acc, 'b', label='Validation acc')
plt.title('Training and validation accuracy')
plt.legend()
plt.figure()
plt.plot(epochs, loss, 'bo', label='Training loss')
plt.plot(epochs, val_loss, 'b', label='Validation loss')
plt.title('Training and validation loss')
plt.legend()
plt.show()
if __name__ == "__main__":
model = create_model()
history = train_model(model)
show_results(history)
predict.py
import os
from keras import layers
from keras import models
from tensorflow import optimizers
import matplotlib.pyplot as plt
from keras.preprocessing.image import ImageDataGenerator
train_dir = "../resources/data/train"
validation_dir = "../resources/data/validation"
# 创建网络
# 检测视频或者摄像头中的人脸
import cv2
from keras.preprocessing import image
from keras.models import load_model
import numpy as np
import dlib
from PIL import Image
model = load_model('../output/model/smileDetect.h5')
detector = dlib.get_frontal_face_detector()
video = cv2.VideoCapture(0)
font = cv2.FONT_HERSHEY_SIMPLEX
def rec(img):
gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
dets = detector(gray, 1)
if dets is not None:
for face in dets:
left = face.left()
top = face.top()
right = face.right()
bottom = face.bottom()
cv2.rectangle(img, (left, top), (right, bottom), (0, 255, 0), 2)
img1 = cv2.resize(img[top:bottom, left:right], dsize=(150, 150))
img1 = cv2.cvtColor(img1, cv2.COLOR_BGR2RGB)
img1 = np.array(img1) / 255.
img_tensor = img1.reshape(-1, 150, 150, 3)
prediction = model.predict(img_tensor)
if prediction[0][0] > 0.5:
result = 'unsmile'
else:
result = 'smile'
cv2.putText(img, result, (left, top), font, 2, (0, 255, 0), 2, cv2.LINE_AA)
cv2.imshow('Video', img)
while video.isOpened():
res, img_rd = video.read()
if not res:
break
rec(img_rd)
if cv2.waitKey(1) & 0xFF == ord('q'):
break
video.release()
cv2.destroyAllWindows()
来源:https://blog.csdn.net/qq_47281915/article/details/122278405


猜你喜欢
- 作者认为最快的学习区块链的方式是自己创建一个,本文就跟随作者用Python来创建一个区块链。对数字货币的崛起感到新奇的我们,并且想知道其背后
- 本文介绍了Scrapy项目实战之爬取某社区用户详情,分享给大家,具有如下:get_cookies.pyfrom selenium impor
- 本文实例讲述了彻底删除thinkphp3.1案例blog标签的方法。分享给大家供大家参考。具体方法如下:thinkphp3.1框架中的案例b
- 环境搭建准备工具:pyton3.5,selenium,phantomjs我的电脑里面已经装好了python3.5安装Seleniumpip3
- 前言JavaScript 中的 Infinity 是一个可以应用于任何变量的数值,表示无穷大。下面就来看看 Infinity 是如何工作的,
- 我们可以用动态产生变量的方法,从表格里捕捉数据,动态地创造“剥离”变量引号并且“清理”它,见下列代码,我们只需键入变量名称,选择 query
- 这最近在PJ的function库里看到的这个函数,感觉思路差了点,不过相对比较完美,只是闭合标签时的顺序问题,呵呵 修改一下数组arrTag
- 在知乎上遇到一个问题,说:计算机中的「null」怎么读?null正确的发音是/n^l/,有点类似四声‘纳儿&rs
- 别人复制你网站的文章时自动加上注释,这个功能你在很多网站应该都有体会过,当我们复制一段内容时,就自动在文章后面加上了网站的一些
- 1.首先准备好VS2019以及mysql数据库,两者都可以去官网下载,我们直接描述连接过程。2.连接:第一步:打开mysql的安装目录,我本
- 在list中嵌套元组,在进行sort排序的时候,产生的是原数组的副本,排序过程中,先根据第一个字段进行从小到大排序,如果第一个字段相同的话,
- 实现了宽度、高度、透明度的渐变,还能以高度宽度中点为中心,还扩展成以任意点为中心渐变(实例中以点击点为中心)。<!DOCTYPE ht
- 前言最近在公司售前售后同事遇到一些奇怪的需求找到我,需要提供公司一些项目数据库所有表的结构信息(字段名、类型、长度、是否主键、***、备注)
- 本文以sql2012为例第一步,(查询日志中文件名和大小):USE 数据库名GOSELECT file_id, name,size,* FR
- 环境:Python+keras,后端为Tensorflow训练集:MNIST对于如何训练一个识别手写数字的神经网络,网上资源十分丰富,并且能
- 背景对接多个外部接口,需要保存请求参数以及返回参数,方便消息的补偿,因为多个外部接口,多个接口字段都不统一,整体使用一个大字段(longte
- 本文实例为大家分享了Python 12306抢火车票的具体代码,供大家参考,具体内容如下# -*- coding: utf-8 -*-fro
- 常用目标检测模型基本都是读取的PASCAL VOC格式的标签,下面代码用于生成VOC格式的代码,根据需要修改即可:from lxml imp
- 数据存储·在javascript中,数据存储的位置会对代码整体性能产生重大的影响。·数据存储共有4种方式:字面量、变量、数组、对象成员。·要
- 按单词反转字符串是一道很常见的面试题。在Python中实现起来非常简单。def reverse_string_by_word(s): lst