内容简介:The FaceNet Keras model is available onTo implement a live Camera feed, we use CameraX. I have used the code available in theAll our classification code will come in the
1. Convert the Keras model to a TFLite model
The FaceNet Keras model is available on
nyoki-mtl/keras-facenet
repo. After downloading the .h5
model, we’ll use the tf.lite.TFLiteConverter
API to convert our Keras model to a TFLite model.
2. Setting up a Preview and ImageAnalyser using CameraX
To implement a live Camera feed, we use CameraX. I have used the code available in the official docs
. Next, we create a FrameAnalyser
class which implements ImageAnalysis
class, which will help us retrieve camera frames and run inference on them.
All our classification code will come in the analyze
method. First, using Firebase MLKit, we’ll get bounding boxes for all faces present in the camera frame ( a Bitmap
object ). We’ll create a FirebaseVisionFaceDetector
which runs the face detection model on an FirebaseVisionInputImage
object.
3. Producing Face Embeddings using FaceNet and Comparing them.
First, we’ll produce face embeddings using our FaceNet model. Before, we’ll create a helper class for handling the FaceNet model. This helper class will,
-
Crop the given camera frame using the bounding box ( as
Rect) which we got from Firebase MLKit. -
Transform this cropped image from a
Bitmapto aByteBufferwith normalized pixel values. -
Finally, feed the
ByteBufferto our FaceNet model using theInterpreterclass provided by TF Lite Android library.
In the snippet below, see the getFaceEmbedding()
method which encapsulates all the above steps.
Now, we have a class that would return us the 128-dimensional embedding for all faces present in the given image. We come back to a FrameAnalyser
‘s analyze()
method. Using the helper class which just created, we’ll produce face embeddings and compare each of them with a set of embeddings that we already have.
Before that, we need to get the set of predefined embeddings, right? These embeddings will refer to the people whom we need to recognize. So, the app will read the images
folder present in the internal storage of the user’s device. If the user wants to recognize, two users, namely Rahul
and Neeta
, then he/she needs to create two separate directories within the images
folder. Then he/she has to place an image of Rahul
and Neeta
in their respective sub-directories.
images ->
rahul ->
image_rahul.png neeta ->
image_neeta.png
Our aim to read these images and produce a HashMap<String,FloatArray>
object where the key ( String
)will the subject’s name like Rahul
or Neeta
and the value ( FloatArray
) will the corresponding face embedding. You’ll get an idea of the process like studying the code below.
We’ll compare the embeddings using the cosine similarity metrics which will return a similarity score in the interval [ -1 , 1 ]
.
The predictions
array is then supplied to the boundingBoxOverlay
class which draws the bounding boxes and also displays the label. In the BoundingBoxOverlay.kt
class,
The Results
Using the app, I have tried to recognize the faces of Jeff Bezos and Elon Musk,
Also, I had stored in the images in my internal storage as such,
The End
I hope you liked the story. I have included any APK in the GitHub repo so that you can try the app on your device. Thanks for reading!
以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持 码农网
猜你喜欢:本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们。
数据结构与算法分析
张琨、张宏、朱保平 / 人民邮电出版社 / 2016-2-1 / 45
本书共分10章,主要包括第1章绪论,第2章线性表,第3章栈和队列,第4章串,第5章数组和广义表,第6章 树和二叉树,第7章图,第8章查找,第9章内部排序,第10章算法分析。其内容模块涵盖了课堂教学、习题课教学、实验教学、自学辅导、综合训练等。立体化教材的使用在提高教学效率、增强教学效果、加大教学信息量、培养学生的应用与实践能力。 作者简介一起来看看 《数据结构与算法分析》 这本书的介绍吧!
JSON 在线解析
在线 JSON 格式化工具
RGB CMYK 转换工具
RGB CMYK 互转工具