Pass on README.
This commit is contained in:
parent
2fc6b0e3de
commit
7d9af6cab1
40
README.md
40
README.md
|
@ -4,7 +4,7 @@ This is a Python and Torch implementation of the CVPR 2015 paper
|
||||||
[FaceNet: A Unified Embedding for Face Recognition and Clustering](http://www.cv-foundation.org/openaccess/content_cvpr_2015/app/1A_089.pdf)
|
[FaceNet: A Unified Embedding for Face Recognition and Clustering](http://www.cv-foundation.org/openaccess/content_cvpr_2015/app/1A_089.pdf)
|
||||||
by Florian Schroff, Dmitry Kalenichenko, and James Philbin at Google
|
by Florian Schroff, Dmitry Kalenichenko, and James Philbin at Google
|
||||||
using publicly available libraries and datasets.
|
using publicly available libraries and datasets.
|
||||||
Torch allows CPU and CUDA
|
Torch allows the network to be executed on a CPU or with CUDA.
|
||||||
|
|
||||||
**Crafted by [Brandon Amos](http://bamos.github.io) in the
|
**Crafted by [Brandon Amos](http://bamos.github.io) in the
|
||||||
[Elijah](http://elijah.cs.cmu.edu) research group at
|
[Elijah](http://elijah.cs.cmu.edu) research group at
|
||||||
|
@ -27,7 +27,7 @@ image of Sylvestor Stallone from the publicly available
|
||||||
to try to make the eyes and nose appear in
|
to try to make the eyes and nose appear in
|
||||||
the same location on each image.
|
the same location on each image.
|
||||||
3. Use a deep neural network to represent (or embed) the face on
|
3. Use a deep neural network to represent (or embed) the face on
|
||||||
a 128-dimensional hypersphere.
|
a 128-dimensional unit hypersphere.
|
||||||
The embedding is a generic representation for anybody's face.
|
The embedding is a generic representation for anybody's face.
|
||||||
Unlike other face representations, this embedding has the nice property
|
Unlike other face representations, this embedding has the nice property
|
||||||
that a larger distance between two face embeddings means
|
that a larger distance between two face embeddings means
|
||||||
|
@ -58,11 +58,18 @@ Please contact Brandon Amos at [bamos@cs.cmu.edu](mailto:bamos@cs.cmu.edu).
|
||||||
# Real-Time Web Demo
|
# Real-Time Web Demo
|
||||||
See [our YouTube video](TODO) of using this in a real-time web application
|
See [our YouTube video](TODO) of using this in a real-time web application
|
||||||
for face recognition.
|
for face recognition.
|
||||||
The source is available in [demos/www](/demos/www).
|
The source is available in [examples/web](/examples/web).
|
||||||
|
|
||||||
TODO: Screenshot
|
TODO: Screenshot
|
||||||
|
|
||||||
# Cool demo, but I want numbers. What's the accuracy?
|
From the `examples/web` directory, install requirements
|
||||||
|
with `./install-deps.sh` and `sudo pip install -r requirements.txt`.
|
||||||
|
|
||||||
|
# Comparison Demo
|
||||||
|
Use `./demos/compare.py` to compute the squared Euclidean
|
||||||
|
distance of faces found in two images.
|
||||||
|
|
||||||
|
# Cool demos, but I want numbers. What's the accuracy?
|
||||||
Even though the public datasets we trained on have orders of magnitude less data
|
Even though the public datasets we trained on have orders of magnitude less data
|
||||||
than private industry datasets, the accuracy is remarkably high and
|
than private industry datasets, the accuracy is remarkably high and
|
||||||
outperforms all other open-source face recognition implementations we
|
outperforms all other open-source face recognition implementations we
|
||||||
|
@ -90,6 +97,17 @@ in `./data/lfw/raw` and `./data/lfw/deepfunneled`.
|
||||||
4. Generate the ROC curve from the `evaluation` directory with `./lfw-roc.py --workDir lfw.nn4.v1.reps`.
|
4. Generate the ROC curve from the `evaluation` directory with `./lfw-roc.py --workDir lfw.nn4.v1.reps`.
|
||||||
This creates `roc.pdf` in the `lfw.nn4.v1.reps` directory.
|
This creates `roc.pdf` in the `lfw.nn4.v1.reps` directory.
|
||||||
|
|
||||||
|
# What's in this repository?
|
||||||
|
+ [batch-represent](/batch-represent): Generate representations from
|
||||||
|
a batch of images, stored in a directory by names.
|
||||||
|
+ [demos/www](/demos/www): Real-time web demo.
|
||||||
|
+ [demos/compare.py](/demos/compare.py): Compare two images.
|
||||||
|
+ [evaluation](/evaluation): LFW accuracy evaluation scripts.
|
||||||
|
+ [facenet](/facenet): Python library code.
|
||||||
|
+ [images](/images): Images used in the README.
|
||||||
|
+ [models](/models): Location of binary models.
|
||||||
|
+ [training](/training): Scripts to train new models.
|
||||||
|
|
||||||
# Setup
|
# Setup
|
||||||
|
|
||||||
## Check out git submodules
|
## Check out git submodules
|
||||||
|
@ -121,6 +139,20 @@ access them from the shared Docker directory.
|
||||||
## By hand
|
## By hand
|
||||||
TODO
|
TODO
|
||||||
|
|
||||||
|
### Install dlib
|
||||||
|
Download dlib from [here](https://github.com/davisking/dlib/releases/download/v18.16/dlib-18.16.tar.bz2).
|
||||||
|
|
||||||
|
```
|
||||||
|
cd ~/src
|
||||||
|
tar xf dlib-18.16.tar.bz2
|
||||||
|
cd dlib-18.16/python_examples
|
||||||
|
mkdir build
|
||||||
|
cd build
|
||||||
|
cmake ../../tools/python
|
||||||
|
cmake --build . --config Release
|
||||||
|
cp dlib.so ..
|
||||||
|
```
|
||||||
|
|
||||||
Dependencies:
|
Dependencies:
|
||||||
+ [torch7](https://github.com/torch/torch7)
|
+ [torch7](https://github.com/torch/torch7)
|
||||||
+ [dpnn](https://github.com/nicholas-leonard/dpnn)
|
+ [dpnn](https://github.com/nicholas-leonard/dpnn)
|
||||||
|
|
Loading…
Reference in New Issue