Commit Graph

8031 Commits

Author SHA1 Message Date
Davis King 0057461a62 Promote some of the sub-network methods into the add_loss_layer interface so users don't have to write .subnet() so often. 2020-03-29 12:17:56 -04:00
Davis King c79f64f52d make update_parameters() a little more uniform 2020-03-29 11:19:37 -04:00
Davis King fd0145345e Fix Use of C++17 deprecated feature: std::iterator #2036 2020-03-29 11:08:07 -04:00
Adrià Arrufat f42f100d0f
Add DCGAN example (#2035)
* wip: dcgan-example

* wip: dcgan-example

* update example to use leaky_relu and remove bias from net

* wip

* it works!

* add more comments

* add visualization code

* add example documentation

* rename example

* fix comment

* better comment format

* fix the noise generator seed

* add message to hit enter for image generation

* fix srand, too

* add std::vector overload to update_parameters

* improve training stability

* better naming of variables

make sure it is clear we update the generator with the discriminator's
gradient using fake samples and true labels

* fix comment: generator -> discriminator

* update leaky_relu docs to match the relu ones

* replace not with !

* add Davis' suggestions to make training more stable

* use tensor instead of resizable_tensor

* do not use dnn_trainer for discriminator
2020-03-29 11:07:38 -04:00
Adrià Arrufat d610e56c2a
add leaky_relu activation layer (#2033)
* add leaky_relu activation layer

* add inplace case for leaky_relu and test_layer

* make clear that alpha is not learned by leaky_relu

* remove branch from cuda kernel
2020-03-21 11:29:20 -04:00
Juha Reunanen 74123841bb
To avoid a GPU memory leak, allow passing thread pools to dnn_trainer from outside (#2027)
* Problem: The CUDA runtime allocates resources for each thread, and apparently those resources are not freed when the corresponding threads terminate. Therefore, each instantiation of dnn_trainer leaks a bit of GPU memory.

Solution: Add possibility to pass thread pools from outside. This way, subsequent dnn_trainer instances can use the same threads, and there's no memory leak.

* Add helpful comments
2020-03-19 07:38:43 -04:00
scott-vsi 6fc503d242
link against openblasp (#2028)
openblasp is a parallel implementation of openblas with pthreads found on Centos/Fedora
2020-03-18 21:45:51 -04:00
Adrià Arrufat 1380e6b95f
add loss multiclass log weighted (#2022)
* add loss_multiclass_log_weighted

* fix class name in loss_abstract

* add loss_multiclass_log_weighted test

* rename test function to match class name

* fix typo

* reuse the weighted label struct across weighted losses

* do not break compatibility with loss_multiclass_log_per_pixel_weighted

* actually test the loss and fix docs

* fix build with gcc 9
2020-03-18 08:33:54 -04:00
hwiesmann 9185a925ce
Integer conversions generating compiler warnings (#2024)
* Prevention of compiler warning due to usage of int instead of a size type

* Conversion of status type to long to prevent compiler warnings

* The returned number of read items from a buffer is specified in numbers of type "streamsize"

Co-authored-by: Hartwig <git@skywind.eu>
2020-03-14 19:12:04 -04:00
Facundo Galán 08aeada7d5
Replace result_of by invoke_result for C++17 and above (#2021)
Co-authored-by: Facundo Galan <fgalan@danaide.com.ar>
2020-03-13 07:53:40 -04:00
scott-vsi c8a175f569
effect -> affect (#2019) 2020-03-11 22:56:07 -04:00
Davis King 7b35d7b234 removed inappropriate assert 2020-03-10 20:42:42 -04:00
hwiesmann e7087e5957
Prevention of compiler warning (#2015)
Co-authored-by: Hartwig <git@skywind.eu>
2020-03-10 20:02:02 -04:00
Adrià Arrufat c832d3b2fc
simplify resnet definition by reusing struct template parameter (#2010)
* simplify definition by reusing struct template parameter

* put resnet into its own namespace

* fix infer names

* rename struct impl to def
2020-03-09 21:21:04 -04:00
Davis King 3a53c78ad2 increment imglab version 2020-02-29 09:34:36 -05:00
Davis King 9a33669610 A little bit of cleanup and docs. Also added missing mutex lock. 2020-02-29 09:33:00 -05:00
martin 4ff365a530
imglab: chinese ("automatic") clustering, keyboard shortcuts for zooming (#2007)
* imglab: add support for using chinese whispers for more automatic clustering

* widgets: refactor out zooming from wheel handling

* tools/imglab/src/metadata_editor.cpp

imglab: add keyboard shortcuts for zooming
2020-02-29 09:31:28 -05:00
Davis King fc6992ac04 A little bit of cleanup 2020-02-07 08:12:18 -05:00
Adrià Arrufat 10d7f119ca
Add dnn_introduction3_ex (#1991)
* Add dnn_introduction3_ex
2020-02-07 07:59:36 -05:00
Davis King c90cb0bc14 Remove unit tests for python 2.7 since that version of pyhton is dead,
and the unit test servers don't even support it anymore.
2020-01-30 19:41:26 -05:00
Davis King f5c828291d Added note about vcpk availability 2020-01-30 19:36:49 -05:00
Davis King b13840a86f Fixed code needing C++14 to use C++11 features instead. 2020-01-29 07:30:59 -05:00
Juha Reunanen 46bcd2059e
If nearest-neighbor interpolation is wanted, then don't use an image pyramid. (#1986) 2020-01-28 21:03:39 -05:00
Hye Sung Jung 443021882c
fix spelling errors (#1985) 2020-01-28 21:02:41 -05:00
Julien Schueller 870f49a636
Do not link to libnsl (#1987)
Dlib does not use nsl symbols, why was this necessary ?
This make conda-forge build fail
2020-01-28 21:01:14 -05:00
Davis King 20d02b80e7 run tests for python 3.8 on travis ci 2020-01-27 07:53:19 -05:00
Davis King 0c415dbb4c Add little test 2020-01-20 07:58:50 -05:00
Davis King f71e49f28e remove unused variables 2020-01-20 07:58:07 -05:00
Davis King d88b2575a1 Make copying non-const cuda_data_ptrs to const ones nicer. 2020-01-20 07:57:33 -05:00
Juha Reunanen bd6994cc66 Add new loss layer for binary loss per pixel (#1976)
* Add new loss layer for binary loss per pixel
2020-01-20 07:47:47 -05:00
Davis King 6bdd289f73 Added static_pointer_cast() for casting cuda_data_void_ptr to
cuda_data_ptr<T>.  Also moved some memcpy() functions to namespace scope
so that calling them like dlib::cuda::memcpy() can referene them.  It
was slightly annoting before.
2020-01-18 13:27:25 -05:00
Davis King 2326a72281 fix code not compiling with some versions of libjpeg as a result of the change I just made. 2020-01-18 13:07:59 -05:00
Davis King e1b667181b Fixed const correctness on the in-memory jpeg loading code. 2020-01-18 11:53:46 -05:00
Davis King a0af6b7afd tweaked docs 2020-01-17 20:33:47 -05:00
Adrià Arrufat 60dad52c12 add visitor to count net parameters (#1977) 2020-01-17 20:32:19 -05:00
Juha Reunanen 356bba38fe Minor fix: print to console only if the verbose flag is on (#1980) 2020-01-16 20:23:47 -05:00
Manjunath Bhat d766f5e82e Adding Mish activation function (#1938)
* Adding Mish activation function

* Bug fixed

* Added test for Mish

* Removed unwanted comments

* Simplified calculation and removed comments

* Kernel added and gradient computation simplified

* Gradient simplified

* Corrected gradient calculations

* Compute output when input greater than 8

* Minor correction

* Remove unnecessary pgrad for Mish

* Removed CUDNN calls

* Add standalone CUDA implementation of the Mish activation function

* Fix in-place gradient in the CUDA version; refactor a little

* Swap delta and omega

* Need to have src (=x) (and not dest) available for Mish

* Add test case that makes sure that cuda::mish and cpu::mish return the same results

* Minor tweaking to keep the previous behaviour

Co-authored-by: Juha Reunanen <juha.reunanen@tomaattinen.com>
2020-01-15 06:04:02 -05:00
Davis King a82bf1536e omg, github, cleanup merge stuff 2020-01-14 21:26:25 -05:00
Davis King b70c0a6f80 Merge branch 'thebhatman-Mish' 2020-01-14 21:24:57 -05:00
thebhatman c454bdc182 Added test for Mish 2020-01-14 21:24:18 -05:00
Davis King 8a91a7c7c1 Merge branch 'thebhatman-Mish' 2020-01-14 21:19:15 -05:00
thebhatman edff12d2e1 Adding Mish activation function 2020-01-14 21:18:28 -05:00
Davis King cd5f0b0554 fixed failing tests due to recent default change in solver stopping criteria 2020-01-13 08:00:51 -05:00
Davis King 0c42dcca8d fixed test failure 2020-01-12 23:17:59 -05:00
Davis King adce342366 adjusted eps so tests still pass 2020-01-12 21:20:25 -05:00
Davis King 931fb52659 fixed test not building due to the commit I just made 2020-01-12 19:54:07 -05:00
Davis King 59d1b9d8c5 Added a relative epsilon termination option to svm_c_linear_trainer 2020-01-12 19:48:26 -05:00
Davis King 45731b863c Always check that the data give to cross_validate_trainer() is valid.
It's a cheap check, and easy for someone to forget about otherwise.
2020-01-12 19:36:53 -05:00
Davis King eae3caf9f8 Fixed function_evaluation_request::set() invalidating function_evaluation_request::x() 2020-01-11 21:12:34 -05:00
Davis King a53354eb79 updated docs 2020-01-09 22:31:16 -05:00