Commit Graph

66 Commits

Author SHA1 Message Date
3969ce30ed Speedup Tensor Cores: 1st layer uses FP32 and pre-allocate GPU memory for Tensor Cores 2018-12-11 23:48:58 +03:00
25f133d6ef Another one minor fix 2018-12-11 21:26:36 +03:00
cb998db949 Some fix for CUDNN_HALF 2018-12-11 21:16:18 +03:00
a621235783 Switch to Tensor Cores after 2000 iterations. 2018-12-10 01:35:08 +03:00
dc7f8a32ae mAP calculation during training, if is used flag -map 2018-12-09 18:18:47 +03:00
742bb7c7ce Compile fix 2018-12-07 22:52:07 +03:00
7c2f302321 Fixed nan issue for training with CUDNN_HALF=1 by using Tensor Cores 2018-12-07 22:40:10 +03:00
21a4ec9390 Saving loss-chart for each 100 iterations automatically 2018-11-26 11:11:56 +03:00
9f7d7c58b5 Minor fixes. Use CUDA 10.0 2018-11-17 02:48:46 +03:00
25f65f6878 Added fast_binarize_weights_gpu() 2018-11-05 22:38:35 +03:00
c0e2512af2 Activation improvement, more robust timer. 2018-09-27 23:10:54 +03:00
7dd97537fb XNOR-net tiny-yolo_xnor.cfg ~2x faster than cuDNN on CUDA (nVidia GPU Maxwell) 2018-09-22 02:01:14 +03:00
03e95320a1 XNOR coalesced memory access, and avoid bank conflicts 2018-09-17 23:39:25 +03:00
ca43bbdaae Fixed openmp bugs for XNOR 2018-09-12 16:22:54 +03:00
c0e01fd63c Test for XNOR-conv on CUDA 2018-09-08 02:46:05 +03:00
b141f85cab Compile fix 2018-09-07 15:07:46 +03:00
007878393f Temporary Slow implementation of XNOR on CUDA (shared_memory) 2018-09-06 23:21:26 +03:00
c4a9e3422e Temporary implementation of XNOR on CUDA 2018-08-31 02:47:58 +03:00
9753b72aeb temp fix, don't use it 2018-08-30 17:24:41 +03:00
cfc5fedbb6 Just used spaces for indents instead of Tabs 2018-07-10 23:29:15 +03:00
9bae70b225 Accelerated by another 5% using FP16/32 Batch-norm for Tensor Cores. 2018-04-17 02:51:11 +03:00
537d135feb Improve training performance - batch-norm using cuDNN. 2018-03-20 02:16:51 +03:00
880cf187d8 Fixed multi-GPU training for Tensor Cores 2018-03-09 19:44:46 +03:00
cad4d1618f Added support for Tensor Cores CC >= 7.0 (V100). For FP16/32 (mixed precision) define CUDNN_HALF should be used. 2018-02-25 16:29:44 +03:00
cd2bdec090 Updated to CUDA 9.1. And fixed no_gpu dependecies. 2018-02-23 15:05:31 +03:00
6332ea99ab one more fix 2018-02-23 00:13:08 +03:00
b2b5756d86 Added __float2half_rn() and __half2float() 2018-02-22 23:52:43 +03:00
dda993f3dd Use half_float16 instead of float32 if defined both CUDNN and CUDNN_HALF. Use Tensor Cores. 2018-02-22 22:54:40 +03:00
9920410ba9 minor fix 2017-07-14 12:11:45 +03:00
d7a30ada7e Fixed behavior if missing library cudnn.lib 2017-01-16 12:51:42 +03:00
3b9afd4cd2 Fixed behavior if missing library cudnn.lib 2017-01-16 00:44:41 +03:00
75fe603722 :vegan: :charizard: 2016-11-24 22:56:23 -08:00
c7a700dc22 new font strategy 2016-11-05 14:09:21 -07:00
352ae7e65b ADAM 2016-10-26 08:35:44 -07:00
73f7aacf35 better multigpu 2016-09-20 11:34:49 -07:00
5c067dc447 good chance I didn't break anything 2016-09-12 13:55:20 -07:00
8f1b4e0962 updates and things 2016-09-01 16:48:41 -07:00
afb8b4f98b CVPR prep 2016-06-22 21:46:32 -07:00
08c7cf9c88 no mean on input binarization 2016-06-19 14:28:15 -07:00
8322a58cf6 hate warnings 2016-06-14 11:30:28 -07:00
729ce43e6e stuff 2016-06-09 17:20:31 -07:00
ec3d050a76 hope i didn't break anything 2016-06-02 15:25:24 -07:00
13209df7bb art, cudnn 2016-05-13 11:59:43 -07:00
c7b10ceadb so much need to commit 2016-05-06 16:25:16 -07:00
cff59ba135 go updates 2016-03-16 04:30:48 -07:00
d1965bdb96 Go 2016-03-13 23:18:42 -07:00
16d06ec0db stuff 2016-02-29 13:54:12 -08:00
913d355ec1 lots of stuff 2016-01-28 12:30:38 -08:00
892923514f fixed darknet, stuff 2015-12-08 15:12:10 -08:00
c2738835f0 Faster batch normalization 2015-12-07 17:18:04 -08:00