--------------------------------------------------------------------------- Training stage 0 Sampled 1631 windows from 4250 images. Done sampling windows (time=13s). Extracting features... done (time=1s). Computing lambdas... done (time=6s). Sampled 5000 windows from 256 images. Done sampling windows (time=2s). Extracting features... done (time=2s). Training AdaBoost: nWeak= 32 nFtrs=1280 pos=3262 neg=5000 i= 16 alpha=0.391 err=0.314 loss=1.00e-01 i= 32 alpha=0.395 err=0.312 loss=2.83e-02 Done training err=0.0047 fp=0.0072 fn=0.0021 (t=0.2s). Done training stage 0 (time=24s). --------------------------------------------------------------------------- Training stage 1 Sampled 5000 windows from 256 images. Done sampling windows (time=4s). Extracting features... done (time=2s). Training AdaBoost: nWeak=128 nFtrs=1280 pos=3262 neg=10000 i= 16 alpha=0.194 err=0.404 loss=4.71e-01 i= 32 alpha=0.191 err=0.406 loss=3.44e-01 i= 48 alpha=0.170 err=0.416 loss=2.67e-01 i= 64 alpha=0.177 err=0.412 loss=2.11e-01 i= 80 alpha=0.164 err=0.419 loss=1.69e-01 i= 96 alpha=0.144 err=0.429 loss=1.36e-01 i= 112 alpha=0.152 err=0.424 loss=1.11e-01 i= 128 alpha=0.147 err=0.427 loss=9.04e-02 Done training err=0.0157 fp=0.0232 fn=0.0083 (t=0.6s). Done training stage 1 (time=7s). --------------------------------------------------------------------------- Training stage 2 Sampled 5000 windows from 256 images. Done sampling windows (time=4s). Extracting features... done (time=2s). Training AdaBoost: nWeak=512 nFtrs=1280 pos=3262 neg=10000 i= 16 alpha=0.176 err=0.413 loss=5.97e-01 i= 32 alpha=0.165 err=0.418 loss=4.74e-01 i= 48 alpha=0.137 err=0.432 loss=3.94e-01 i= 64 alpha=0.170 err=0.416 loss=3.35e-01 i= 80 alpha=0.129 err=0.436 loss=2.86e-01 i= 96 alpha=0.116 err=0.442 loss=2.49e-01 i= 112 alpha=0.130 err=0.435 loss=2.17e-01 i= 128 alpha=0.117 err=0.442 loss=1.89e-01 i= 144 alpha=0.120 err=0.441 loss=1.68e-01 i= 160 alpha=0.112 err=0.444 loss=1.46e-01 i= 176 alpha=0.114 err=0.443 loss=1.28e-01 i= 192 alpha=0.121 err=0.440 loss=1.13e-01 i= 208 alpha=0.131 err=0.435 loss=9.91e-02 i= 224 alpha=0.114 err=0.443 loss=8.76e-02 i= 240 alpha=0.125 err=0.438 loss=7.64e-02 i= 256 alpha=0.112 err=0.444 loss=6.81e-02 i= 272 alpha=0.119 err=0.441 loss=6.04e-02 i= 288 alpha=0.122 err=0.439 loss=5.38e-02 i= 304 alpha=0.128 err=0.436 loss=4.78e-02 i= 320 alpha=0.124 err=0.438 loss=4.23e-02 i= 336 alpha=0.125 err=0.438 loss=3.77e-02 i= 352 alpha=0.117 err=0.442 loss=3.35e-02 i= 368 alpha=0.127 err=0.437 loss=2.99e-02 i= 384 alpha=0.131 err=0.435 loss=2.65e-02 i= 400 alpha=0.108 err=0.446 loss=2.37e-02 i= 416 alpha=0.143 err=0.429 loss=2.09e-02 i= 432 alpha=0.120 err=0.440 loss=1.84e-02 i= 448 alpha=0.124 err=0.439 loss=1.63e-02 i= 464 alpha=0.106 err=0.447 loss=1.45e-02 i= 480 alpha=0.104 err=0.448 loss=1.30e-02 i= 496 alpha=0.108 err=0.446 loss=1.15e-02 i= 512 alpha=0.110 err=0.445 loss=1.03e-02 Done training err=0.0000 fp=0.0000 fn=0.0000 (t=1.8s). Done training stage 2 (time=7s). --------------------------------------------------------------------------- Training stage 3 Sampled 5000 windows from 384 images. Done sampling windows (time=5s). Extracting features... done (time=2s). Training AdaBoost: nWeak=2048 nFtrs=1280 pos=3262 neg=10000 i= 16 alpha=0.172 err=0.415 loss=6.86e-01 i= 32 alpha=0.141 err=0.430 loss=5.73e-01 i= 48 alpha=0.140 err=0.431 loss=4.89e-01 i= 64 alpha=0.113 err=0.444 loss=4.31e-01 i= 80 alpha=0.136 err=0.432 loss=3.81e-01 i= 96 alpha=0.126 err=0.437 loss=3.36e-01 i= 112 alpha=0.127 err=0.437 loss=2.98e-01 i= 128 alpha=0.110 err=0.445 loss=2.69e-01 i= 144 alpha=0.122 err=0.439 loss=2.39e-01 i= 160 alpha=0.143 err=0.429 loss=2.14e-01 i= 176 alpha=0.099 err=0.451 loss=1.92e-01 i= 192 alpha=0.114 err=0.443 loss=1.75e-01 i= 208 alpha=0.096 err=0.452 loss=1.58e-01 i= 224 alpha=0.111 err=0.445 loss=1.44e-01 i= 240 alpha=0.095 err=0.453 loss=1.30e-01 i= 256 alpha=0.100 err=0.450 loss=1.19e-01 i= 272 alpha=0.111 err=0.445 loss=1.08e-01 i= 288 alpha=0.100 err=0.450 loss=9.84e-02 i= 304 alpha=0.112 err=0.444 loss=8.99e-02 i= 320 alpha=0.109 err=0.446 loss=8.13e-02 i= 336 alpha=0.108 err=0.446 loss=7.36e-02 i= 352 alpha=0.104 err=0.448 loss=6.68e-02 i= 368 alpha=0.114 err=0.443 loss=6.12e-02 i= 384 alpha=0.091 err=0.454 loss=5.56e-02 i= 400 alpha=0.110 err=0.445 loss=5.05e-02 i= 416 alpha=0.108 err=0.446 loss=4.59e-02 i= 432 alpha=0.104 err=0.448 loss=4.18e-02 i= 448 alpha=0.113 err=0.444 loss=3.80e-02 i= 464 alpha=0.109 err=0.446 loss=3.48e-02 i= 480 alpha=0.111 err=0.445 loss=3.17e-02 i= 496 alpha=0.102 err=0.449 loss=2.89e-02 i= 512 alpha=0.108 err=0.446 loss=2.64e-02 i= 528 alpha=0.104 err=0.448 loss=2.41e-02 i= 544 alpha=0.119 err=0.441 loss=2.19e-02 i= 560 alpha=0.115 err=0.443 loss=1.99e-02 i= 576 alpha=0.116 err=0.442 loss=1.80e-02 i= 592 alpha=0.101 err=0.449 loss=1.64e-02 i= 608 alpha=0.095 err=0.453 loss=1.50e-02 i= 624 alpha=0.118 err=0.441 loss=1.36e-02 i= 640 alpha=0.109 err=0.446 loss=1.24e-02 i= 656 alpha=0.096 err=0.452 loss=1.14e-02 i= 672 alpha=0.106 err=0.447 loss=1.05e-02 i= 688 alpha=0.104 err=0.448 loss=9.64e-03 i= 704 alpha=0.108 err=0.446 loss=8.80e-03 i= 720 alpha=0.111 err=0.445 loss=7.98e-03 i= 736 alpha=0.111 err=0.445 loss=7.21e-03 i= 752 alpha=0.102 err=0.449 loss=6.59e-03 i= 768 alpha=0.106 err=0.447 loss=6.00e-03 i= 784 alpha=0.107 err=0.447 loss=5.46e-03 i= 800 alpha=0.118 err=0.441 loss=4.98e-03 i= 816 alpha=0.111 err=0.445 loss=4.55e-03 i= 832 alpha=0.098 err=0.451 loss=4.17e-03 i= 848 alpha=0.095 err=0.452 loss=3.83e-03 i= 864 alpha=0.097 err=0.451 loss=3.50e-03 i= 880 alpha=0.118 err=0.441 loss=3.20e-03 i= 896 alpha=0.097 err=0.452 loss=2.92e-03 i= 912 alpha=0.101 err=0.450 loss=2.67e-03 i= 928 alpha=0.114 err=0.443 loss=2.43e-03 i= 944 alpha=0.107 err=0.447 loss=2.23e-03 i= 960 alpha=0.094 err=0.453 loss=2.05e-03 i= 976 alpha=0.108 err=0.446 loss=1.88e-03 i= 992 alpha=0.089 err=0.456 loss=1.74e-03 i=1008 alpha=0.093 err=0.454 loss=1.58e-03 i=1024 alpha=0.106 err=0.447 loss=1.45e-03 i=1040 alpha=0.095 err=0.453 loss=1.33e-03 i=1056 alpha=0.132 err=0.435 loss=1.20e-03 i=1072 alpha=0.107 err=0.446 loss=1.09e-03 i=1088 alpha=0.101 err=0.450 loss=9.95e-04 i=1104 alpha=0.093 err=0.454 loss=9.12e-04 i=1120 alpha=0.097 err=0.452 loss=8.33e-04 i=1136 alpha=0.095 err=0.453 loss=7.63e-04 i=1152 alpha=0.109 err=0.446 loss=6.99e-04 i=1168 alpha=0.094 err=0.453 loss=6.41e-04 i=1184 alpha=0.100 err=0.450 loss=5.88e-04 i=1200 alpha=0.100 err=0.450 loss=5.38e-04 i=1216 alpha=0.096 err=0.452 loss=4.98e-04 i=1232 alpha=0.095 err=0.453 loss=4.58e-04 i=1248 alpha=0.108 err=0.446 loss=4.15e-04 i=1264 alpha=0.090 err=0.455 loss=3.80e-04 i=1280 alpha=0.109 err=0.446 loss=3.48e-04 i=1296 alpha=0.097 err=0.451 loss=3.19e-04 i=1312 alpha=0.110 err=0.445 loss=2.91e-04 i=1328 alpha=0.120 err=0.440 loss=2.64e-04 i=1344 alpha=0.112 err=0.444 loss=2.42e-04 i=1360 alpha=0.101 err=0.450 loss=2.22e-04 i=1376 alpha=0.101 err=0.450 loss=2.02e-04 i=1392 alpha=0.092 err=0.454 loss=1.84e-04 i=1408 alpha=0.089 err=0.456 loss=1.70e-04 i=1424 alpha=0.113 err=0.444 loss=1.55e-04 i=1440 alpha=0.114 err=0.443 loss=1.42e-04 i=1456 alpha=0.116 err=0.442 loss=1.30e-04 i=1472 alpha=0.107 err=0.447 loss=1.18e-04 i=1488 alpha=0.111 err=0.445 loss=1.07e-04 i=1504 alpha=0.110 err=0.445 loss=9.84e-05 i=1520 alpha=0.105 err=0.448 loss=9.00e-05 i=1536 alpha=0.093 err=0.454 loss=8.29e-05 i=1552 alpha=0.118 err=0.441 loss=7.57e-05 i=1568 alpha=0.092 err=0.454 loss=6.94e-05 i=1584 alpha=0.094 err=0.453 loss=6.36e-05 i=1600 alpha=0.107 err=0.447 loss=5.81e-05 i=1616 alpha=0.098 err=0.451 loss=5.31e-05 i=1632 alpha=0.100 err=0.450 loss=4.85e-05 i=1648 alpha=0.100 err=0.450 loss=4.46e-05 i=1664 alpha=0.108 err=0.446 loss=4.07e-05 i=1680 alpha=0.093 err=0.453 loss=3.74e-05 i=1696 alpha=0.109 err=0.446 loss=3.44e-05 i=1712 alpha=0.104 err=0.448 loss=3.13e-05 i=1728 alpha=0.114 err=0.443 loss=2.88e-05 i=1744 alpha=0.096 err=0.452 loss=2.64e-05 i=1760 alpha=0.097 err=0.452 loss=2.42e-05 i=1776 alpha=0.112 err=0.444 loss=2.21e-05 i=1792 alpha=0.105 err=0.448 loss=2.02e-05 i=1808 alpha=0.114 err=0.443 loss=1.85e-05 i=1824 alpha=0.126 err=0.437 loss=1.69e-05 i=1840 alpha=0.100 err=0.450 loss=1.55e-05 i=1856 alpha=0.109 err=0.446 loss=1.42e-05 i=1872 alpha=0.108 err=0.446 loss=1.30e-05 i=1888 alpha=0.112 err=0.444 loss=1.19e-05 i=1904 alpha=0.096 err=0.452 loss=1.09e-05 i=1920 alpha=0.084 err=0.458 loss=1.00e-05 i=1936 alpha=0.105 err=0.448 loss=9.21e-06 i=1952 alpha=0.119 err=0.441 loss=8.43e-06 i=1968 alpha=0.102 err=0.449 loss=7.72e-06 i=1984 alpha=0.117 err=0.442 loss=7.06e-06 i=2000 alpha=0.116 err=0.442 loss=6.48e-06 i=2016 alpha=0.100 err=0.450 loss=5.91e-06 i=2032 alpha=0.118 err=0.441 loss=5.36e-06 i=2048 alpha=0.114 err=0.443 loss=4.89e-06 Done training err=0.0000 fp=0.0000 fn=0.0000 (t=7.1s). Done training stage 3 (time=14s). --------------------------------------------------------------------------- Done training (time=53s).